Cloning into '/tmp/jaeger-tests'... Switched to a new branch 'integration-tests' branch 'integration-tests' set up to track 'origin/integration-tests'. Installing kuttl Try 0... curl -sLo /tmp/jaeger-tests/hack/install/../../bin/kubectl-kuttl https://github.com/kudobuilder/kuttl/releases/download/v0.15.0/kubectl-kuttl_0.15.0_linux_x86_64 KUBECONFIG file is: /tmp/kubeconfig-2531963025 for suite in elasticsearch examples generate miscellaneous sidecar streaming ui upgrade; do \ make run-e2e-tests-$suite ; \ done make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh elasticsearch false true + '[' 3 -ne 3 ']' + test_suite_name=elasticsearch + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/elasticsearch.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-elasticsearch make[2]: Entering directory '/tmp/jaeger-tests' >>>> Elasticsearch image not loaded because SKIP_ES_EXTERNAL is true SKIP_ES_EXTERNAL=true \ KAFKA_VERSION=3.6.0 \ SKIP_KAFKA=false \ ./tests/e2e/elasticsearch/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.18.0-0.nightly-2025-02-01-185117 True False 6m35s Cluster version is 4.18.0-0.nightly-2025-02-01-185117' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.18.0-0.nightly-2025-02-01-185117 True False 6m35s Cluster version is 4.18.0-0.nightly-2025-02-01-185117' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/elasticsearch/render.sh ++ export SUITE_DIR=./tests/e2e/elasticsearch ++ SUITE_DIR=./tests/e2e/elasticsearch ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate Try 0... curl -sLo /tmp/jaeger-tests/hack/install/../../bin/gomplate https://github.com/hairyhenderson/gomplate/releases/download/v3.10.0/gomplate_linux-amd64-slim ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq Try 0... curl -sLo /tmp/jaeger-tests/hack/install/../../bin/yq https://github.com/mikefarah/yq/releases/download/v4.20.2/yq_linux_amd64 ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize Try 0... curl -sLo /tmp/kustomize.tar.gz https://github.com/kubernetes-sigs/kustomize/releases/download/kustomize%2Fv4.5.7/kustomize_v4.5.7_linux_amd64.tar.gz Try 1... curl -sLo /tmp/kustomize.tar.gz https://github.com/kubernetes-sigs/kustomize/releases/download/kustomize%2Fv4.5.7/kustomize_v4.5.7_linux_amd64.tar.gz Try 2... curl -sLo /tmp/kustomize.tar.gz https://github.com/kubernetes-sigs/kustomize/releases/download/kustomize%2Fv4.5.7/kustomize_v4.5.7_linux_amd64.tar.gz Try 3... curl -sLo /tmp/kustomize.tar.gz https://github.com/kubernetes-sigs/kustomize/releases/download/kustomize%2Fv4.5.7/kustomize_v4.5.7_linux_amd64.tar.gz Try 4... curl -sLo /tmp/kustomize.tar.gz https://github.com/kubernetes-sigs/kustomize/releases/download/kustomize%2Fv4.5.7/kustomize_v4.5.7_linux_amd64.tar.gz ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/elasticsearch ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + is_secured=false + '[' true = true ']' + is_secured=true + start_test es-from-aio-to-production + '[' 1 -ne 1 ']' + test_name=es-from-aio-to-production + echo =========================================================================== =========================================================================== + info 'Rendering files for test es-from-aio-to-production' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test es-from-aio-to-production\e[0m' Rendering files for test es-from-aio-to-production + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build + '[' _build '!=' _build ']' + mkdir -p es-from-aio-to-production + cd es-from-aio-to-production + jaeger_name=my-jaeger + render_install_jaeger my-jaeger allInOne 00 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=allInOne + test_step=00 + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test my-jaeger true 01 + '[' 3 -ne 3 ']' + jaeger=my-jaeger + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + jaeger_deploy_mode=production + [[ true = true ]] + [[ true = true ]] + jaeger_deploy_mode=production_autoprovisioned + render_install_jaeger my-jaeger production_autoprovisioned 03 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=production_autoprovisioned + test_step=03 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./03-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./03-assert.yaml + [[ true = true ]] + [[ true = true ]] + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.elasticsearch.redundancyPolicy="ZeroRedundancy"' ./03-install.yaml + render_smoke_test my-jaeger true 04 + '[' 3 -ne 3 ']' + jaeger=my-jaeger + is_secured=true + test_step=04 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./04-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./04-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test es-increasing-replicas + '[' 1 -ne 1 ']' + test_name=es-increasing-replicas + echo =========================================================================== =========================================================================== + info 'Rendering files for test es-increasing-replicas' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test es-increasing-replicas\e[0m' Rendering files for test es-increasing-replicas + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build/es-from-aio-to-production + '[' es-from-aio-to-production '!=' _build ']' + cd .. + mkdir -p es-increasing-replicas + cd es-increasing-replicas + jaeger_name=simple-prod + '[' true = true ']' + jaeger_deployment_mode=production_autoprovisioned + render_install_jaeger simple-prod production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=simple-prod + JAEGER_NAME=simple-prod + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + cp ./01-install.yaml ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.collector.replicas=2 ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.query.replicas=2 ./02-install.yaml + cp ./01-assert.yaml ./02-assert.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.replicas=2 ./02-assert.yaml + /tmp/jaeger-tests/bin/yq e -i .status.readyReplicas=2 ./02-assert.yaml + render_smoke_test simple-prod true 03 + '[' 3 -ne 3 ']' + jaeger=simple-prod + is_secured=true + test_step=03 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + export JAEGER_NAME=simple-prod + JAEGER_NAME=simple-prod + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./03-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./03-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + '[' true = true ']' + cp ./02-install.yaml ./04-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.storage.elasticsearch.nodeCount=2 ./04-install.yaml + /tmp/jaeger-tests/bin/gomplate -f ./openshift-check-es-nodes.yaml.template -o ./05-check-es-nodes.yaml + '[' true = true ']' + skip_test es-index-cleaner-upstream 'SKIP_ES_EXTERNAL is true' + '[' 2 -ne 2 ']' + test_name=es-index-cleaner-upstream + message='SKIP_ES_EXTERNAL is true' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build/es-increasing-replicas + '[' es-increasing-replicas '!=' _build ']' + cd .. + rm -rf es-index-cleaner-upstream + warning 'es-index-cleaner-upstream: SKIP_ES_EXTERNAL is true' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: es-index-cleaner-upstream: SKIP_ES_EXTERNAL is true\e[0m' WAR: es-index-cleaner-upstream: SKIP_ES_EXTERNAL is true + '[' true = true ']' + es_index_cleaner -autoprov production_autoprovisioned + '[' 2 -ne 2 ']' + postfix=-autoprov + jaeger_deployment_strategy=production_autoprovisioned + start_test es-index-cleaner-autoprov + '[' 1 -ne 1 ']' + test_name=es-index-cleaner-autoprov + echo =========================================================================== =========================================================================== + info 'Rendering files for test es-index-cleaner-autoprov' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test es-index-cleaner-autoprov\e[0m' Rendering files for test es-index-cleaner-autoprov + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build + '[' _build '!=' _build ']' + mkdir -p es-index-cleaner-autoprov + cd es-index-cleaner-autoprov + jaeger_name=test-es-index-cleaner-with-prefix + cronjob_name=test-es-index-cleaner-with-prefix-es-index-cleaner + secured_es_connection=false + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_managed_es ']' + ELASTICSEARCH_URL=https://elasticsearch + secured_es_connection=true + cp ../../es-index-cleaner-upstream/04-assert.yaml ../../es-index-cleaner-upstream/README.md . + render_install_jaeger test-es-index-cleaner-with-prefix production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=test-es-index-cleaner-with-prefix + JAEGER_NAME=test-es-index-cleaner-with-prefix + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.options.es.index-prefix=""' ./01-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.storage.esIndexCleaner.enabled=false ./01-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.storage.esIndexCleaner.numberOfDays=0 ./01-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.esIndexCleaner.schedule="*/1 * * * *"' ./01-install.yaml + render_report_spans test-es-index-cleaner-with-prefix true 5 00 true 02 + '[' 6 -ne 6 ']' + jaeger=test-es-index-cleaner-with-prefix + is_secured=true + number_of_spans=5 + job_number=00 + ensure_reported_spans=true + test_step=02 + export JAEGER_NAME=test-es-index-cleaner-with-prefix + JAEGER_NAME=test-es-index-cleaner-with-prefix + export JAEGER_COLLECTOR_ENDPOINT=http://test-es-index-cleaner-with-prefix-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://test-es-index-cleaner-with-prefix-collector-headless:14268 + export JOB_NUMBER=00 + JOB_NUMBER=00 + export DAYS=5 + DAYS=5 + '[' true = true ']' + protocol=https:// + query_port= + template=/tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template + '[' true = true ']' + export ENSURE_REPORTED_SPANS=true + ENSURE_REPORTED_SPANS=true + export JAEGER_QUERY_ENDPOINT=https://test-es-index-cleaner-with-prefix-query + JAEGER_QUERY_ENDPOINT=https://test-es-index-cleaner-with-prefix-query + params= + '[' true = true ']' + '[' true = true ']' + '[' '' '!=' allInOne ']' + params='-t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template -t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template -o ./02-report-spans.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-report-spans.yaml.template -o ./02-assert.yaml + unset JAEGER_COLLECTOR_ENDPOINT + unset JAEGER_QUERY_ENDPOINT + unset JOB_NUMBER + unset DAYS + unset ENSURE_REPORTED_SPANS + sed 's~enabled: false~enabled: true~gi' ./01-install.yaml + CRONJOB_NAME=test-es-index-cleaner-with-prefix-es-index-cleaner + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/wait-for-cronjob-execution.yaml.template -o ./04-wait-es-index-cleaner.yaml + /tmp/jaeger-tests/bin/gomplate -f ./01-install.yaml -o ./05-install.yaml + render_check_indices true ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' 00 06 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + job_number=00 + test_step=06 + escape_command ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + '[' 1 -ne 1 ']' + command=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' ++ sed 's/\\/\\\\/g' ++ echo ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + export 'CMD_PARAMETERS='\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=test-es-index-cleaner-with-prefix-curator + JOB_NUMBER=00 + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + MOUNT_SECRET=test-es-index-cleaner-with-prefix-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./06-check-indices.yaml + JOB_NUMBER=00 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./06-assert.yaml + '[' true = true ']' + get_elasticsearch_openshift_operator_version + export ESO_OPERATOR_VERSION + '[' true = true ']' ++ kubectl get pods -l name=elasticsearch-operator --all-namespaces '-o=jsonpath={.items[0].metadata.annotations.operatorframework\.io/properties}' + properties='{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.18},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.16"}}]}' + '[' -z '{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.18},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.16"}}]}' ']' ++ echo '{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.18},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.16"}}]}' ++ /tmp/jaeger-tests/bin/yq e -P '.properties.[] | select(.value.packageName == "elasticsearch-operator") | .value.version' + ESO_OPERATOR_VERSION=5.8.16 ++ version_ge 5.8.16 5.4 +++ echo 5.8.16 5.4 +++ sort -rV +++ tr ' ' '\n' +++ head -n 1 ++ test 5.8.16 == 5.8.16 + '[' -n '' ']' + skip_test es-index-cleaner-managed 'Test only supported with Elasticsearch OpenShift Operator >= 5.4' + '[' 2 -ne 2 ']' + test_name=es-index-cleaner-managed + message='Test only supported with Elasticsearch OpenShift Operator >= 5.4' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build/es-index-cleaner-autoprov + '[' es-index-cleaner-autoprov '!=' _build ']' + cd .. + rm -rf es-index-cleaner-managed + warning 'es-index-cleaner-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: es-index-cleaner-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4\e[0m' WAR: es-index-cleaner-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4 + '[' true = true ']' + start_test es-multiinstance + '[' 1 -ne 1 ']' + test_name=es-multiinstance + echo =========================================================================== =========================================================================== + info 'Rendering files for test es-multiinstance' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test es-multiinstance\e[0m' Rendering files for test es-multiinstance + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build + '[' _build '!=' _build ']' + mkdir -p es-multiinstance + cd es-multiinstance + jaeger_name=instance-1 + render_install_jaeger instance-1 production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=instance-1 + JAEGER_NAME=instance-1 + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + /tmp/jaeger-tests/bin/gomplate -f ./03-create-second-instance.yaml.template -o 03-create-second-instance.yaml + '[' true = true ']' + skip_test es-rollover-upstream 'SKIP_ES_EXTERNAL is true' + '[' 2 -ne 2 ']' + test_name=es-rollover-upstream + message='SKIP_ES_EXTERNAL is true' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build/es-multiinstance + '[' es-multiinstance '!=' _build ']' + cd .. + rm -rf es-rollover-upstream + warning 'es-rollover-upstream: SKIP_ES_EXTERNAL is true' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: es-rollover-upstream: SKIP_ES_EXTERNAL is true\e[0m' WAR: es-rollover-upstream: SKIP_ES_EXTERNAL is true + '[' true = true ']' + es_rollover -autoprov production_autoprovisioned + '[' 2 -ne 2 ']' + postfix=-autoprov + jaeger_deployment_strategy=production_autoprovisioned + start_test es-rollover-autoprov + '[' 1 -ne 1 ']' + test_name=es-rollover-autoprov + echo =========================================================================== =========================================================================== + info 'Rendering files for test es-rollover-autoprov' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test es-rollover-autoprov\e[0m' Rendering files for test es-rollover-autoprov + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build + '[' _build '!=' _build ']' + mkdir -p es-rollover-autoprov + cd es-rollover-autoprov + cp ../../es-rollover-upstream/05-assert.yaml ../../es-rollover-upstream/05-install.yaml ../../es-rollover-upstream/README.md . + jaeger_name=my-jaeger + secured_es_connection=false + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_managed_es ']' + ELASTICSEARCH_URL=https://elasticsearch + secured_es_connection=true + render_install_jaeger my-jaeger production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + render_report_spans my-jaeger true 2 00 true 02 + '[' 6 -ne 6 ']' + jaeger=my-jaeger + is_secured=true + number_of_spans=2 + job_number=00 + ensure_reported_spans=true + test_step=02 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JOB_NUMBER=00 + JOB_NUMBER=00 + export DAYS=2 + DAYS=2 + '[' true = true ']' + protocol=https:// + query_port= + template=/tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template + '[' true = true ']' + export ENSURE_REPORTED_SPANS=true + ENSURE_REPORTED_SPANS=true + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + params= + '[' true = true ']' + '[' true = true ']' + '[' '' '!=' allInOne ']' + params='-t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template -t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template -o ./02-report-spans.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-report-spans.yaml.template -o ./02-assert.yaml + unset JAEGER_COLLECTOR_ENDPOINT + unset JAEGER_QUERY_ENDPOINT + unset JOB_NUMBER + unset DAYS + unset ENSURE_REPORTED_SPANS + render_check_indices true ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' 00 03 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + job_number=00 + test_step=03 + escape_command ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + '[' 1 -ne 1 ']' + command=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' ++ echo ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-exist'\'',' + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-exist'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=00 + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./03-check-indices.yaml + JOB_NUMBER=00 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./03-assert.yaml + render_check_indices true ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' 01 04 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + job_number=01 + test_step=04 + escape_command ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + '[' 1 -ne 1 ']' + command=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' ++ echo ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--pattern'\'', '\''jaeger-span-\\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=01 + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./04-check-indices.yaml + JOB_NUMBER=01 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./04-assert.yaml + render_report_spans my-jaeger true 2 02 true 06 + '[' 6 -ne 6 ']' + jaeger=my-jaeger + is_secured=true + number_of_spans=2 + job_number=02 + ensure_reported_spans=true + test_step=06 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JOB_NUMBER=02 + JOB_NUMBER=02 + export DAYS=2 + DAYS=2 + '[' true = true ']' + protocol=https:// + query_port= + template=/tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template + '[' true = true ']' + export ENSURE_REPORTED_SPANS=true + ENSURE_REPORTED_SPANS=true + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + params= + '[' true = true ']' + '[' true = true ']' + '[' '' '!=' allInOne ']' + params='-t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template -t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template -o ./06-report-spans.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-report-spans.yaml.template -o ./06-assert.yaml + unset JAEGER_COLLECTOR_ENDPOINT + unset JAEGER_QUERY_ENDPOINT + unset JOB_NUMBER + unset DAYS + unset ENSURE_REPORTED_SPANS + render_check_indices true ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' 02 07 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + job_number=02 + test_step=07 + escape_command ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + '[' 1 -ne 1 ']' + command=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' ++ echo ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-exist'\'',' + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-exist'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=02 + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./07-check-indices.yaml + JOB_NUMBER=02 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./07-assert.yaml + render_check_indices true ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' 03 08 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' + job_number=03 + test_step=08 + escape_command ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' + '[' 1 -ne 1 ']' + command=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' ++ echo ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--pattern'\'', '\''jaeger-span-\\d{6}'\'', '\''--assert-exist'\'',' + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\\d{6}'\'', '\''--assert-exist'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=03 + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./08-check-indices.yaml + JOB_NUMBER=03 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./08-assert.yaml + render_check_indices true ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' 04 09 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' + job_number=04 + test_step=09 + escape_command ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' + '[' 1 -ne 1 ']' + command=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' ++ echo ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=04 + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./09-check-indices.yaml + JOB_NUMBER=04 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./09-assert.yaml + render_report_spans my-jaeger true 2 03 true 10 + '[' 6 -ne 6 ']' + jaeger=my-jaeger + is_secured=true + number_of_spans=2 + job_number=03 + ensure_reported_spans=true + test_step=10 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JOB_NUMBER=03 + JOB_NUMBER=03 + export DAYS=2 + DAYS=2 + '[' true = true ']' + protocol=https:// + query_port= + template=/tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template + '[' true = true ']' + export ENSURE_REPORTED_SPANS=true + ENSURE_REPORTED_SPANS=true + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + params= + '[' true = true ']' + '[' true = true ']' + '[' '' '!=' allInOne ']' + params='-t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template -t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template -o ./10-report-spans.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-report-spans.yaml.template -o ./10-assert.yaml + unset JAEGER_COLLECTOR_ENDPOINT + unset JAEGER_QUERY_ENDPOINT + unset JOB_NUMBER + unset DAYS + unset ENSURE_REPORTED_SPANS + CRONJOB_NAME=my-jaeger-es-rollover + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/wait-for-cronjob-execution.yaml.template -o ./11-wait-rollover.yaml + render_check_indices true ''\''--name'\'', '\''jaeger-span-000002'\'',' 05 11 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--name'\'', '\''jaeger-span-000002'\'',' + job_number=05 + test_step=11 + escape_command ''\''--name'\'', '\''jaeger-span-000002'\'',' + '[' 1 -ne 1 ']' + command=''\''--name'\'', '\''jaeger-span-000002'\'',' ++ echo ''\''--name'\'', '\''jaeger-span-000002'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--name'\'', '\''jaeger-span-000002'\'',' + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-000002'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=05 + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-000002'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./11-check-indices.yaml + JOB_NUMBER=05 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./11-assert.yaml + render_check_indices true ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' 06 12 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' + job_number=06 + test_step=12 + escape_command ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' + '[' 1 -ne 1 ']' + command=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' ++ echo ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=06 + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./12-check-indices.yaml + JOB_NUMBER=06 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./12-assert.yaml + '[' true = true ']' + get_elasticsearch_openshift_operator_version + export ESO_OPERATOR_VERSION + '[' true = true ']' ++ kubectl get pods -l name=elasticsearch-operator --all-namespaces '-o=jsonpath={.items[0].metadata.annotations.operatorframework\.io/properties}' + properties='{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.18},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.16"}}]}' + '[' -z '{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.18},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.16"}}]}' ']' ++ echo '{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.18},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.16"}}]}' ++ /tmp/jaeger-tests/bin/yq e -P '.properties.[] | select(.value.packageName == "elasticsearch-operator") | .value.version' + ESO_OPERATOR_VERSION=5.8.16 ++ version_ge 5.8.16 5.4 +++ echo 5.8.16 5.4 +++ tr ' ' '\n' +++ sort -rV +++ head -n 1 ++ test 5.8.16 == 5.8.16 + '[' -n '' ']' + skip_test es-rollover-managed 'Test only supported with Elasticsearch OpenShift Operator >= 5.4' + '[' 2 -ne 2 ']' + test_name=es-rollover-managed + message='Test only supported with Elasticsearch OpenShift Operator >= 5.4' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build/es-rollover-autoprov + '[' es-rollover-autoprov '!=' _build ']' + cd .. + rm -rf es-rollover-managed + warning 'es-rollover-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: es-rollover-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4\e[0m' WAR: es-rollover-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4 + '[' true = true ']' + skip_test es-spark-dependencies 'This test is not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=es-spark-dependencies + message='This test is not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build + '[' _build '!=' _build ']' + rm -rf es-spark-dependencies + warning 'es-spark-dependencies: This test is not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: es-spark-dependencies: This test is not supported in OpenShift\e[0m' WAR: es-spark-dependencies: This test is not supported in OpenShift make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running elasticsearch E2E tests' Running elasticsearch E2E tests + cd tests/e2e/elasticsearch/_build + set +e + KUBECONFIG=/tmp/kubeconfig-2531963025 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-d102g9k0-f1a62.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 7 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/es-from-aio-to-production === PAUSE kuttl/harness/es-from-aio-to-production === RUN kuttl/harness/es-increasing-replicas === PAUSE kuttl/harness/es-increasing-replicas === RUN kuttl/harness/es-index-cleaner-autoprov === PAUSE kuttl/harness/es-index-cleaner-autoprov === RUN kuttl/harness/es-multiinstance === PAUSE kuttl/harness/es-multiinstance === RUN kuttl/harness/es-rollover-autoprov === PAUSE kuttl/harness/es-rollover-autoprov === RUN kuttl/harness/es-simple-prod === PAUSE kuttl/harness/es-simple-prod === CONT kuttl/harness/artifacts logger.go:42: 07:16:14 | artifacts | Creating namespace: kuttl-test-nice-dogfish logger.go:42: 07:16:14 | artifacts | artifacts events from ns kuttl-test-nice-dogfish: logger.go:42: 07:16:14 | artifacts | Deleting namespace: kuttl-test-nice-dogfish === CONT kuttl/harness/es-multiinstance logger.go:42: 07:16:21 | es-multiinstance | Ignoring 03-create-second-instance.yaml.template as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:16:21 | es-multiinstance | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:16:21 | es-multiinstance | Creating namespace: kuttl-test-giving-bug logger.go:42: 07:16:21 | es-multiinstance/0-clear-namespace | starting test step 0-clear-namespace logger.go:42: 07:16:21 | es-multiinstance/0-clear-namespace | running command: [sh -c kubectl delete namespace jaeger-e2e-multiinstance-test --ignore-not-found=true] logger.go:42: 07:16:21 | es-multiinstance/0-clear-namespace | test step completed 0-clear-namespace logger.go:42: 07:16:21 | es-multiinstance/1-install | starting test step 1-install logger.go:42: 07:16:21 | es-multiinstance/1-install | Jaeger:kuttl-test-giving-bug/instance-1 created logger.go:42: 07:17:28 | es-multiinstance/1-install | test step completed 1-install logger.go:42: 07:17:28 | es-multiinstance/2-create-namespace | starting test step 2-create-namespace logger.go:42: 07:17:28 | es-multiinstance/2-create-namespace | running command: [sh -c kubectl create namespace jaeger-e2e-multiinstance-test] logger.go:42: 07:17:28 | es-multiinstance/2-create-namespace | namespace/jaeger-e2e-multiinstance-test created logger.go:42: 07:17:28 | es-multiinstance/2-create-namespace | test step completed 2-create-namespace logger.go:42: 07:17:28 | es-multiinstance/3-create-second-instance | starting test step 3-create-second-instance logger.go:42: 07:17:28 | es-multiinstance/3-create-second-instance | running command: [sh -c kubectl apply -f ./01-install.yaml -n jaeger-e2e-multiinstance-test] logger.go:42: 07:17:28 | es-multiinstance/3-create-second-instance | jaeger.jaegertracing.io/instance-1 created logger.go:42: 07:17:28 | es-multiinstance/3-create-second-instance | running command: [sh -c /tmp/jaeger-tests/bin/kubectl-kuttl assert ./01-assert.yaml -n jaeger-e2e-multiinstance-test --timeout 1000] logger.go:42: 07:18:32 | es-multiinstance/3-create-second-instance | assert is valid logger.go:42: 07:18:32 | es-multiinstance/3-create-second-instance | test step completed 3-create-second-instance logger.go:42: 07:18:32 | es-multiinstance/4-check-secrets | starting test step 4-check-secrets logger.go:42: 07:18:32 | es-multiinstance/4-check-secrets | running command: [sh -c kubectl get secrets elasticsearch -o jsonpath='{.data.logging-es\.crt}' -n $NAMESPACE > secret1] logger.go:42: 07:18:32 | es-multiinstance/4-check-secrets | running command: [sh -c kubectl get secrets elasticsearch -o jsonpath='{.data.logging-es\.crt}' -n jaeger-e2e-multiinstance-test > secret2] logger.go:42: 07:18:32 | es-multiinstance/4-check-secrets | running command: [sh -c cmp --silent secret1 secret2 || exit 0] logger.go:42: 07:18:32 | es-multiinstance/4-check-secrets | test step completed 4-check-secrets logger.go:42: 07:18:32 | es-multiinstance/5-delete | starting test step 5-delete logger.go:42: 07:18:32 | es-multiinstance/5-delete | running command: [sh -c kubectl delete namespace jaeger-e2e-multiinstance-test --wait=false] logger.go:42: 07:18:32 | es-multiinstance/5-delete | namespace "jaeger-e2e-multiinstance-test" deleted logger.go:42: 07:18:32 | es-multiinstance/5-delete | test step completed 5-delete logger.go:42: 07:18:32 | es-multiinstance | es-multiinstance events from ns kuttl-test-giving-bug: logger.go:42: 07:18:32 | es-multiinstance | 2025-02-03 07:16:39 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestgivingbuginstance1-1-cb5bb7f45-wghj6 Binding Scheduled Successfully assigned kuttl-test-giving-bug/elasticsearch-cdm-kuttltestgivingbuginstance1-1-cb5bb7f45-wghj6 to ip-10-0-66-98.us-east-2.compute.internal default-scheduler logger.go:42: 07:18:32 | es-multiinstance | 2025-02-03 07:16:39 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestgivingbuginstance1-1-cb5bb7f45-wghj6 AddedInterface Add eth0 [10.129.2.20/23] from ovn-kubernetes multus logger.go:42: 07:18:32 | es-multiinstance | 2025-02-03 07:16:39 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestgivingbuginstance1-1-cb5bb7f45-wghj6.spec.containers{elasticsearch} Pulling Pulling image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:410674bf2f77ff1d431f494cadfab3e81249b35695eb65aedfdb8498079b4d35" kubelet logger.go:42: 07:18:32 | es-multiinstance | 2025-02-03 07:16:39 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestgivingbuginstance1-1-cb5bb7f45 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestgivingbuginstance1-1-cb5bb7f45-wghj6 replicaset-controller logger.go:42: 07:18:32 | es-multiinstance | 2025-02-03 07:16:39 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestgivingbuginstance1-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestgivingbuginstance1-1-cb5bb7f45 to 1 deployment-controller logger.go:42: 07:18:32 | es-multiinstance | 2025-02-03 07:16:47 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestgivingbuginstance1-1-cb5bb7f45-wghj6.spec.containers{elasticsearch} Pulled Successfully pulled image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:410674bf2f77ff1d431f494cadfab3e81249b35695eb65aedfdb8498079b4d35" in 8.074s (8.074s including waiting). Image size: 539381180 bytes. kubelet logger.go:42: 07:18:32 | es-multiinstance | 2025-02-03 07:16:48 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestgivingbuginstance1-1-cb5bb7f45-wghj6.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:18:32 | es-multiinstance | 2025-02-03 07:16:48 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestgivingbuginstance1-1-cb5bb7f45-wghj6.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:18:32 | es-multiinstance | 2025-02-03 07:16:48 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestgivingbuginstance1-1-cb5bb7f45-wghj6.spec.containers{proxy} Pulling Pulling image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:fe2439e5f9e14d828caace425214edfa3edc179d72d225c43f69e9e4c9a5837b" kubelet logger.go:42: 07:18:32 | es-multiinstance | 2025-02-03 07:16:50 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestgivingbuginstance1-1-cb5bb7f45-wghj6.spec.containers{proxy} Pulled Successfully pulled image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:fe2439e5f9e14d828caace425214edfa3edc179d72d225c43f69e9e4c9a5837b" in 1.94s (1.94s including waiting). Image size: 287327885 bytes. kubelet logger.go:42: 07:18:32 | es-multiinstance | 2025-02-03 07:16:50 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestgivingbuginstance1-1-cb5bb7f45-wghj6.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:18:32 | es-multiinstance | 2025-02-03 07:16:50 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestgivingbuginstance1-1-cb5bb7f45-wghj6.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:18:32 | es-multiinstance | 2025-02-03 07:16:59 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestgivingbuginstance1-1-cb5bb7f45-wghj6.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 07:18:32 | es-multiinstance | 2025-02-03 07:17:04 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestgivingbuginstance1-1-cb5bb7f45-wghj6.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:18:32 | es-multiinstance | 2025-02-03 07:17:16 +0000 UTC Normal Pod instance-1-collector-6dcfd8d596-xsxmj Binding Scheduled Successfully assigned kuttl-test-giving-bug/instance-1-collector-6dcfd8d596-xsxmj to ip-10-0-6-172.us-east-2.compute.internal default-scheduler logger.go:42: 07:18:32 | es-multiinstance | 2025-02-03 07:17:16 +0000 UTC Normal Pod instance-1-collector-6dcfd8d596-xsxmj AddedInterface Add eth0 [10.131.0.22/23] from ovn-kubernetes multus logger.go:42: 07:18:32 | es-multiinstance | 2025-02-03 07:17:16 +0000 UTC Normal Pod instance-1-collector-6dcfd8d596-xsxmj.spec.containers{jaeger-collector} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:40a1c7fa23aa5ffa64e8e22aa38022f5d4d7ff644c46a3da7169b713d486c3c1" kubelet logger.go:42: 07:18:32 | es-multiinstance | 2025-02-03 07:17:16 +0000 UTC Normal ReplicaSet.apps instance-1-collector-6dcfd8d596 SuccessfulCreate Created pod: instance-1-collector-6dcfd8d596-xsxmj replicaset-controller logger.go:42: 07:18:32 | es-multiinstance | 2025-02-03 07:17:16 +0000 UTC Normal Deployment.apps instance-1-collector ScalingReplicaSet Scaled up replica set instance-1-collector-6dcfd8d596 to 1 deployment-controller logger.go:42: 07:18:32 | es-multiinstance | 2025-02-03 07:17:16 +0000 UTC Normal Pod instance-1-query-d7d5db969-x7bj4 Binding Scheduled Successfully assigned kuttl-test-giving-bug/instance-1-query-d7d5db969-x7bj4 to ip-10-0-26-111.us-east-2.compute.internal default-scheduler logger.go:42: 07:18:32 | es-multiinstance | 2025-02-03 07:17:16 +0000 UTC Normal ReplicaSet.apps instance-1-query-d7d5db969 SuccessfulCreate Created pod: instance-1-query-d7d5db969-x7bj4 replicaset-controller logger.go:42: 07:18:32 | es-multiinstance | 2025-02-03 07:17:16 +0000 UTC Normal Deployment.apps instance-1-query ScalingReplicaSet Scaled up replica set instance-1-query-d7d5db969 to 1 deployment-controller logger.go:42: 07:18:32 | es-multiinstance | 2025-02-03 07:17:17 +0000 UTC Normal Pod instance-1-query-d7d5db969-x7bj4 AddedInterface Add eth0 [10.128.2.22/23] from ovn-kubernetes multus logger.go:42: 07:18:32 | es-multiinstance | 2025-02-03 07:17:17 +0000 UTC Normal Pod instance-1-query-d7d5db969-x7bj4.spec.containers{jaeger-query} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:84020ce29bcb5a4bd018e6596188ae919c5cd600e08f78a546c0e76ea477685e" kubelet logger.go:42: 07:18:32 | es-multiinstance | 2025-02-03 07:17:19 +0000 UTC Normal Pod instance-1-collector-6dcfd8d596-xsxmj.spec.containers{jaeger-collector} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:40a1c7fa23aa5ffa64e8e22aa38022f5d4d7ff644c46a3da7169b713d486c3c1" in 2.494s (2.494s including waiting). Image size: 139779827 bytes. kubelet logger.go:42: 07:18:32 | es-multiinstance | 2025-02-03 07:17:19 +0000 UTC Normal Pod instance-1-collector-6dcfd8d596-xsxmj.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:18:32 | es-multiinstance | 2025-02-03 07:17:19 +0000 UTC Normal Pod instance-1-collector-6dcfd8d596-xsxmj.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:18:32 | es-multiinstance | 2025-02-03 07:17:19 +0000 UTC Normal Pod instance-1-query-d7d5db969-x7bj4.spec.containers{jaeger-query} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:84020ce29bcb5a4bd018e6596188ae919c5cd600e08f78a546c0e76ea477685e" in 2.272s (2.272s including waiting). Image size: 142020742 bytes. kubelet logger.go:42: 07:18:32 | es-multiinstance | 2025-02-03 07:17:19 +0000 UTC Normal Pod instance-1-query-d7d5db969-x7bj4.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:18:32 | es-multiinstance | 2025-02-03 07:17:19 +0000 UTC Normal Pod instance-1-query-d7d5db969-x7bj4.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:18:32 | es-multiinstance | 2025-02-03 07:17:19 +0000 UTC Normal Pod instance-1-query-d7d5db969-x7bj4.spec.containers{oauth-proxy} Pulling Pulling image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" kubelet logger.go:42: 07:18:32 | es-multiinstance | 2025-02-03 07:17:23 +0000 UTC Normal Pod instance-1-query-d7d5db969-x7bj4.spec.containers{oauth-proxy} Pulled Successfully pulled image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" in 3.998s (3.998s including waiting). Image size: 438322369 bytes. kubelet logger.go:42: 07:18:32 | es-multiinstance | 2025-02-03 07:17:23 +0000 UTC Normal Pod instance-1-query-d7d5db969-x7bj4.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:18:32 | es-multiinstance | 2025-02-03 07:17:23 +0000 UTC Normal Pod instance-1-query-d7d5db969-x7bj4.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:18:32 | es-multiinstance | 2025-02-03 07:17:23 +0000 UTC Normal Pod instance-1-query-d7d5db969-x7bj4.spec.containers{jaeger-agent} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:65342c7d622f336741a8ae0b9dead79c7ecabc155d02cdd42b7b49ca36680e74" kubelet logger.go:42: 07:18:32 | es-multiinstance | 2025-02-03 07:17:26 +0000 UTC Normal Pod instance-1-query-d7d5db969-x7bj4.spec.containers{jaeger-agent} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:65342c7d622f336741a8ae0b9dead79c7ecabc155d02cdd42b7b49ca36680e74" in 2.216s (2.216s including waiting). Image size: 112614125 bytes. kubelet logger.go:42: 07:18:32 | es-multiinstance | 2025-02-03 07:17:26 +0000 UTC Normal Pod instance-1-query-d7d5db969-x7bj4.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:18:32 | es-multiinstance | 2025-02-03 07:17:26 +0000 UTC Normal Pod instance-1-query-d7d5db969-x7bj4.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:18:32 | es-multiinstance | 2025-02-03 07:17:44 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling instance-1-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:18:32 | es-multiinstance | 2025-02-03 07:17:44 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling instance-1-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:18:32 | es-multiinstance | 2025-02-03 07:17:44 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling instance-1-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:18:32 | es-multiinstance | 2025-02-03 07:18:14 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling instance-1-collector FailedGetResourceMetric failed to get cpu utilization: missing request for cpu in container jaeger-collector of Pod instance-1-collector-6dcfd8d596-xsxmj horizontal-pod-autoscaler logger.go:42: 07:18:32 | es-multiinstance | 2025-02-03 07:18:14 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling instance-1-collector FailedGetResourceMetric failed to get memory utilization: missing request for memory in container jaeger-collector of Pod instance-1-collector-6dcfd8d596-xsxmj horizontal-pod-autoscaler logger.go:42: 07:18:32 | es-multiinstance | 2025-02-03 07:18:14 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling instance-1-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: missing request for cpu in container jaeger-collector of Pod instance-1-collector-6dcfd8d596-xsxmj horizontal-pod-autoscaler logger.go:42: 07:18:32 | es-multiinstance | Deleting namespace: kuttl-test-giving-bug === CONT kuttl/harness/es-simple-prod logger.go:42: 07:18:40 | es-simple-prod | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:18:40 | es-simple-prod | Creating namespace: kuttl-test-mature-eel logger.go:42: 07:18:40 | es-simple-prod | es-simple-prod events from ns kuttl-test-mature-eel: logger.go:42: 07:18:40 | es-simple-prod | Deleting namespace: kuttl-test-mature-eel === CONT kuttl/harness/es-rollover-autoprov logger.go:42: 07:18:46 | es-rollover-autoprov | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:18:46 | es-rollover-autoprov | Creating namespace: kuttl-test-settling-wallaby logger.go:42: 07:18:46 | es-rollover-autoprov/1-install | starting test step 1-install logger.go:42: 07:18:46 | es-rollover-autoprov/1-install | Jaeger:kuttl-test-settling-wallaby/my-jaeger created logger.go:42: 07:19:42 | es-rollover-autoprov/1-install | test step completed 1-install logger.go:42: 07:19:42 | es-rollover-autoprov/2-report-spans | starting test step 2-report-spans logger.go:42: 07:19:42 | es-rollover-autoprov/2-report-spans | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:19:43 | es-rollover-autoprov/2-report-spans | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:19:53 | es-rollover-autoprov/2-report-spans | running command: [sh -c DAYS=2 ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JOB_NUMBER=00 JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/report-spans.yaml.template -o report-span-00-job.yaml] logger.go:42: 07:19:53 | es-rollover-autoprov/2-report-spans | running command: [sh -c kubectl apply -f report-span-00-job.yaml -n $NAMESPACE] logger.go:42: 07:19:53 | es-rollover-autoprov/2-report-spans | job.batch/00-report-span created logger.go:42: 07:20:20 | es-rollover-autoprov/2-report-spans | test step completed 2-report-spans logger.go:42: 07:20:20 | es-rollover-autoprov/3-check-indices | starting test step 3-check-indices logger.go:42: 07:20:20 | es-rollover-autoprov/3-check-indices | Job:kuttl-test-settling-wallaby/00-check-indices created logger.go:42: 07:20:25 | es-rollover-autoprov/3-check-indices | test step completed 3-check-indices logger.go:42: 07:20:25 | es-rollover-autoprov/4-check-indices | starting test step 4-check-indices logger.go:42: 07:20:25 | es-rollover-autoprov/4-check-indices | Job:kuttl-test-settling-wallaby/01-check-indices created logger.go:42: 07:20:29 | es-rollover-autoprov/4-check-indices | test step completed 4-check-indices logger.go:42: 07:20:29 | es-rollover-autoprov/5-install | starting test step 5-install logger.go:42: 07:20:29 | es-rollover-autoprov/5-install | Jaeger:kuttl-test-settling-wallaby/my-jaeger updated logger.go:42: 07:20:36 | es-rollover-autoprov/5-install | test step completed 5-install logger.go:42: 07:20:36 | es-rollover-autoprov/6-report-spans | starting test step 6-report-spans logger.go:42: 07:20:36 | es-rollover-autoprov/6-report-spans | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:20:45 | es-rollover-autoprov/6-report-spans | running command: [sh -c DAYS=2 ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JOB_NUMBER=02 JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/report-spans.yaml.template -o report-span-02-job.yaml] logger.go:42: 07:20:46 | es-rollover-autoprov/6-report-spans | running command: [sh -c kubectl apply -f report-span-02-job.yaml -n $NAMESPACE] logger.go:42: 07:20:46 | es-rollover-autoprov/6-report-spans | job.batch/02-report-span created logger.go:42: 07:21:11 | es-rollover-autoprov/6-report-spans | test step completed 6-report-spans logger.go:42: 07:21:11 | es-rollover-autoprov/7-check-indices | starting test step 7-check-indices logger.go:42: 07:21:11 | es-rollover-autoprov/7-check-indices | Job:kuttl-test-settling-wallaby/02-check-indices created logger.go:42: 07:21:16 | es-rollover-autoprov/7-check-indices | test step completed 7-check-indices logger.go:42: 07:21:16 | es-rollover-autoprov/8-check-indices | starting test step 8-check-indices logger.go:42: 07:21:16 | es-rollover-autoprov/8-check-indices | Job:kuttl-test-settling-wallaby/03-check-indices created logger.go:42: 07:21:21 | es-rollover-autoprov/8-check-indices | test step completed 8-check-indices logger.go:42: 07:21:21 | es-rollover-autoprov/9-check-indices | starting test step 9-check-indices logger.go:42: 07:21:21 | es-rollover-autoprov/9-check-indices | Job:kuttl-test-settling-wallaby/04-check-indices created logger.go:42: 07:21:26 | es-rollover-autoprov/9-check-indices | test step completed 9-check-indices logger.go:42: 07:21:26 | es-rollover-autoprov/10-report-spans | starting test step 10-report-spans logger.go:42: 07:21:26 | es-rollover-autoprov/10-report-spans | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:21:34 | es-rollover-autoprov/10-report-spans | running command: [sh -c DAYS=2 ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JOB_NUMBER=03 JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/report-spans.yaml.template -o report-span-03-job.yaml] logger.go:42: 07:21:34 | es-rollover-autoprov/10-report-spans | running command: [sh -c kubectl apply -f report-span-03-job.yaml -n $NAMESPACE] logger.go:42: 07:21:34 | es-rollover-autoprov/10-report-spans | job.batch/03-report-span created logger.go:42: 07:22:00 | es-rollover-autoprov/10-report-spans | test step completed 10-report-spans logger.go:42: 07:22:00 | es-rollover-autoprov/11-check-indices | starting test step 11-check-indices logger.go:42: 07:22:00 | es-rollover-autoprov/11-check-indices | running command: [sh -c go run ../../../../cmd-utils/wait-cronjob/main.go --cronjob my-jaeger-es-rollover --namespace $NAMESPACE] logger.go:42: 07:22:13 | es-rollover-autoprov/11-check-indices | time="2025-02-03T07:22:13Z" level=debug msg="Checking if the my-jaeger-es-rollover CronJob exists" logger.go:42: 07:22:13 | es-rollover-autoprov/11-check-indices | time="2025-02-03T07:22:13Z" level=debug msg="No BatchV1beta1/Cronjobs were found" logger.go:42: 07:22:13 | es-rollover-autoprov/11-check-indices | time="2025-02-03T07:22:13Z" level=info msg="Cronjob my-jaeger-es-rollover found successfully" logger.go:42: 07:22:13 | es-rollover-autoprov/11-check-indices | time="2025-02-03T07:22:13Z" level=debug msg="Waiting for the next scheduled job from my-jaeger-es-rollover cronjob" logger.go:42: 07:22:13 | es-rollover-autoprov/11-check-indices | time="2025-02-03T07:22:13Z" level=debug msg="Waiting for next job from my-jaeger-es-rollover to succeed" logger.go:42: 07:22:23 | es-rollover-autoprov/11-check-indices | time="2025-02-03T07:22:23Z" level=debug msg="Waiting for next job from my-jaeger-es-rollover to succeed" logger.go:42: 07:22:33 | es-rollover-autoprov/11-check-indices | time="2025-02-03T07:22:33Z" level=debug msg="Waiting for next job from my-jaeger-es-rollover to succeed" logger.go:42: 07:22:43 | es-rollover-autoprov/11-check-indices | time="2025-02-03T07:22:43Z" level=debug msg="Waiting for next job from my-jaeger-es-rollover to succeed" logger.go:42: 07:22:53 | es-rollover-autoprov/11-check-indices | time="2025-02-03T07:22:53Z" level=debug msg="Waiting for next job from my-jaeger-es-rollover to succeed" logger.go:42: 07:23:03 | es-rollover-autoprov/11-check-indices | time="2025-02-03T07:23:03Z" level=debug msg="Waiting for next job from my-jaeger-es-rollover to succeed" logger.go:42: 07:23:13 | es-rollover-autoprov/11-check-indices | time="2025-02-03T07:23:13Z" level=info msg="Job of owner my-jaeger-es-rollover succeeded after my-jaeger-es-rollover 1m0.01316968s" logger.go:42: 07:23:13 | es-rollover-autoprov/11-check-indices | Job:kuttl-test-settling-wallaby/05-check-indices created logger.go:42: 07:23:18 | es-rollover-autoprov/11-check-indices | test step completed 11-check-indices logger.go:42: 07:23:18 | es-rollover-autoprov/12-check-indices | starting test step 12-check-indices logger.go:42: 07:23:18 | es-rollover-autoprov/12-check-indices | Job:kuttl-test-settling-wallaby/06-check-indices created logger.go:42: 07:23:22 | es-rollover-autoprov/12-check-indices | test step completed 12-check-indices logger.go:42: 07:23:22 | es-rollover-autoprov | es-rollover-autoprov events from ns kuttl-test-settling-wallaby: logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:19:03 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsettlingwallabymyjaeger-1-64bf7bj78b Binding Scheduled Successfully assigned kuttl-test-settling-wallaby/elasticsearch-cdm-kuttltestsettlingwallabymyjaeger-1-64bf7bj78b to ip-10-0-6-172.us-east-2.compute.internal default-scheduler logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:19:03 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestsettlingwallabymyjaeger-1-64bf7f4cc5 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestsettlingwallabymyjaeger-1-64bf7bj78b replicaset-controller logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:19:03 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestsettlingwallabymyjaeger-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestsettlingwallabymyjaeger-1-64bf7f4cc5 to 1 deployment-controller logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:19:04 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsettlingwallabymyjaeger-1-64bf7bj78b AddedInterface Add eth0 [10.131.0.24/23] from ovn-kubernetes multus logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:19:04 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsettlingwallabymyjaeger-1-64bf7bj78b.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:410674bf2f77ff1d431f494cadfab3e81249b35695eb65aedfdb8498079b4d35" already present on machine kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:19:04 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsettlingwallabymyjaeger-1-64bf7bj78b.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:19:04 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsettlingwallabymyjaeger-1-64bf7bj78b.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:19:04 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsettlingwallabymyjaeger-1-64bf7bj78b.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:fe2439e5f9e14d828caace425214edfa3edc179d72d225c43f69e9e4c9a5837b" already present on machine kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:19:04 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsettlingwallabymyjaeger-1-64bf7bj78b.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:19:04 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsettlingwallabymyjaeger-1-64bf7bj78b.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:19:14 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestsettlingwallabymyjaeger-1-64bf7bj78b.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:19:19 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestsettlingwallabymyjaeger-1-64bf7bj78b.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:19:31 +0000 UTC Normal Pod my-jaeger-collector-56d6d6b78b-sv5v7 Binding Scheduled Successfully assigned kuttl-test-settling-wallaby/my-jaeger-collector-56d6d6b78b-sv5v7 to ip-10-0-26-111.us-east-2.compute.internal default-scheduler logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:19:31 +0000 UTC Normal Pod my-jaeger-collector-56d6d6b78b-sv5v7 AddedInterface Add eth0 [10.128.2.25/23] from ovn-kubernetes multus logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:19:31 +0000 UTC Normal Pod my-jaeger-collector-56d6d6b78b-sv5v7.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:40a1c7fa23aa5ffa64e8e22aa38022f5d4d7ff644c46a3da7169b713d486c3c1" already present on machine kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:19:31 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-56d6d6b78b SuccessfulCreate Created pod: my-jaeger-collector-56d6d6b78b-sv5v7 replicaset-controller logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:19:31 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-56d6d6b78b to 1 deployment-controller logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:19:31 +0000 UTC Normal Pod my-jaeger-query-765cd4568f-ljvcr Binding Scheduled Successfully assigned kuttl-test-settling-wallaby/my-jaeger-query-765cd4568f-ljvcr to ip-10-0-66-98.us-east-2.compute.internal default-scheduler logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:19:31 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-765cd4568f SuccessfulCreate Created pod: my-jaeger-query-765cd4568f-ljvcr replicaset-controller logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:19:31 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-765cd4568f to 1 deployment-controller logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:19:32 +0000 UTC Normal Pod my-jaeger-collector-56d6d6b78b-sv5v7.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:19:32 +0000 UTC Normal Pod my-jaeger-collector-56d6d6b78b-sv5v7.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:19:32 +0000 UTC Normal Pod my-jaeger-query-765cd4568f-ljvcr AddedInterface Add eth0 [10.129.2.21/23] from ovn-kubernetes multus logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:19:32 +0000 UTC Normal Pod my-jaeger-query-765cd4568f-ljvcr.spec.containers{jaeger-query} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:84020ce29bcb5a4bd018e6596188ae919c5cd600e08f78a546c0e76ea477685e" kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:19:34 +0000 UTC Normal Pod my-jaeger-query-765cd4568f-ljvcr.spec.containers{jaeger-query} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:84020ce29bcb5a4bd018e6596188ae919c5cd600e08f78a546c0e76ea477685e" in 2.657s (2.657s including waiting). Image size: 142020742 bytes. kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:19:34 +0000 UTC Normal Pod my-jaeger-query-765cd4568f-ljvcr.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:19:34 +0000 UTC Normal Pod my-jaeger-query-765cd4568f-ljvcr.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:19:34 +0000 UTC Normal Pod my-jaeger-query-765cd4568f-ljvcr.spec.containers{oauth-proxy} Pulling Pulling image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:19:38 +0000 UTC Normal Pod my-jaeger-query-765cd4568f-ljvcr.spec.containers{oauth-proxy} Pulled Successfully pulled image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" in 3.735s (3.735s including waiting). Image size: 438322369 bytes. kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:19:38 +0000 UTC Normal Pod my-jaeger-query-765cd4568f-ljvcr.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:19:38 +0000 UTC Normal Pod my-jaeger-query-765cd4568f-ljvcr.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:19:38 +0000 UTC Normal Pod my-jaeger-query-765cd4568f-ljvcr.spec.containers{jaeger-agent} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:65342c7d622f336741a8ae0b9dead79c7ecabc155d02cdd42b7b49ca36680e74" kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:19:40 +0000 UTC Normal Pod my-jaeger-query-765cd4568f-ljvcr.spec.containers{jaeger-agent} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:65342c7d622f336741a8ae0b9dead79c7ecabc155d02cdd42b7b49ca36680e74" in 1.803s (1.803s including waiting). Image size: 112614125 bytes. kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:19:40 +0000 UTC Normal Pod my-jaeger-query-765cd4568f-ljvcr.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:19:40 +0000 UTC Normal Pod my-jaeger-query-765cd4568f-ljvcr.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:19:49 +0000 UTC Normal Pod my-jaeger-query-765cd4568f-ljvcr.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:19:49 +0000 UTC Normal Pod my-jaeger-query-765cd4568f-ljvcr.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:19:49 +0000 UTC Normal Pod my-jaeger-query-765cd4568f-ljvcr.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:19:49 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-765cd4568f SuccessfulDelete Deleted pod: my-jaeger-query-765cd4568f-ljvcr replicaset-controller logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:19:49 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled down replica set my-jaeger-query-765cd4568f to 0 from 1 deployment-controller logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:19:50 +0000 UTC Normal Pod my-jaeger-query-56c47d9676-x9chj Binding Scheduled Successfully assigned kuttl-test-settling-wallaby/my-jaeger-query-56c47d9676-x9chj to ip-10-0-66-98.us-east-2.compute.internal default-scheduler logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:19:50 +0000 UTC Normal Pod my-jaeger-query-56c47d9676-x9chj AddedInterface Add eth0 [10.129.2.22/23] from ovn-kubernetes multus logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:19:50 +0000 UTC Normal Pod my-jaeger-query-56c47d9676-x9chj.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:84020ce29bcb5a4bd018e6596188ae919c5cd600e08f78a546c0e76ea477685e" already present on machine kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:19:50 +0000 UTC Normal Pod my-jaeger-query-56c47d9676-x9chj.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:19:50 +0000 UTC Normal Pod my-jaeger-query-56c47d9676-x9chj.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:19:50 +0000 UTC Normal Pod my-jaeger-query-56c47d9676-x9chj.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" already present on machine kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:19:50 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-56c47d9676 SuccessfulCreate Created pod: my-jaeger-query-56c47d9676-x9chj replicaset-controller logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:19:50 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-56c47d9676 to 1 deployment-controller logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:19:51 +0000 UTC Normal Pod my-jaeger-query-56c47d9676-x9chj.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:19:51 +0000 UTC Normal Pod my-jaeger-query-56c47d9676-x9chj.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:19:51 +0000 UTC Normal Pod my-jaeger-query-56c47d9676-x9chj.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:65342c7d622f336741a8ae0b9dead79c7ecabc155d02cdd42b7b49ca36680e74" already present on machine kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:19:51 +0000 UTC Normal Pod my-jaeger-query-56c47d9676-x9chj.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:19:51 +0000 UTC Normal Pod my-jaeger-query-56c47d9676-x9chj.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:19:53 +0000 UTC Normal Pod 00-report-span-pzf6h Binding Scheduled Successfully assigned kuttl-test-settling-wallaby/00-report-span-pzf6h to ip-10-0-26-111.us-east-2.compute.internal default-scheduler logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:19:53 +0000 UTC Normal Job.batch 00-report-span SuccessfulCreate Created pod: 00-report-span-pzf6h job-controller logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:19:54 +0000 UTC Normal Pod 00-report-span-pzf6h AddedInterface Add eth0 [10.128.2.26/23] from ovn-kubernetes multus logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:19:54 +0000 UTC Normal Pod 00-report-span-pzf6h.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:19:57 +0000 UTC Normal Pod 00-report-span-pzf6h.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 2.812s (2.812s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:19:57 +0000 UTC Normal Pod 00-report-span-pzf6h.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:19:57 +0000 UTC Normal Pod 00-report-span-pzf6h.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:19:57 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:19:57 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:19:57 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:20:12 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get cpu utilization: missing request for cpu in container jaeger-collector of Pod my-jaeger-collector-56d6d6b78b-sv5v7 horizontal-pod-autoscaler logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:20:12 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get memory utilization: missing request for memory in container jaeger-collector of Pod my-jaeger-collector-56d6d6b78b-sv5v7 horizontal-pod-autoscaler logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:20:12 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: missing request for cpu in container jaeger-collector of Pod my-jaeger-collector-56d6d6b78b-sv5v7 horizontal-pod-autoscaler logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:20:19 +0000 UTC Normal Job.batch 00-report-span Completed Job completed job-controller logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:20:20 +0000 UTC Normal Pod 00-check-indices-6hxbl Binding Scheduled Successfully assigned kuttl-test-settling-wallaby/00-check-indices-6hxbl to ip-10-0-26-111.us-east-2.compute.internal default-scheduler logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:20:20 +0000 UTC Normal Job.batch 00-check-indices SuccessfulCreate Created pod: 00-check-indices-6hxbl job-controller logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:20:21 +0000 UTC Normal Pod 00-check-indices-6hxbl AddedInterface Add eth0 [10.128.2.27/23] from ovn-kubernetes multus logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:20:21 +0000 UTC Normal Pod 00-check-indices-6hxbl.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:20:21 +0000 UTC Normal Pod 00-check-indices-6hxbl.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 612ms (612ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:20:21 +0000 UTC Normal Pod 00-check-indices-6hxbl.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:20:21 +0000 UTC Normal Pod 00-check-indices-6hxbl.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:20:25 +0000 UTC Normal Job.batch 00-check-indices Completed Job completed job-controller logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:20:25 +0000 UTC Normal Pod 01-check-indices-g5kg7 Binding Scheduled Successfully assigned kuttl-test-settling-wallaby/01-check-indices-g5kg7 to ip-10-0-26-111.us-east-2.compute.internal default-scheduler logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:20:25 +0000 UTC Normal Pod 01-check-indices-g5kg7 AddedInterface Add eth0 [10.128.2.28/23] from ovn-kubernetes multus logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:20:25 +0000 UTC Normal Pod 01-check-indices-g5kg7.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:20:25 +0000 UTC Normal Job.batch 01-check-indices SuccessfulCreate Created pod: 01-check-indices-g5kg7 job-controller logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:20:26 +0000 UTC Normal Pod 01-check-indices-g5kg7.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 533ms (533ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:20:26 +0000 UTC Normal Pod 01-check-indices-g5kg7.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:20:26 +0000 UTC Normal Pod 01-check-indices-g5kg7.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:20:29 +0000 UTC Normal Job.batch 01-check-indices Completed Job completed job-controller logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:20:30 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-jrbtx Binding Scheduled Successfully assigned kuttl-test-settling-wallaby/my-jaeger-es-rollover-create-mapping-jrbtx to ip-10-0-26-111.us-east-2.compute.internal default-scheduler logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:20:30 +0000 UTC Normal Job.batch my-jaeger-es-rollover-create-mapping SuccessfulCreate Created pod: my-jaeger-es-rollover-create-mapping-jrbtx job-controller logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:20:31 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-jrbtx AddedInterface Add eth0 [10.128.2.29/23] from ovn-kubernetes multus logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:20:31 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-jrbtx.spec.containers{my-jaeger-es-rollover-create-mapping} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:146ea9aab82c1e7dd871e269a58e0491439b4ce3b98c281e46214fb813cd8194" kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:20:32 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-jrbtx.spec.containers{my-jaeger-es-rollover-create-mapping} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:146ea9aab82c1e7dd871e269a58e0491439b4ce3b98c281e46214fb813cd8194" in 1.625s (1.625s including waiting). Image size: 104079089 bytes. kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:20:33 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-jrbtx.spec.containers{my-jaeger-es-rollover-create-mapping} Created Created container my-jaeger-es-rollover-create-mapping kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:20:33 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-jrbtx.spec.containers{my-jaeger-es-rollover-create-mapping} Started Started container my-jaeger-es-rollover-create-mapping kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:20:36 +0000 UTC Normal Pod my-jaeger-collector-56d6d6b78b-sv5v7.spec.containers{jaeger-collector} Killing Stopping container jaeger-collector kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:20:36 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-56d6d6b78b SuccessfulDelete Deleted pod: my-jaeger-collector-56d6d6b78b-sv5v7 replicaset-controller logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:20:36 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled down replica set my-jaeger-collector-56d6d6b78b to 0 from 1 deployment-controller logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:20:36 +0000 UTC Normal Job.batch my-jaeger-es-rollover-create-mapping Completed Job completed job-controller logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:20:36 +0000 UTC Normal Pod my-jaeger-query-56c47d9676-x9chj.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:20:36 +0000 UTC Normal Pod my-jaeger-query-56c47d9676-x9chj.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:20:36 +0000 UTC Normal Pod my-jaeger-query-56c47d9676-x9chj.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:20:36 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-56c47d9676 SuccessfulDelete Deleted pod: my-jaeger-query-56c47d9676-x9chj replicaset-controller logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:20:36 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled down replica set my-jaeger-query-56c47d9676 to 0 from 1 deployment-controller logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:20:37 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-54f8979cd7 SuccessfulCreate Created pod: my-jaeger-collector-54f8979cd7-zd67q replicaset-controller logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:20:37 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-54f8979cd7 to 1 deployment-controller logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:20:38 +0000 UTC Normal Pod my-jaeger-collector-54f8979cd7-zd67q Binding Scheduled Successfully assigned kuttl-test-settling-wallaby/my-jaeger-collector-54f8979cd7-zd67q to ip-10-0-26-111.us-east-2.compute.internal default-scheduler logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:20:38 +0000 UTC Normal Pod my-jaeger-collector-54f8979cd7-zd67q AddedInterface Add eth0 [10.128.2.30/23] from ovn-kubernetes multus logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:20:38 +0000 UTC Normal Pod my-jaeger-collector-54f8979cd7-zd67q.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:40a1c7fa23aa5ffa64e8e22aa38022f5d4d7ff644c46a3da7169b713d486c3c1" already present on machine kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:20:38 +0000 UTC Normal Pod my-jaeger-collector-54f8979cd7-zd67q.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:20:38 +0000 UTC Normal Pod my-jaeger-collector-54f8979cd7-zd67q.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:20:38 +0000 UTC Normal Pod my-jaeger-query-5956d576c7-l9wj5 Binding Scheduled Successfully assigned kuttl-test-settling-wallaby/my-jaeger-query-5956d576c7-l9wj5 to ip-10-0-66-98.us-east-2.compute.internal default-scheduler logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:20:38 +0000 UTC Normal Pod my-jaeger-query-5956d576c7-l9wj5 AddedInterface Add eth0 [10.129.2.23/23] from ovn-kubernetes multus logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:20:38 +0000 UTC Normal Pod my-jaeger-query-5956d576c7-l9wj5.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:84020ce29bcb5a4bd018e6596188ae919c5cd600e08f78a546c0e76ea477685e" already present on machine kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:20:38 +0000 UTC Normal Pod my-jaeger-query-5956d576c7-l9wj5.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:20:38 +0000 UTC Normal Pod my-jaeger-query-5956d576c7-l9wj5.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:20:38 +0000 UTC Normal Pod my-jaeger-query-5956d576c7-l9wj5.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" already present on machine kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:20:38 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-5956d576c7 SuccessfulCreate Created pod: my-jaeger-query-5956d576c7-l9wj5 replicaset-controller logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:20:38 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-5956d576c7 to 1 deployment-controller logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:20:39 +0000 UTC Normal Pod my-jaeger-query-5956d576c7-l9wj5.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:20:39 +0000 UTC Normal Pod my-jaeger-query-5956d576c7-l9wj5.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:20:39 +0000 UTC Normal Pod my-jaeger-query-5956d576c7-l9wj5.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:65342c7d622f336741a8ae0b9dead79c7ecabc155d02cdd42b7b49ca36680e74" already present on machine kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:20:39 +0000 UTC Normal Pod my-jaeger-query-5956d576c7-l9wj5.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:20:39 +0000 UTC Normal Pod my-jaeger-query-5956d576c7-l9wj5.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:20:46 +0000 UTC Normal Pod 02-report-span-xbbrr Binding Scheduled Successfully assigned kuttl-test-settling-wallaby/02-report-span-xbbrr to ip-10-0-26-111.us-east-2.compute.internal default-scheduler logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:20:46 +0000 UTC Normal Job.batch 02-report-span SuccessfulCreate Created pod: 02-report-span-xbbrr job-controller logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:20:47 +0000 UTC Normal Pod 02-report-span-xbbrr AddedInterface Add eth0 [10.128.2.31/23] from ovn-kubernetes multus logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:20:47 +0000 UTC Normal Pod 02-report-span-xbbrr.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:20:47 +0000 UTC Normal Pod 02-report-span-xbbrr.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 704ms (704ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:20:47 +0000 UTC Normal Pod 02-report-span-xbbrr.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:20:48 +0000 UTC Normal Pod 02-report-span-xbbrr.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:21:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28976121-jghnk Binding Scheduled Successfully assigned kuttl-test-settling-wallaby/my-jaeger-es-lookback-28976121-jghnk to ip-10-0-66-98.us-east-2.compute.internal default-scheduler logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:21:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28976121-jghnk AddedInterface Add eth0 [10.129.2.24/23] from ovn-kubernetes multus logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:21:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28976121-jghnk.spec.containers{my-jaeger-es-lookback} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:146ea9aab82c1e7dd871e269a58e0491439b4ce3b98c281e46214fb813cd8194" kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:21:00 +0000 UTC Normal Job.batch my-jaeger-es-lookback-28976121 SuccessfulCreate Created pod: my-jaeger-es-lookback-28976121-jghnk job-controller logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:21:00 +0000 UTC Normal CronJob.batch my-jaeger-es-lookback SuccessfulCreate Created job my-jaeger-es-lookback-28976121 cronjob-controller logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:21:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28976121-r2t8p Binding Scheduled Successfully assigned kuttl-test-settling-wallaby/my-jaeger-es-rollover-28976121-r2t8p to ip-10-0-26-111.us-east-2.compute.internal default-scheduler logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:21:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28976121-r2t8p AddedInterface Add eth0 [10.128.2.32/23] from ovn-kubernetes multus logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:21:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28976121-r2t8p.spec.containers{my-jaeger-es-rollover} Pulled Container image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:146ea9aab82c1e7dd871e269a58e0491439b4ce3b98c281e46214fb813cd8194" already present on machine kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:21:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28976121-r2t8p.spec.containers{my-jaeger-es-rollover} Created Created container my-jaeger-es-rollover kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:21:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28976121-r2t8p.spec.containers{my-jaeger-es-rollover} Started Started container my-jaeger-es-rollover kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:21:00 +0000 UTC Normal Job.batch my-jaeger-es-rollover-28976121 SuccessfulCreate Created pod: my-jaeger-es-rollover-28976121-r2t8p job-controller logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:21:00 +0000 UTC Normal CronJob.batch my-jaeger-es-rollover SuccessfulCreate Created job my-jaeger-es-rollover-28976121 cronjob-controller logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:21:01 +0000 UTC Normal Pod my-jaeger-es-lookback-28976121-jghnk.spec.containers{my-jaeger-es-lookback} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:146ea9aab82c1e7dd871e269a58e0491439b4ce3b98c281e46214fb813cd8194" in 1.082s (1.082s including waiting). Image size: 104079089 bytes. kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:21:02 +0000 UTC Normal Pod my-jaeger-es-lookback-28976121-jghnk.spec.containers{my-jaeger-es-lookback} Created Created container my-jaeger-es-lookback kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:21:02 +0000 UTC Normal Pod my-jaeger-es-lookback-28976121-jghnk.spec.containers{my-jaeger-es-lookback} Started Started container my-jaeger-es-lookback kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:21:04 +0000 UTC Normal Job.batch my-jaeger-es-rollover-28976121 Completed Job completed job-controller logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:21:04 +0000 UTC Normal CronJob.batch my-jaeger-es-rollover SawCompletedJob Saw completed job: my-jaeger-es-rollover-28976121, condition: Complete cronjob-controller logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:21:05 +0000 UTC Normal Job.batch my-jaeger-es-lookback-28976121 Completed Job completed job-controller logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:21:05 +0000 UTC Normal CronJob.batch my-jaeger-es-lookback SawCompletedJob Saw completed job: my-jaeger-es-lookback-28976121, condition: Complete cronjob-controller logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:21:11 +0000 UTC Normal Pod 02-check-indices-4qjvg Binding Scheduled Successfully assigned kuttl-test-settling-wallaby/02-check-indices-4qjvg to ip-10-0-26-111.us-east-2.compute.internal default-scheduler logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:21:11 +0000 UTC Normal Job.batch 02-check-indices SuccessfulCreate Created pod: 02-check-indices-4qjvg job-controller logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:21:11 +0000 UTC Normal Job.batch 02-report-span Completed Job completed job-controller logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:21:12 +0000 UTC Normal Pod 02-check-indices-4qjvg AddedInterface Add eth0 [10.128.2.33/23] from ovn-kubernetes multus logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:21:12 +0000 UTC Normal Pod 02-check-indices-4qjvg.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:21:13 +0000 UTC Normal Pod 02-check-indices-4qjvg.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 1.313s (1.313s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:21:13 +0000 UTC Normal Pod 02-check-indices-4qjvg.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:21:13 +0000 UTC Normal Pod 02-check-indices-4qjvg.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:21:16 +0000 UTC Normal Job.batch 02-check-indices Completed Job completed job-controller logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:21:16 +0000 UTC Normal Pod 03-check-indices-dgwqb Binding Scheduled Successfully assigned kuttl-test-settling-wallaby/03-check-indices-dgwqb to ip-10-0-26-111.us-east-2.compute.internal default-scheduler logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:21:16 +0000 UTC Normal Job.batch 03-check-indices SuccessfulCreate Created pod: 03-check-indices-dgwqb job-controller logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:21:17 +0000 UTC Normal Pod 03-check-indices-dgwqb AddedInterface Add eth0 [10.128.2.34/23] from ovn-kubernetes multus logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:21:17 +0000 UTC Normal Pod 03-check-indices-dgwqb.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:21:17 +0000 UTC Normal Pod 03-check-indices-dgwqb.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 515ms (515ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:21:18 +0000 UTC Normal Pod 03-check-indices-dgwqb.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:21:18 +0000 UTC Normal Pod 03-check-indices-dgwqb.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:21:21 +0000 UTC Normal Job.batch 03-check-indices Completed Job completed job-controller logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:21:21 +0000 UTC Normal Pod 04-check-indices-z5hjg Binding Scheduled Successfully assigned kuttl-test-settling-wallaby/04-check-indices-z5hjg to ip-10-0-26-111.us-east-2.compute.internal default-scheduler logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:21:21 +0000 UTC Normal Job.batch 04-check-indices SuccessfulCreate Created pod: 04-check-indices-z5hjg job-controller logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:21:22 +0000 UTC Normal Pod 04-check-indices-z5hjg AddedInterface Add eth0 [10.128.2.35/23] from ovn-kubernetes multus logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:21:22 +0000 UTC Normal Pod 04-check-indices-z5hjg.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:21:23 +0000 UTC Normal Pod 04-check-indices-z5hjg.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 549ms (549ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:21:23 +0000 UTC Normal Pod 04-check-indices-z5hjg.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:21:23 +0000 UTC Normal Pod 04-check-indices-z5hjg.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:21:26 +0000 UTC Normal Job.batch 04-check-indices Completed Job completed job-controller logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:21:27 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get cpu utilization: missing request for cpu in container jaeger-collector of Pod my-jaeger-collector-54f8979cd7-zd67q horizontal-pod-autoscaler logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:21:27 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get memory utilization: missing request for memory in container jaeger-collector of Pod my-jaeger-collector-54f8979cd7-zd67q horizontal-pod-autoscaler logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:21:27 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: missing request for cpu in container jaeger-collector of Pod my-jaeger-collector-54f8979cd7-zd67q horizontal-pod-autoscaler logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:21:35 +0000 UTC Normal Pod 03-report-span-rvn6j Binding Scheduled Successfully assigned kuttl-test-settling-wallaby/03-report-span-rvn6j to ip-10-0-26-111.us-east-2.compute.internal default-scheduler logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:21:35 +0000 UTC Normal Pod 03-report-span-rvn6j AddedInterface Add eth0 [10.128.2.36/23] from ovn-kubernetes multus logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:21:35 +0000 UTC Normal Pod 03-report-span-rvn6j.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:21:35 +0000 UTC Normal Job.batch 03-report-span SuccessfulCreate Created pod: 03-report-span-rvn6j job-controller logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:21:36 +0000 UTC Normal Pod 03-report-span-rvn6j.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 572ms (572ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:21:36 +0000 UTC Normal Pod 03-report-span-rvn6j.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:21:36 +0000 UTC Normal Pod 03-report-span-rvn6j.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:21:59 +0000 UTC Normal Job.batch 03-report-span Completed Job completed job-controller logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:22:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28976122-47npx Binding Scheduled Successfully assigned kuttl-test-settling-wallaby/my-jaeger-es-lookback-28976122-47npx to ip-10-0-26-111.us-east-2.compute.internal default-scheduler logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:22:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28976122-47npx AddedInterface Add eth0 [10.128.2.37/23] from ovn-kubernetes multus logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:22:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28976122-47npx.spec.containers{my-jaeger-es-lookback} Pulled Container image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:146ea9aab82c1e7dd871e269a58e0491439b4ce3b98c281e46214fb813cd8194" already present on machine kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:22:00 +0000 UTC Normal Job.batch my-jaeger-es-lookback-28976122 SuccessfulCreate Created pod: my-jaeger-es-lookback-28976122-47npx job-controller logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:22:00 +0000 UTC Normal CronJob.batch my-jaeger-es-lookback SuccessfulCreate Created job my-jaeger-es-lookback-28976122 cronjob-controller logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:22:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28976122-n6w48 Binding Scheduled Successfully assigned kuttl-test-settling-wallaby/my-jaeger-es-rollover-28976122-n6w48 to ip-10-0-26-111.us-east-2.compute.internal default-scheduler logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:22:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28976122-n6w48 AddedInterface Add eth0 [10.128.2.38/23] from ovn-kubernetes multus logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:22:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28976122-n6w48.spec.containers{my-jaeger-es-rollover} Pulled Container image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:146ea9aab82c1e7dd871e269a58e0491439b4ce3b98c281e46214fb813cd8194" already present on machine kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:22:00 +0000 UTC Normal Job.batch my-jaeger-es-rollover-28976122 SuccessfulCreate Created pod: my-jaeger-es-rollover-28976122-n6w48 job-controller logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:22:00 +0000 UTC Normal CronJob.batch my-jaeger-es-rollover SuccessfulCreate Created job my-jaeger-es-rollover-28976122 cronjob-controller logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:22:01 +0000 UTC Normal Pod my-jaeger-es-lookback-28976122-47npx.spec.containers{my-jaeger-es-lookback} Created Created container my-jaeger-es-lookback kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:22:01 +0000 UTC Normal Pod my-jaeger-es-lookback-28976122-47npx.spec.containers{my-jaeger-es-lookback} Started Started container my-jaeger-es-lookback kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:22:01 +0000 UTC Normal Pod my-jaeger-es-rollover-28976122-n6w48.spec.containers{my-jaeger-es-rollover} Created Created container my-jaeger-es-rollover kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:22:01 +0000 UTC Normal Pod my-jaeger-es-rollover-28976122-n6w48.spec.containers{my-jaeger-es-rollover} Started Started container my-jaeger-es-rollover kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:22:04 +0000 UTC Normal Job.batch my-jaeger-es-lookback-28976122 Completed Job completed job-controller logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:22:04 +0000 UTC Normal CronJob.batch my-jaeger-es-lookback SawCompletedJob Saw completed job: my-jaeger-es-lookback-28976122, condition: Complete cronjob-controller logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:22:04 +0000 UTC Normal Job.batch my-jaeger-es-rollover-28976122 Completed Job completed job-controller logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:22:04 +0000 UTC Normal CronJob.batch my-jaeger-es-rollover SawCompletedJob Saw completed job: my-jaeger-es-rollover-28976122, condition: Complete cronjob-controller logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:23:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28976123-7fzqq Binding Scheduled Successfully assigned kuttl-test-settling-wallaby/my-jaeger-es-lookback-28976123-7fzqq to ip-10-0-26-111.us-east-2.compute.internal default-scheduler logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:23:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28976123-7fzqq AddedInterface Add eth0 [10.128.2.40/23] from ovn-kubernetes multus logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:23:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28976123-7fzqq.spec.containers{my-jaeger-es-lookback} Pulled Container image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:146ea9aab82c1e7dd871e269a58e0491439b4ce3b98c281e46214fb813cd8194" already present on machine kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:23:00 +0000 UTC Normal Job.batch my-jaeger-es-lookback-28976123 SuccessfulCreate Created pod: my-jaeger-es-lookback-28976123-7fzqq job-controller logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:23:00 +0000 UTC Normal CronJob.batch my-jaeger-es-lookback SuccessfulCreate Created job my-jaeger-es-lookback-28976123 cronjob-controller logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:23:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28976123-bs8wb Binding Scheduled Successfully assigned kuttl-test-settling-wallaby/my-jaeger-es-rollover-28976123-bs8wb to ip-10-0-26-111.us-east-2.compute.internal default-scheduler logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:23:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28976123-bs8wb AddedInterface Add eth0 [10.128.2.39/23] from ovn-kubernetes multus logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:23:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28976123-bs8wb.spec.containers{my-jaeger-es-rollover} Pulled Container image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:146ea9aab82c1e7dd871e269a58e0491439b4ce3b98c281e46214fb813cd8194" already present on machine kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:23:00 +0000 UTC Normal Job.batch my-jaeger-es-rollover-28976123 SuccessfulCreate Created pod: my-jaeger-es-rollover-28976123-bs8wb job-controller logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:23:00 +0000 UTC Normal CronJob.batch my-jaeger-es-rollover SuccessfulCreate Created job my-jaeger-es-rollover-28976123 cronjob-controller logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:23:01 +0000 UTC Normal Pod my-jaeger-es-lookback-28976123-7fzqq.spec.containers{my-jaeger-es-lookback} Created Created container my-jaeger-es-lookback kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:23:01 +0000 UTC Normal Pod my-jaeger-es-lookback-28976123-7fzqq.spec.containers{my-jaeger-es-lookback} Started Started container my-jaeger-es-lookback kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:23:01 +0000 UTC Normal Pod my-jaeger-es-rollover-28976123-bs8wb.spec.containers{my-jaeger-es-rollover} Created Created container my-jaeger-es-rollover kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:23:01 +0000 UTC Normal Pod my-jaeger-es-rollover-28976123-bs8wb.spec.containers{my-jaeger-es-rollover} Started Started container my-jaeger-es-rollover kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:23:03 +0000 UTC Normal Job.batch my-jaeger-es-lookback-28976123 Completed Job completed job-controller logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:23:03 +0000 UTC Normal CronJob.batch my-jaeger-es-lookback SawCompletedJob Saw completed job: my-jaeger-es-lookback-28976123, condition: Complete cronjob-controller logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:23:03 +0000 UTC Normal Job.batch my-jaeger-es-rollover-28976123 Completed Job completed job-controller logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:23:03 +0000 UTC Normal CronJob.batch my-jaeger-es-rollover SawCompletedJob Saw completed job: my-jaeger-es-rollover-28976123, condition: Complete cronjob-controller logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:23:13 +0000 UTC Normal Pod 05-check-indices-9hbtz Binding Scheduled Successfully assigned kuttl-test-settling-wallaby/05-check-indices-9hbtz to ip-10-0-26-111.us-east-2.compute.internal default-scheduler logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:23:13 +0000 UTC Normal Job.batch 05-check-indices SuccessfulCreate Created pod: 05-check-indices-9hbtz job-controller logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:23:14 +0000 UTC Normal Pod 05-check-indices-9hbtz AddedInterface Add eth0 [10.128.2.41/23] from ovn-kubernetes multus logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:23:14 +0000 UTC Normal Pod 05-check-indices-9hbtz.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:23:15 +0000 UTC Normal Pod 05-check-indices-9hbtz.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 832ms (832ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:23:15 +0000 UTC Normal Pod 05-check-indices-9hbtz.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:23:15 +0000 UTC Normal Pod 05-check-indices-9hbtz.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:23:18 +0000 UTC Normal Job.batch 05-check-indices Completed Job completed job-controller logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:23:18 +0000 UTC Normal Pod 06-check-indices-zrprk Binding Scheduled Successfully assigned kuttl-test-settling-wallaby/06-check-indices-zrprk to ip-10-0-26-111.us-east-2.compute.internal default-scheduler logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:23:18 +0000 UTC Normal Job.batch 06-check-indices SuccessfulCreate Created pod: 06-check-indices-zrprk job-controller logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:23:19 +0000 UTC Normal Pod 06-check-indices-zrprk AddedInterface Add eth0 [10.128.2.42/23] from ovn-kubernetes multus logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:23:19 +0000 UTC Normal Pod 06-check-indices-zrprk.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:23:20 +0000 UTC Normal Pod 06-check-indices-zrprk.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 695ms (695ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:23:20 +0000 UTC Normal Pod 06-check-indices-zrprk.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:23:20 +0000 UTC Normal Pod 06-check-indices-zrprk.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:23:22 | es-rollover-autoprov | 2025-02-03 07:23:22 +0000 UTC Normal Job.batch 06-check-indices Completed Job completed job-controller logger.go:42: 07:23:22 | es-rollover-autoprov | Deleting namespace: kuttl-test-settling-wallaby === CONT kuttl/harness/es-increasing-replicas logger.go:42: 07:23:31 | es-increasing-replicas | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:23:31 | es-increasing-replicas | Ignoring check-es-nodes.sh as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:23:31 | es-increasing-replicas | Ignoring openshift-check-es-nodes.yaml.template as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:23:31 | es-increasing-replicas | Creating namespace: kuttl-test-polite-bull logger.go:42: 07:23:31 | es-increasing-replicas/1-install | starting test step 1-install logger.go:42: 07:23:31 | es-increasing-replicas/1-install | Jaeger:kuttl-test-polite-bull/simple-prod created logger.go:42: 07:24:26 | es-increasing-replicas/1-install | test step completed 1-install logger.go:42: 07:24:26 | es-increasing-replicas/2-install | starting test step 2-install logger.go:42: 07:24:26 | es-increasing-replicas/2-install | Jaeger:kuttl-test-polite-bull/simple-prod updated logger.go:42: 07:24:32 | es-increasing-replicas/2-install | test step completed 2-install logger.go:42: 07:24:32 | es-increasing-replicas/3-smoke-test | starting test step 3-smoke-test logger.go:42: 07:24:32 | es-increasing-replicas/3-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE simple-prod /dev/null] logger.go:42: 07:24:33 | es-increasing-replicas/3-smoke-test | Warning: resource jaegers/simple-prod is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:24:39 | es-increasing-replicas/3-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:24:40 | es-increasing-replicas/3-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:24:40 | es-increasing-replicas/3-smoke-test | job.batch/report-span created logger.go:42: 07:24:40 | es-increasing-replicas/3-smoke-test | job.batch/check-span created logger.go:42: 07:24:54 | es-increasing-replicas/3-smoke-test | test step completed 3-smoke-test logger.go:42: 07:24:54 | es-increasing-replicas/4-install | starting test step 4-install logger.go:42: 07:24:54 | es-increasing-replicas/4-install | Jaeger:kuttl-test-polite-bull/simple-prod updated logger.go:42: 07:24:54 | es-increasing-replicas/4-install | test step completed 4-install logger.go:42: 07:24:54 | es-increasing-replicas/5-check-es-nodes | starting test step 5-check-es-nodes logger.go:42: 07:24:54 | es-increasing-replicas/5-check-es-nodes | running command: [sh -c ./check-es-nodes.sh $NAMESPACE] logger.go:42: 07:24:54 | es-increasing-replicas/5-check-es-nodes | Checking if the number of ES instances is the expected logger.go:42: 07:24:54 | es-increasing-replicas/5-check-es-nodes | false logger.go:42: 07:24:54 | es-increasing-replicas/5-check-es-nodes | Error: no matches found logger.go:42: 07:24:59 | es-increasing-replicas/5-check-es-nodes | Checking if the number of ES instances is the expected logger.go:42: 07:24:59 | es-increasing-replicas/5-check-es-nodes | true logger.go:42: 07:24:59 | es-increasing-replicas/5-check-es-nodes | test step completed 5-check-es-nodes logger.go:42: 07:24:59 | es-increasing-replicas | es-increasing-replicas events from ns kuttl-test-polite-bull: logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:23:56 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestpolitebullsimpleprod-1-5fcf64496c SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestpolitebullsimpleprod-1-5fcf6449b8mmv replicaset-controller logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:23:56 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpolitebullsimpleprod-1-5fcf6449b8mmv Binding Scheduled Successfully assigned kuttl-test-polite-bull/elasticsearch-cdm-kuttltestpolitebullsimpleprod-1-5fcf6449b8mmv to ip-10-0-6-172.us-east-2.compute.internal default-scheduler logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:23:56 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestpolitebullsimpleprod-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestpolitebullsimpleprod-1-5fcf64496c to 1 deployment-controller logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:23:57 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpolitebullsimpleprod-1-5fcf6449b8mmv AddedInterface Add eth0 [10.131.0.25/23] from ovn-kubernetes multus logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:23:57 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpolitebullsimpleprod-1-5fcf6449b8mmv.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:410674bf2f77ff1d431f494cadfab3e81249b35695eb65aedfdb8498079b4d35" already present on machine kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:23:57 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpolitebullsimpleprod-1-5fcf6449b8mmv.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:23:57 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpolitebullsimpleprod-1-5fcf6449b8mmv.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:23:57 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpolitebullsimpleprod-1-5fcf6449b8mmv.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:fe2439e5f9e14d828caace425214edfa3edc179d72d225c43f69e9e4c9a5837b" already present on machine kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:23:57 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpolitebullsimpleprod-1-5fcf6449b8mmv.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:23:57 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpolitebullsimpleprod-1-5fcf6449b8mmv.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:07 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestpolitebullsimpleprod-1-5fcf6449b8mmv.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:12 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestpolitebullsimpleprod-1-5fcf6449b8mmv.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:23 +0000 UTC Normal Pod simple-prod-collector-5b5cf99c8-k5n57 Binding Scheduled Successfully assigned kuttl-test-polite-bull/simple-prod-collector-5b5cf99c8-k5n57 to ip-10-0-26-111.us-east-2.compute.internal default-scheduler logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:23 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-5b5cf99c8 SuccessfulCreate Created pod: simple-prod-collector-5b5cf99c8-k5n57 replicaset-controller logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:23 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-5b5cf99c8 to 1 deployment-controller logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:23 +0000 UTC Normal Pod simple-prod-query-68ccc6bc7f-jt8b2 Binding Scheduled Successfully assigned kuttl-test-polite-bull/simple-prod-query-68ccc6bc7f-jt8b2 to ip-10-0-66-98.us-east-2.compute.internal default-scheduler logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:23 +0000 UTC Normal ReplicaSet.apps simple-prod-query-68ccc6bc7f SuccessfulCreate Created pod: simple-prod-query-68ccc6bc7f-jt8b2 replicaset-controller logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:23 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-68ccc6bc7f to 1 deployment-controller logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:24 +0000 UTC Normal Pod simple-prod-collector-5b5cf99c8-k5n57 AddedInterface Add eth0 [10.128.2.43/23] from ovn-kubernetes multus logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:24 +0000 UTC Normal Pod simple-prod-collector-5b5cf99c8-k5n57.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:40a1c7fa23aa5ffa64e8e22aa38022f5d4d7ff644c46a3da7169b713d486c3c1" already present on machine kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:24 +0000 UTC Normal Pod simple-prod-collector-5b5cf99c8-k5n57.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:24 +0000 UTC Normal Pod simple-prod-collector-5b5cf99c8-k5n57.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:24 +0000 UTC Normal Pod simple-prod-query-68ccc6bc7f-jt8b2 AddedInterface Add eth0 [10.129.2.25/23] from ovn-kubernetes multus logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:24 +0000 UTC Normal Pod simple-prod-query-68ccc6bc7f-jt8b2.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:84020ce29bcb5a4bd018e6596188ae919c5cd600e08f78a546c0e76ea477685e" already present on machine kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:24 +0000 UTC Normal Pod simple-prod-query-68ccc6bc7f-jt8b2.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:24 +0000 UTC Normal Pod simple-prod-query-68ccc6bc7f-jt8b2.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:24 +0000 UTC Normal Pod simple-prod-query-68ccc6bc7f-jt8b2.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" already present on machine kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:24 +0000 UTC Normal Pod simple-prod-query-68ccc6bc7f-jt8b2.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:24 +0000 UTC Normal Pod simple-prod-query-68ccc6bc7f-jt8b2.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:24 +0000 UTC Normal Pod simple-prod-query-68ccc6bc7f-jt8b2.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:65342c7d622f336741a8ae0b9dead79c7ecabc155d02cdd42b7b49ca36680e74" already present on machine kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:24 +0000 UTC Normal Pod simple-prod-query-68ccc6bc7f-jt8b2.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:24 +0000 UTC Normal Pod simple-prod-query-68ccc6bc7f-jt8b2.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:27 +0000 UTC Normal Pod simple-prod-collector-5b5cf99c8-9hmlv Binding Scheduled Successfully assigned kuttl-test-polite-bull/simple-prod-collector-5b5cf99c8-9hmlv to ip-10-0-66-98.us-east-2.compute.internal default-scheduler logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:27 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-5b5cf99c8 SuccessfulCreate Created pod: simple-prod-collector-5b5cf99c8-9hmlv replicaset-controller logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:27 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-5b5cf99c8 to 2 from 1 deployment-controller logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:27 +0000 UTC Normal Pod simple-prod-query-68ccc6bc7f-l8ndq Binding Scheduled Successfully assigned kuttl-test-polite-bull/simple-prod-query-68ccc6bc7f-l8ndq to ip-10-0-26-111.us-east-2.compute.internal default-scheduler logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:27 +0000 UTC Normal ReplicaSet.apps simple-prod-query-68ccc6bc7f SuccessfulCreate Created pod: simple-prod-query-68ccc6bc7f-l8ndq replicaset-controller logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:27 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-68ccc6bc7f to 2 from 1 deployment-controller logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:28 +0000 UTC Normal Pod simple-prod-collector-5b5cf99c8-9hmlv AddedInterface Add eth0 [10.129.2.26/23] from ovn-kubernetes multus logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:28 +0000 UTC Normal Pod simple-prod-collector-5b5cf99c8-9hmlv.spec.containers{jaeger-collector} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:40a1c7fa23aa5ffa64e8e22aa38022f5d4d7ff644c46a3da7169b713d486c3c1" kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:28 +0000 UTC Normal Pod simple-prod-query-68ccc6bc7f-l8ndq AddedInterface Add eth0 [10.128.2.44/23] from ovn-kubernetes multus logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:28 +0000 UTC Normal Pod simple-prod-query-68ccc6bc7f-l8ndq.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:84020ce29bcb5a4bd018e6596188ae919c5cd600e08f78a546c0e76ea477685e" already present on machine kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:28 +0000 UTC Normal Pod simple-prod-query-68ccc6bc7f-l8ndq.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:28 +0000 UTC Normal Pod simple-prod-query-68ccc6bc7f-l8ndq.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:28 +0000 UTC Normal Pod simple-prod-query-68ccc6bc7f-l8ndq.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" already present on machine kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:28 +0000 UTC Normal Pod simple-prod-query-68ccc6bc7f-l8ndq.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:28 +0000 UTC Normal Pod simple-prod-query-68ccc6bc7f-l8ndq.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:28 +0000 UTC Normal Pod simple-prod-query-68ccc6bc7f-l8ndq.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:65342c7d622f336741a8ae0b9dead79c7ecabc155d02cdd42b7b49ca36680e74" already present on machine kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:29 +0000 UTC Normal Pod simple-prod-query-68ccc6bc7f-l8ndq.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:29 +0000 UTC Normal Pod simple-prod-query-68ccc6bc7f-l8ndq.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:30 +0000 UTC Normal Pod simple-prod-collector-5b5cf99c8-9hmlv.spec.containers{jaeger-collector} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:40a1c7fa23aa5ffa64e8e22aa38022f5d4d7ff644c46a3da7169b713d486c3c1" in 2.15s (2.15s including waiting). Image size: 139779827 bytes. kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:30 +0000 UTC Normal Pod simple-prod-collector-5b5cf99c8-9hmlv.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:30 +0000 UTC Normal Pod simple-prod-collector-5b5cf99c8-9hmlv.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:35 +0000 UTC Normal Pod simple-prod-query-68ccc6bc7f-jt8b2.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:35 +0000 UTC Normal Pod simple-prod-query-68ccc6bc7f-jt8b2.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:35 +0000 UTC Normal Pod simple-prod-query-68ccc6bc7f-jt8b2.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:35 +0000 UTC Normal Pod simple-prod-query-68ccc6bc7f-l8ndq.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:35 +0000 UTC Normal Pod simple-prod-query-68ccc6bc7f-l8ndq.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:35 +0000 UTC Normal Pod simple-prod-query-68ccc6bc7f-l8ndq.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:35 +0000 UTC Normal ReplicaSet.apps simple-prod-query-68ccc6bc7f SuccessfulDelete Deleted pod: simple-prod-query-68ccc6bc7f-l8ndq replicaset-controller logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:35 +0000 UTC Normal ReplicaSet.apps simple-prod-query-68ccc6bc7f SuccessfulDelete Deleted pod: simple-prod-query-68ccc6bc7f-jt8b2 replicaset-controller logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:35 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled down replica set simple-prod-query-68ccc6bc7f to 0 from 2 deployment-controller logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:36 +0000 UTC Normal Pod simple-prod-query-bc57b7f66-kpshl Binding Scheduled Successfully assigned kuttl-test-polite-bull/simple-prod-query-bc57b7f66-kpshl to ip-10-0-66-98.us-east-2.compute.internal default-scheduler logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:36 +0000 UTC Normal Pod simple-prod-query-bc57b7f66-krxfp Binding Scheduled Successfully assigned kuttl-test-polite-bull/simple-prod-query-bc57b7f66-krxfp to ip-10-0-26-111.us-east-2.compute.internal default-scheduler logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:36 +0000 UTC Normal ReplicaSet.apps simple-prod-query-bc57b7f66 SuccessfulCreate Created pod: simple-prod-query-bc57b7f66-kpshl replicaset-controller logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:36 +0000 UTC Normal ReplicaSet.apps simple-prod-query-bc57b7f66 SuccessfulCreate Created pod: simple-prod-query-bc57b7f66-krxfp replicaset-controller logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:36 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-bc57b7f66 to 2 deployment-controller logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:37 +0000 UTC Normal Pod simple-prod-query-bc57b7f66-kpshl AddedInterface Add eth0 [10.129.2.27/23] from ovn-kubernetes multus logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:37 +0000 UTC Normal Pod simple-prod-query-bc57b7f66-kpshl.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:84020ce29bcb5a4bd018e6596188ae919c5cd600e08f78a546c0e76ea477685e" already present on machine kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:37 +0000 UTC Normal Pod simple-prod-query-bc57b7f66-kpshl.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:37 +0000 UTC Normal Pod simple-prod-query-bc57b7f66-kpshl.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:37 +0000 UTC Normal Pod simple-prod-query-bc57b7f66-kpshl.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" already present on machine kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:37 +0000 UTC Normal Pod simple-prod-query-bc57b7f66-kpshl.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:37 +0000 UTC Normal Pod simple-prod-query-bc57b7f66-kpshl.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:37 +0000 UTC Normal Pod simple-prod-query-bc57b7f66-kpshl.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:65342c7d622f336741a8ae0b9dead79c7ecabc155d02cdd42b7b49ca36680e74" already present on machine kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:37 +0000 UTC Normal Pod simple-prod-query-bc57b7f66-kpshl.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:37 +0000 UTC Normal Pod simple-prod-query-bc57b7f66-kpshl.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:37 +0000 UTC Normal Pod simple-prod-query-bc57b7f66-krxfp AddedInterface Add eth0 [10.128.2.45/23] from ovn-kubernetes multus logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:37 +0000 UTC Normal Pod simple-prod-query-bc57b7f66-krxfp.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:84020ce29bcb5a4bd018e6596188ae919c5cd600e08f78a546c0e76ea477685e" already present on machine kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:37 +0000 UTC Normal Pod simple-prod-query-bc57b7f66-krxfp.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:37 +0000 UTC Normal Pod simple-prod-query-bc57b7f66-krxfp.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:37 +0000 UTC Normal Pod simple-prod-query-bc57b7f66-krxfp.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" already present on machine kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:37 +0000 UTC Normal Pod simple-prod-query-bc57b7f66-krxfp.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:37 +0000 UTC Normal Pod simple-prod-query-bc57b7f66-krxfp.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:37 +0000 UTC Normal Pod simple-prod-query-bc57b7f66-krxfp.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:65342c7d622f336741a8ae0b9dead79c7ecabc155d02cdd42b7b49ca36680e74" already present on machine kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:37 +0000 UTC Normal Pod simple-prod-query-bc57b7f66-krxfp.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:37 +0000 UTC Normal Pod simple-prod-query-bc57b7f66-krxfp.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:40 +0000 UTC Normal Pod check-span-q5xg4 Binding Scheduled Successfully assigned kuttl-test-polite-bull/check-span-q5xg4 to ip-10-0-66-98.us-east-2.compute.internal default-scheduler logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:40 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-q5xg4 job-controller logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:40 +0000 UTC Normal Pod report-span-js57d Binding Scheduled Successfully assigned kuttl-test-polite-bull/report-span-js57d to ip-10-0-66-98.us-east-2.compute.internal default-scheduler logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:40 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-js57d job-controller logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:41 +0000 UTC Normal Pod check-span-q5xg4 AddedInterface Add eth0 [10.129.2.29/23] from ovn-kubernetes multus logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:41 +0000 UTC Normal Pod check-span-q5xg4.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:41 +0000 UTC Normal Pod report-span-js57d AddedInterface Add eth0 [10.129.2.28/23] from ovn-kubernetes multus logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:41 +0000 UTC Normal Pod report-span-js57d.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:43 +0000 UTC Normal Pod check-span-q5xg4.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 2.003s (2.003s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:43 +0000 UTC Normal Pod check-span-q5xg4.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:43 +0000 UTC Normal Pod check-span-q5xg4.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:43 +0000 UTC Normal Pod report-span-js57d.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 2.044s (2.044s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:43 +0000 UTC Normal Pod report-span-js57d.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:43 +0000 UTC Normal Pod report-span-js57d.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:53 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:56 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestpolitebullsimpleprod-2-6d645445fc SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestpolitebullsimpleprod-2-6d645445hmpx2 replicaset-controller logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:56 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpolitebullsimpleprod-2-6d645445hmpx2 Binding Scheduled Successfully assigned kuttl-test-polite-bull/elasticsearch-cdm-kuttltestpolitebullsimpleprod-2-6d645445hmpx2 to ip-10-0-66-98.us-east-2.compute.internal default-scheduler logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:56 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestpolitebullsimpleprod-2 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestpolitebullsimpleprod-2-6d645445fc to 1 deployment-controller logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:56 +0000 UTC Normal Pod simple-prod-collector-5b5cf99c8-9hmlv.spec.containers{jaeger-collector} Killing Stopping container jaeger-collector kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:56 +0000 UTC Normal Pod simple-prod-collector-5b5cf99c8-k5n57.spec.containers{jaeger-collector} Killing Stopping container jaeger-collector kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:56 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-5b5cf99c8 SuccessfulDelete Deleted pod: simple-prod-collector-5b5cf99c8-9hmlv replicaset-controller logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:56 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-5b5cf99c8 SuccessfulDelete Deleted pod: simple-prod-collector-5b5cf99c8-k5n57 replicaset-controller logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:56 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-855467d9d5 SuccessfulCreate Created pod: simple-prod-collector-855467d9d5-fr94p replicaset-controller logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:56 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled down replica set simple-prod-collector-5b5cf99c8 to 0 from 2 deployment-controller logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:56 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-855467d9d5 to 2 deployment-controller logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:56 +0000 UTC Normal Pod simple-prod-query-7546f8f485-k8d9t Binding Scheduled Successfully assigned kuttl-test-polite-bull/simple-prod-query-7546f8f485-k8d9t to ip-10-0-26-111.us-east-2.compute.internal default-scheduler logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:56 +0000 UTC Normal ReplicaSet.apps simple-prod-query-7546f8f485 SuccessfulCreate Created pod: simple-prod-query-7546f8f485-k8d9t replicaset-controller logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:56 +0000 UTC Normal Pod simple-prod-query-bc57b7f66-kpshl.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:56 +0000 UTC Normal Pod simple-prod-query-bc57b7f66-kpshl.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:56 +0000 UTC Normal Pod simple-prod-query-bc57b7f66-kpshl.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:56 +0000 UTC Normal Pod simple-prod-query-bc57b7f66-krxfp.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:56 +0000 UTC Normal Pod simple-prod-query-bc57b7f66-krxfp.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:56 +0000 UTC Normal Pod simple-prod-query-bc57b7f66-krxfp.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:56 +0000 UTC Normal ReplicaSet.apps simple-prod-query-bc57b7f66 SuccessfulDelete Deleted pod: simple-prod-query-bc57b7f66-krxfp replicaset-controller logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:56 +0000 UTC Normal ReplicaSet.apps simple-prod-query-bc57b7f66 SuccessfulDelete Deleted pod: simple-prod-query-bc57b7f66-kpshl replicaset-controller logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:56 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled down replica set simple-prod-query-bc57b7f66 to 0 from 2 deployment-controller logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:56 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-7546f8f485 to 2 deployment-controller logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:57 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpolitebullsimpleprod-2-6d645445hmpx2 AddedInterface Add eth0 [10.129.2.30/23] from ovn-kubernetes multus logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:57 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpolitebullsimpleprod-2-6d645445hmpx2.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:410674bf2f77ff1d431f494cadfab3e81249b35695eb65aedfdb8498079b4d35" already present on machine kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:57 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpolitebullsimpleprod-2-6d645445hmpx2.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:57 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpolitebullsimpleprod-2-6d645445hmpx2.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:57 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpolitebullsimpleprod-2-6d645445hmpx2.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:fe2439e5f9e14d828caace425214edfa3edc179d72d225c43f69e9e4c9a5837b" already present on machine kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:57 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpolitebullsimpleprod-2-6d645445hmpx2.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:57 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpolitebullsimpleprod-2-6d645445hmpx2.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:57 +0000 UTC Normal Pod simple-prod-collector-855467d9d5-fr94p Binding Scheduled Successfully assigned kuttl-test-polite-bull/simple-prod-collector-855467d9d5-fr94p to ip-10-0-26-111.us-east-2.compute.internal default-scheduler logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:57 +0000 UTC Normal Pod simple-prod-collector-855467d9d5-fr94p AddedInterface Add eth0 [10.128.2.47/23] from ovn-kubernetes multus logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:57 +0000 UTC Normal Pod simple-prod-collector-855467d9d5-fr94p.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:40a1c7fa23aa5ffa64e8e22aa38022f5d4d7ff644c46a3da7169b713d486c3c1" already present on machine kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:57 +0000 UTC Normal Pod simple-prod-collector-855467d9d5-fr94p.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:57 +0000 UTC Normal Pod simple-prod-collector-855467d9d5-trzt7 Binding Scheduled Successfully assigned kuttl-test-polite-bull/simple-prod-collector-855467d9d5-trzt7 to ip-10-0-66-98.us-east-2.compute.internal default-scheduler logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:57 +0000 UTC Normal Pod simple-prod-collector-855467d9d5-trzt7 AddedInterface Add eth0 [10.129.2.32/23] from ovn-kubernetes multus logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:57 +0000 UTC Normal Pod simple-prod-collector-855467d9d5-trzt7.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:40a1c7fa23aa5ffa64e8e22aa38022f5d4d7ff644c46a3da7169b713d486c3c1" already present on machine kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:57 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-855467d9d5 SuccessfulCreate Created pod: simple-prod-collector-855467d9d5-trzt7 replicaset-controller logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:57 +0000 UTC Normal Pod simple-prod-query-7546f8f485-64ngf Binding Scheduled Successfully assigned kuttl-test-polite-bull/simple-prod-query-7546f8f485-64ngf to ip-10-0-66-98.us-east-2.compute.internal default-scheduler logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:57 +0000 UTC Normal Pod simple-prod-query-7546f8f485-64ngf AddedInterface Add eth0 [10.129.2.31/23] from ovn-kubernetes multus logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:57 +0000 UTC Normal Pod simple-prod-query-7546f8f485-64ngf.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:84020ce29bcb5a4bd018e6596188ae919c5cd600e08f78a546c0e76ea477685e" already present on machine kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:57 +0000 UTC Normal Pod simple-prod-query-7546f8f485-k8d9t AddedInterface Add eth0 [10.128.2.46/23] from ovn-kubernetes multus logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:57 +0000 UTC Normal Pod simple-prod-query-7546f8f485-k8d9t.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:84020ce29bcb5a4bd018e6596188ae919c5cd600e08f78a546c0e76ea477685e" already present on machine kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:57 +0000 UTC Normal Pod simple-prod-query-7546f8f485-k8d9t.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:57 +0000 UTC Normal Pod simple-prod-query-7546f8f485-k8d9t.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:57 +0000 UTC Normal Pod simple-prod-query-7546f8f485-k8d9t.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" already present on machine kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:57 +0000 UTC Normal ReplicaSet.apps simple-prod-query-7546f8f485 SuccessfulCreate Created pod: simple-prod-query-7546f8f485-64ngf replicaset-controller logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:58 +0000 UTC Normal Pod simple-prod-collector-855467d9d5-fr94p.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:58 +0000 UTC Normal Pod simple-prod-collector-855467d9d5-trzt7.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:58 +0000 UTC Normal Pod simple-prod-collector-855467d9d5-trzt7.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:58 +0000 UTC Normal Pod simple-prod-query-7546f8f485-64ngf.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:58 +0000 UTC Normal Pod simple-prod-query-7546f8f485-64ngf.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:58 +0000 UTC Normal Pod simple-prod-query-7546f8f485-64ngf.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" already present on machine kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:58 +0000 UTC Normal Pod simple-prod-query-7546f8f485-64ngf.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:58 +0000 UTC Normal Pod simple-prod-query-7546f8f485-64ngf.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:58 +0000 UTC Normal Pod simple-prod-query-7546f8f485-64ngf.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:65342c7d622f336741a8ae0b9dead79c7ecabc155d02cdd42b7b49ca36680e74" already present on machine kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:58 +0000 UTC Normal Pod simple-prod-query-7546f8f485-64ngf.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:58 +0000 UTC Normal Pod simple-prod-query-7546f8f485-64ngf.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:58 +0000 UTC Normal Pod simple-prod-query-7546f8f485-k8d9t.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:58 +0000 UTC Normal Pod simple-prod-query-7546f8f485-k8d9t.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:58 +0000 UTC Normal Pod simple-prod-query-7546f8f485-k8d9t.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:65342c7d622f336741a8ae0b9dead79c7ecabc155d02cdd42b7b49ca36680e74" already present on machine kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:58 +0000 UTC Normal Pod simple-prod-query-7546f8f485-k8d9t.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:24:59 | es-increasing-replicas | 2025-02-03 07:24:58 +0000 UTC Normal Pod simple-prod-query-7546f8f485-k8d9t.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:24:59 | es-increasing-replicas | Deleting namespace: kuttl-test-polite-bull === CONT kuttl/harness/es-index-cleaner-autoprov logger.go:42: 07:25:36 | es-index-cleaner-autoprov | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:25:36 | es-index-cleaner-autoprov | Creating namespace: kuttl-test-liked-heron logger.go:42: 07:25:37 | es-index-cleaner-autoprov/1-install | starting test step 1-install logger.go:42: 07:25:37 | es-index-cleaner-autoprov/1-install | Jaeger:kuttl-test-liked-heron/test-es-index-cleaner-with-prefix created logger.go:42: 07:26:24 | es-index-cleaner-autoprov/1-install | test step completed 1-install logger.go:42: 07:26:24 | es-index-cleaner-autoprov/2-report-spans | starting test step 2-report-spans logger.go:42: 07:26:24 | es-index-cleaner-autoprov/2-report-spans | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE test-es-index-cleaner-with-prefix /dev/null] logger.go:42: 07:26:25 | es-index-cleaner-autoprov/2-report-spans | Warning: resource jaegers/test-es-index-cleaner-with-prefix is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:26:31 | es-index-cleaner-autoprov/2-report-spans | running command: [sh -c DAYS=5 ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JOB_NUMBER=00 JAEGER_COLLECTOR_ENDPOINT=http://test-es-index-cleaner-with-prefix-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://test-es-index-cleaner-with-prefix-query MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/report-spans.yaml.template -o report-span-00-job.yaml] logger.go:42: 07:26:32 | es-index-cleaner-autoprov/2-report-spans | running command: [sh -c kubectl apply -f report-span-00-job.yaml -n $NAMESPACE] logger.go:42: 07:26:32 | es-index-cleaner-autoprov/2-report-spans | job.batch/00-report-span created logger.go:42: 07:36:33 | es-index-cleaner-autoprov/2-report-spans | test step failed 2-report-spans case.go:364: failed in step 2-report-spans case.go:366: --- Job:kuttl-test-liked-heron/00-report-span +++ Job:kuttl-test-liked-heron/00-report-span @@ -1,8 +1,162 @@ apiVersion: batch/v1 kind: Job metadata: + annotations: + kubectl.kubernetes.io/last-applied-configuration: | + {"apiVersion":"batch/v1","kind":"Job","metadata":{"annotations":{},"name":"00-report-span","namespace":"kuttl-test-liked-heron"},"spec":{"backoffLimit":15,"template":{"spec":{"containers":[{"command":["./reporter","--days","5","--verbose"],"env":[{"name":"JAEGER_SERVICE_NAME","value":"smoke-test-service"},{"name":"OPERATION_NAME","value":"smoke-test-operation"},{"name":"JAEGER_ENDPOINT","value":"http://test-es-index-cleaner-with-prefix-collector-headless:14268/api/traces"},{"name":"JAEGER_QUERY","value":"https://test-es-index-cleaner-with-prefix-query/api/traces"},{"name":"SECRET_PATH","value":"/var/run/secrets/api-token/token"}],"image":"quay.io/rhn_support_ikanse/jaeger-asserts:latest","name":"asserts-container","volumeMounts":[{"mountPath":"/var/run/secrets/api-token","name":"token-api-volume"}]}],"restartPolicy":"OnFailure","volumes":[{"name":"token-api-volume","secret":{"secretName":"e2e-test"}}]}}}} + labels: + batch.kubernetes.io/controller-uid: ab2a7871-3fb8-4b63-940a-0e4eaba59ce5 + batch.kubernetes.io/job-name: 00-report-span + controller-uid: ab2a7871-3fb8-4b63-940a-0e4eaba59ce5 + job-name: 00-report-span + managedFields: + - apiVersion: batch/v1 + fieldsType: FieldsV1 + fieldsV1: + f:metadata: + f:annotations: + .: {} + f:kubectl.kubernetes.io/last-applied-configuration: {} + f:spec: + f:backoffLimit: {} + f:completionMode: {} + f:completions: {} + f:manualSelector: {} + f:parallelism: {} + f:podReplacementPolicy: {} + f:suspend: {} + f:template: + f:spec: + f:containers: + k:{"name":"asserts-container"}: + .: {} + f:command: {} + f:env: + .: {} + k:{"name":"JAEGER_ENDPOINT"}: + .: {} + f:name: {} + f:value: {} + k:{"name":"JAEGER_QUERY"}: + .: {} + f:name: {} + f:value: {} + k:{"name":"JAEGER_SERVICE_NAME"}: + .: {} + f:name: {} + f:value: {} + k:{"name":"OPERATION_NAME"}: + .: {} + f:name: {} + f:value: {} + k:{"name":"SECRET_PATH"}: + .: {} + f:name: {} + f:value: {} + f:image: {} + f:imagePullPolicy: {} + f:name: {} + f:resources: {} + f:terminationMessagePath: {} + f:terminationMessagePolicy: {} + f:volumeMounts: + .: {} + k:{"mountPath":"/var/run/secrets/api-token"}: + .: {} + f:mountPath: {} + f:name: {} + f:dnsPolicy: {} + f:restartPolicy: {} + f:schedulerName: {} + f:securityContext: {} + f:terminationGracePeriodSeconds: {} + f:volumes: + .: {} + k:{"name":"token-api-volume"}: + .: {} + f:name: {} + f:secret: + .: {} + f:defaultMode: {} + f:secretName: {} + manager: kubectl-client-side-apply + operation: Update + time: "2025-02-03T07:26:32Z" + - apiVersion: batch/v1 + fieldsType: FieldsV1 + fieldsV1: + f:status: + f:active: {} + f:ready: {} + f:startTime: {} + f:terminating: {} + f:uncountedTerminatedPods: {} + manager: kube-controller-manager + operation: Update + subresource: status + time: "2025-02-03T07:26:36Z" name: 00-report-span namespace: kuttl-test-liked-heron +spec: + backoffLimit: 15 + completionMode: NonIndexed + completions: 1 + manualSelector: false + parallelism: 1 + podReplacementPolicy: TerminatingOrFailed + selector: + matchLabels: + batch.kubernetes.io/controller-uid: ab2a7871-3fb8-4b63-940a-0e4eaba59ce5 + suspend: false + template: + metadata: + creationTimestamp: null + labels: + batch.kubernetes.io/controller-uid: ab2a7871-3fb8-4b63-940a-0e4eaba59ce5 + batch.kubernetes.io/job-name: 00-report-span + controller-uid: ab2a7871-3fb8-4b63-940a-0e4eaba59ce5 + job-name: 00-report-span + spec: + containers: + - command: + - ./reporter + - --days + - "5" + - --verbose + env: + - name: JAEGER_SERVICE_NAME + value: smoke-test-service + - name: OPERATION_NAME + value: smoke-test-operation + - name: JAEGER_ENDPOINT + value: http://test-es-index-cleaner-with-prefix-collector-headless:14268/api/traces + - name: JAEGER_QUERY + value: https://test-es-index-cleaner-with-prefix-query/api/traces + - name: SECRET_PATH + value: /var/run/secrets/api-token/token + image: quay.io/rhn_support_ikanse/jaeger-asserts:latest + imagePullPolicy: Always + name: asserts-container + resources: {} + terminationMessagePath: /dev/termination-log + terminationMessagePolicy: File + volumeMounts: + - mountPath: /var/run/secrets/api-token + name: token-api-volume + dnsPolicy: ClusterFirst + restartPolicy: OnFailure + schedulerName: default-scheduler + securityContext: {} + terminationGracePeriodSeconds: 30 + volumes: + - name: token-api-volume + secret: + defaultMode: 420 + secretName: e2e-test status: - succeeded: 1 + active: 1 + ready: 1 + startTime: "2025-02-03T07:26:32Z" + terminating: 0 + uncountedTerminatedPods: {} case.go:366: resource Job:kuttl-test-liked-heron/00-report-span: .status.succeeded: key is missing from map logger.go:42: 07:36:33 | es-index-cleaner-autoprov | es-index-cleaner-autoprov events from ns kuttl-test-liked-heron: logger.go:42: 07:36:33 | es-index-cleaner-autoprov | 2025-02-03 07:25:52 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestlikedherontestesindexcleane-1-6886466f47 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestlikedherontestesindexcleane-1-6th2gh replicaset-controller logger.go:42: 07:36:33 | es-index-cleaner-autoprov | 2025-02-03 07:25:52 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestlikedherontestesindexcleane-1-6th2gh Binding Scheduled Successfully assigned kuttl-test-liked-heron/elasticsearch-cdm-kuttltestlikedherontestesindexcleane-1-6th2gh to ip-10-0-66-98.us-east-2.compute.internal default-scheduler logger.go:42: 07:36:33 | es-index-cleaner-autoprov | 2025-02-03 07:25:52 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestlikedherontestesindexcleane-1-6th2gh FailedMount MountVolume.SetUp failed for volume "elasticsearch-metrics" : secret "elasticsearch-metrics" not found kubelet logger.go:42: 07:36:33 | es-index-cleaner-autoprov | 2025-02-03 07:25:52 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestlikedherontestesindexcleane-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestlikedherontestesindexcleane-1-6886466f47 to 1 deployment-controller logger.go:42: 07:36:33 | es-index-cleaner-autoprov | 2025-02-03 07:25:53 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestlikedherontestesindexcleane-1-6th2gh AddedInterface Add eth0 [10.129.2.33/23] from ovn-kubernetes multus logger.go:42: 07:36:33 | es-index-cleaner-autoprov | 2025-02-03 07:25:53 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestlikedherontestesindexcleane-1-6th2gh.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:410674bf2f77ff1d431f494cadfab3e81249b35695eb65aedfdb8498079b4d35" already present on machine kubelet logger.go:42: 07:36:33 | es-index-cleaner-autoprov | 2025-02-03 07:25:54 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestlikedherontestesindexcleane-1-6th2gh.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:36:33 | es-index-cleaner-autoprov | 2025-02-03 07:25:54 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestlikedherontestesindexcleane-1-6th2gh.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:36:33 | es-index-cleaner-autoprov | 2025-02-03 07:25:54 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestlikedherontestesindexcleane-1-6th2gh.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:fe2439e5f9e14d828caace425214edfa3edc179d72d225c43f69e9e4c9a5837b" already present on machine kubelet logger.go:42: 07:36:33 | es-index-cleaner-autoprov | 2025-02-03 07:25:54 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestlikedherontestesindexcleane-1-6th2gh.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:36:33 | es-index-cleaner-autoprov | 2025-02-03 07:25:54 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestlikedherontestesindexcleane-1-6th2gh.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:36:33 | es-index-cleaner-autoprov | 2025-02-03 07:26:08 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestlikedherontestesindexcleane-1-6th2gh.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:36:33 | es-index-cleaner-autoprov | 2025-02-03 07:26:20 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-collector-6f6cf6d785-f6m8w Binding Scheduled Successfully assigned kuttl-test-liked-heron/test-es-index-cleaner-with-prefix-collector-6f6cf6d785-f6m8w to ip-10-0-6-172.us-east-2.compute.internal default-scheduler logger.go:42: 07:36:33 | es-index-cleaner-autoprov | 2025-02-03 07:26:20 +0000 UTC Warning Pod test-es-index-cleaner-with-prefix-collector-6f6cf6d785-f6m8w FailedMount MountVolume.SetUp failed for volume "test-es-index-cleaner-with-prefix-collector-tls-config-volume" : secret "test-es-index-cleaner-with-prefix-collector-headless-tls" not found kubelet logger.go:42: 07:36:33 | es-index-cleaner-autoprov | 2025-02-03 07:26:20 +0000 UTC Normal ReplicaSet.apps test-es-index-cleaner-with-prefix-collector-6f6cf6d785 SuccessfulCreate Created pod: test-es-index-cleaner-with-prefix-collector-6f6cf6d785-f6m8w replicaset-controller logger.go:42: 07:36:33 | es-index-cleaner-autoprov | 2025-02-03 07:26:20 +0000 UTC Normal Deployment.apps test-es-index-cleaner-with-prefix-collector ScalingReplicaSet Scaled up replica set test-es-index-cleaner-with-prefix-collector-6f6cf6d785 to 1 deployment-controller logger.go:42: 07:36:33 | es-index-cleaner-autoprov | 2025-02-03 07:26:20 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-b5b5bf98b-dh59n Binding Scheduled Successfully assigned kuttl-test-liked-heron/test-es-index-cleaner-with-prefix-query-b5b5bf98b-dh59n to ip-10-0-26-111.us-east-2.compute.internal default-scheduler logger.go:42: 07:36:33 | es-index-cleaner-autoprov | 2025-02-03 07:26:20 +0000 UTC Warning Pod test-es-index-cleaner-with-prefix-query-b5b5bf98b-dh59n FailedMount MountVolume.SetUp failed for volume "test-es-index-cleaner-with-prefix-ui-oauth-proxy-tls" : secret "test-es-index-cleaner-with-prefix-ui-oauth-proxy-tls" not found kubelet logger.go:42: 07:36:33 | es-index-cleaner-autoprov | 2025-02-03 07:26:20 +0000 UTC Normal ReplicaSet.apps test-es-index-cleaner-with-prefix-query-b5b5bf98b SuccessfulCreate Created pod: test-es-index-cleaner-with-prefix-query-b5b5bf98b-dh59n replicaset-controller logger.go:42: 07:36:33 | es-index-cleaner-autoprov | 2025-02-03 07:26:20 +0000 UTC Normal Deployment.apps test-es-index-cleaner-with-prefix-query ScalingReplicaSet Scaled up replica set test-es-index-cleaner-with-prefix-query-b5b5bf98b to 1 deployment-controller logger.go:42: 07:36:33 | es-index-cleaner-autoprov | 2025-02-03 07:26:21 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-collector-6f6cf6d785-f6m8w AddedInterface Add eth0 [10.131.0.26/23] from ovn-kubernetes multus logger.go:42: 07:36:33 | es-index-cleaner-autoprov | 2025-02-03 07:26:21 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-collector-6f6cf6d785-f6m8w.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:40a1c7fa23aa5ffa64e8e22aa38022f5d4d7ff644c46a3da7169b713d486c3c1" already present on machine kubelet logger.go:42: 07:36:33 | es-index-cleaner-autoprov | 2025-02-03 07:26:21 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-b5b5bf98b-dh59n AddedInterface Add eth0 [10.128.2.48/23] from ovn-kubernetes multus logger.go:42: 07:36:33 | es-index-cleaner-autoprov | 2025-02-03 07:26:21 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-b5b5bf98b-dh59n.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:84020ce29bcb5a4bd018e6596188ae919c5cd600e08f78a546c0e76ea477685e" already present on machine kubelet logger.go:42: 07:36:33 | es-index-cleaner-autoprov | 2025-02-03 07:26:21 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-b5b5bf98b-dh59n.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:36:33 | es-index-cleaner-autoprov | 2025-02-03 07:26:22 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-collector-6f6cf6d785-f6m8w.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:36:33 | es-index-cleaner-autoprov | 2025-02-03 07:26:22 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-collector-6f6cf6d785-f6m8w.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:36:33 | es-index-cleaner-autoprov | 2025-02-03 07:26:22 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-b5b5bf98b-dh59n.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:36:33 | es-index-cleaner-autoprov | 2025-02-03 07:26:22 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-b5b5bf98b-dh59n.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" already present on machine kubelet logger.go:42: 07:36:33 | es-index-cleaner-autoprov | 2025-02-03 07:26:22 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-b5b5bf98b-dh59n.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:36:33 | es-index-cleaner-autoprov | 2025-02-03 07:26:22 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-b5b5bf98b-dh59n.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:36:33 | es-index-cleaner-autoprov | 2025-02-03 07:26:22 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-b5b5bf98b-dh59n.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:65342c7d622f336741a8ae0b9dead79c7ecabc155d02cdd42b7b49ca36680e74" already present on machine kubelet logger.go:42: 07:36:33 | es-index-cleaner-autoprov | 2025-02-03 07:26:22 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-b5b5bf98b-dh59n.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:36:33 | es-index-cleaner-autoprov | 2025-02-03 07:26:22 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-b5b5bf98b-dh59n.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:36:33 | es-index-cleaner-autoprov | 2025-02-03 07:26:32 +0000 UTC Normal Pod 00-report-span-5m2gp Binding Scheduled Successfully assigned kuttl-test-liked-heron/00-report-span-5m2gp to ip-10-0-6-172.us-east-2.compute.internal default-scheduler logger.go:42: 07:36:33 | es-index-cleaner-autoprov | 2025-02-03 07:26:32 +0000 UTC Normal Job.batch 00-report-span SuccessfulCreate Created pod: 00-report-span-5m2gp job-controller logger.go:42: 07:36:33 | es-index-cleaner-autoprov | 2025-02-03 07:26:33 +0000 UTC Normal Pod 00-report-span-5m2gp AddedInterface Add eth0 [10.131.0.27/23] from ovn-kubernetes multus logger.go:42: 07:36:33 | es-index-cleaner-autoprov | 2025-02-03 07:26:33 +0000 UTC Normal Pod 00-report-span-5m2gp.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:36:33 | es-index-cleaner-autoprov | 2025-02-03 07:26:34 +0000 UTC Normal Pod 00-report-span-5m2gp.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 1.73s (1.73s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:36:33 | es-index-cleaner-autoprov | 2025-02-03 07:26:35 +0000 UTC Normal Pod 00-report-span-5m2gp.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:36:33 | es-index-cleaner-autoprov | 2025-02-03 07:26:35 +0000 UTC Normal Pod 00-report-span-5m2gp.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:36:33 | es-index-cleaner-autoprov | 2025-02-03 07:26:35 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedGetResourceMetric failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 07:36:33 | es-index-cleaner-autoprov | 2025-02-03 07:26:35 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:36:33 | es-index-cleaner-autoprov | 2025-02-03 07:26:35 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 07:36:33 | es-index-cleaner-autoprov | 2025-02-03 07:26:50 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedGetResourceMetric failed to get memory utilization: missing request for memory in container jaeger-collector of Pod test-es-index-cleaner-with-prefix-collector-6f6cf6d785-f6m8w horizontal-pod-autoscaler logger.go:42: 07:36:33 | es-index-cleaner-autoprov | 2025-02-03 07:27:05 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedGetResourceMetric failed to get cpu utilization: missing request for cpu in container jaeger-collector of Pod test-es-index-cleaner-with-prefix-collector-6f6cf6d785-f6m8w horizontal-pod-autoscaler logger.go:42: 07:36:33 | es-index-cleaner-autoprov | 2025-02-03 07:27:05 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: missing request for cpu in container jaeger-collector of Pod test-es-index-cleaner-with-prefix-collector-6f6cf6d785-f6m8w horizontal-pod-autoscaler logger.go:42: 07:36:33 | es-index-cleaner-autoprov | Deleting namespace: kuttl-test-liked-heron === CONT kuttl/harness/es-from-aio-to-production logger.go:42: 07:36:46 | es-from-aio-to-production | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:36:46 | es-from-aio-to-production | Creating namespace: kuttl-test-mutual-bengal logger.go:42: 07:36:46 | es-from-aio-to-production/0-install | starting test step 0-install logger.go:42: 07:36:46 | es-from-aio-to-production/0-install | Jaeger:kuttl-test-mutual-bengal/my-jaeger created logger.go:42: 07:36:55 | es-from-aio-to-production/0-install | test step completed 0-install logger.go:42: 07:36:55 | es-from-aio-to-production/1-smoke-test | starting test step 1-smoke-test logger.go:42: 07:36:55 | es-from-aio-to-production/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:36:56 | es-from-aio-to-production/1-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:37:05 | es-from-aio-to-production/1-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:37:05 | es-from-aio-to-production/1-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:37:05 | es-from-aio-to-production/1-smoke-test | job.batch/report-span created logger.go:42: 07:37:05 | es-from-aio-to-production/1-smoke-test | job.batch/check-span created logger.go:42: 07:37:18 | es-from-aio-to-production/1-smoke-test | test step completed 1-smoke-test logger.go:42: 07:37:18 | es-from-aio-to-production/3-install | starting test step 3-install logger.go:42: 07:37:18 | es-from-aio-to-production/3-install | Jaeger:kuttl-test-mutual-bengal/my-jaeger updated logger.go:42: 07:38:05 | es-from-aio-to-production/3-install | test step completed 3-install logger.go:42: 07:38:05 | es-from-aio-to-production/4-smoke-test | starting test step 4-smoke-test logger.go:42: 07:38:05 | es-from-aio-to-production/4-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:38:12 | es-from-aio-to-production/4-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:38:13 | es-from-aio-to-production/4-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:38:13 | es-from-aio-to-production/4-smoke-test | job.batch/report-span unchanged logger.go:42: 07:38:13 | es-from-aio-to-production/4-smoke-test | job.batch/check-span unchanged logger.go:42: 07:38:13 | es-from-aio-to-production/4-smoke-test | test step completed 4-smoke-test logger.go:42: 07:38:13 | es-from-aio-to-production | es-from-aio-to-production events from ns kuttl-test-mutual-bengal: logger.go:42: 07:38:13 | es-from-aio-to-production | 2025-02-03 07:36:50 +0000 UTC Normal Pod my-jaeger-65556db95f-mdz7k Binding Scheduled Successfully assigned kuttl-test-mutual-bengal/my-jaeger-65556db95f-mdz7k to ip-10-0-66-98.us-east-2.compute.internal default-scheduler logger.go:42: 07:38:13 | es-from-aio-to-production | 2025-02-03 07:36:50 +0000 UTC Normal ReplicaSet.apps my-jaeger-65556db95f SuccessfulCreate Created pod: my-jaeger-65556db95f-mdz7k replicaset-controller logger.go:42: 07:38:13 | es-from-aio-to-production | 2025-02-03 07:36:50 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-65556db95f to 1 deployment-controller logger.go:42: 07:38:13 | es-from-aio-to-production | 2025-02-03 07:36:51 +0000 UTC Normal Pod my-jaeger-65556db95f-mdz7k AddedInterface Add eth0 [10.129.2.34/23] from ovn-kubernetes multus logger.go:42: 07:38:13 | es-from-aio-to-production | 2025-02-03 07:36:51 +0000 UTC Normal Pod my-jaeger-65556db95f-mdz7k.spec.containers{jaeger} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:fa9aaf9ae026228265ce7f10fb451d5c52a4e0269ce1fa9024cbefddf765ae8d" kubelet logger.go:42: 07:38:13 | es-from-aio-to-production | 2025-02-03 07:36:53 +0000 UTC Normal Pod my-jaeger-65556db95f-mdz7k.spec.containers{jaeger} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:fa9aaf9ae026228265ce7f10fb451d5c52a4e0269ce1fa9024cbefddf765ae8d" in 2.136s (2.136s including waiting). Image size: 144304495 bytes. kubelet logger.go:42: 07:38:13 | es-from-aio-to-production | 2025-02-03 07:36:53 +0000 UTC Normal Pod my-jaeger-65556db95f-mdz7k.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:38:13 | es-from-aio-to-production | 2025-02-03 07:36:53 +0000 UTC Normal Pod my-jaeger-65556db95f-mdz7k.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:38:13 | es-from-aio-to-production | 2025-02-03 07:36:53 +0000 UTC Normal Pod my-jaeger-65556db95f-mdz7k.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" already present on machine kubelet logger.go:42: 07:38:13 | es-from-aio-to-production | 2025-02-03 07:36:53 +0000 UTC Normal Pod my-jaeger-65556db95f-mdz7k.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:38:13 | es-from-aio-to-production | 2025-02-03 07:36:53 +0000 UTC Normal Pod my-jaeger-65556db95f-mdz7k.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:38:13 | es-from-aio-to-production | 2025-02-03 07:37:00 +0000 UTC Normal Pod my-jaeger-65556db95f-mdz7k.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:38:13 | es-from-aio-to-production | 2025-02-03 07:37:00 +0000 UTC Normal Pod my-jaeger-65556db95f-mdz7k.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:38:13 | es-from-aio-to-production | 2025-02-03 07:37:00 +0000 UTC Normal ReplicaSet.apps my-jaeger-65556db95f SuccessfulDelete Deleted pod: my-jaeger-65556db95f-mdz7k replicaset-controller logger.go:42: 07:38:13 | es-from-aio-to-production | 2025-02-03 07:37:00 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled down replica set my-jaeger-65556db95f to 0 from 1 deployment-controller logger.go:42: 07:38:13 | es-from-aio-to-production | 2025-02-03 07:37:01 +0000 UTC Normal ReplicaSet.apps my-jaeger-6c4db85c5b SuccessfulCreate Created pod: my-jaeger-6c4db85c5b-twgr5 replicaset-controller logger.go:42: 07:38:13 | es-from-aio-to-production | 2025-02-03 07:37:01 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-6c4db85c5b to 1 deployment-controller logger.go:42: 07:38:13 | es-from-aio-to-production | 2025-02-03 07:37:02 +0000 UTC Warning Pod my-jaeger-65556db95f-mdz7k.spec.containers{jaeger} Unhealthy Readiness probe failed: Get "http://10.129.2.34:14269/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:38:13 | es-from-aio-to-production | 2025-02-03 07:37:02 +0000 UTC Normal Pod my-jaeger-6c4db85c5b-twgr5 Binding Scheduled Successfully assigned kuttl-test-mutual-bengal/my-jaeger-6c4db85c5b-twgr5 to ip-10-0-66-98.us-east-2.compute.internal default-scheduler logger.go:42: 07:38:13 | es-from-aio-to-production | 2025-02-03 07:37:02 +0000 UTC Normal Pod my-jaeger-6c4db85c5b-twgr5 AddedInterface Add eth0 [10.129.2.35/23] from ovn-kubernetes multus logger.go:42: 07:38:13 | es-from-aio-to-production | 2025-02-03 07:37:02 +0000 UTC Normal Pod my-jaeger-6c4db85c5b-twgr5.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:fa9aaf9ae026228265ce7f10fb451d5c52a4e0269ce1fa9024cbefddf765ae8d" already present on machine kubelet logger.go:42: 07:38:13 | es-from-aio-to-production | 2025-02-03 07:37:02 +0000 UTC Normal Pod my-jaeger-6c4db85c5b-twgr5.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:38:13 | es-from-aio-to-production | 2025-02-03 07:37:02 +0000 UTC Normal Pod my-jaeger-6c4db85c5b-twgr5.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:38:13 | es-from-aio-to-production | 2025-02-03 07:37:02 +0000 UTC Normal Pod my-jaeger-6c4db85c5b-twgr5.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" already present on machine kubelet logger.go:42: 07:38:13 | es-from-aio-to-production | 2025-02-03 07:37:02 +0000 UTC Normal Pod my-jaeger-6c4db85c5b-twgr5.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:38:13 | es-from-aio-to-production | 2025-02-03 07:37:03 +0000 UTC Normal Pod my-jaeger-6c4db85c5b-twgr5.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:38:13 | es-from-aio-to-production | 2025-02-03 07:37:05 +0000 UTC Normal Pod check-span-bxlch Binding Scheduled Successfully assigned kuttl-test-mutual-bengal/check-span-bxlch to ip-10-0-26-111.us-east-2.compute.internal default-scheduler logger.go:42: 07:38:13 | es-from-aio-to-production | 2025-02-03 07:37:05 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-bxlch job-controller logger.go:42: 07:38:13 | es-from-aio-to-production | 2025-02-03 07:37:05 +0000 UTC Normal Pod report-span-cbqsx Binding Scheduled Successfully assigned kuttl-test-mutual-bengal/report-span-cbqsx to ip-10-0-6-172.us-east-2.compute.internal default-scheduler logger.go:42: 07:38:13 | es-from-aio-to-production | 2025-02-03 07:37:05 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-cbqsx job-controller logger.go:42: 07:38:13 | es-from-aio-to-production | 2025-02-03 07:37:06 +0000 UTC Normal Pod check-span-bxlch AddedInterface Add eth0 [10.128.2.49/23] from ovn-kubernetes multus logger.go:42: 07:38:13 | es-from-aio-to-production | 2025-02-03 07:37:06 +0000 UTC Normal Pod check-span-bxlch.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:38:13 | es-from-aio-to-production | 2025-02-03 07:37:06 +0000 UTC Normal Pod report-span-cbqsx AddedInterface Add eth0 [10.131.0.29/23] from ovn-kubernetes multus logger.go:42: 07:38:13 | es-from-aio-to-production | 2025-02-03 07:37:06 +0000 UTC Normal Pod report-span-cbqsx.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:38:13 | es-from-aio-to-production | 2025-02-03 07:37:07 +0000 UTC Normal Pod check-span-bxlch.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 573ms (573ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:38:13 | es-from-aio-to-production | 2025-02-03 07:37:07 +0000 UTC Normal Pod check-span-bxlch.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:38:13 | es-from-aio-to-production | 2025-02-03 07:37:07 +0000 UTC Normal Pod check-span-bxlch.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:38:13 | es-from-aio-to-production | 2025-02-03 07:37:07 +0000 UTC Normal Pod report-span-cbqsx.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 861ms (861ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:38:13 | es-from-aio-to-production | 2025-02-03 07:37:07 +0000 UTC Normal Pod report-span-cbqsx.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:38:13 | es-from-aio-to-production | 2025-02-03 07:37:07 +0000 UTC Normal Pod report-span-cbqsx.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:38:13 | es-from-aio-to-production | 2025-02-03 07:37:17 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:38:13 | es-from-aio-to-production | 2025-02-03 07:37:36 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmutualbengalmyjaeger-1-5d4b486d6wnzx Binding Scheduled Successfully assigned kuttl-test-mutual-bengal/elasticsearch-cdm-kuttltestmutualbengalmyjaeger-1-5d4b486d6wnzx to ip-10-0-6-172.us-east-2.compute.internal default-scheduler logger.go:42: 07:38:13 | es-from-aio-to-production | 2025-02-03 07:37:36 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmutualbengalmyjaeger-1-5d4b486d6wnzx AddedInterface Add eth0 [10.131.0.30/23] from ovn-kubernetes multus logger.go:42: 07:38:13 | es-from-aio-to-production | 2025-02-03 07:37:36 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestmutualbengalmyjaeger-1-5d4b486d95 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestmutualbengalmyjaeger-1-5d4b486d6wnzx replicaset-controller logger.go:42: 07:38:13 | es-from-aio-to-production | 2025-02-03 07:37:36 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestmutualbengalmyjaeger-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestmutualbengalmyjaeger-1-5d4b486d95 to 1 deployment-controller logger.go:42: 07:38:13 | es-from-aio-to-production | 2025-02-03 07:37:37 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmutualbengalmyjaeger-1-5d4b486d6wnzx.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:410674bf2f77ff1d431f494cadfab3e81249b35695eb65aedfdb8498079b4d35" already present on machine kubelet logger.go:42: 07:38:13 | es-from-aio-to-production | 2025-02-03 07:37:37 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmutualbengalmyjaeger-1-5d4b486d6wnzx.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:38:13 | es-from-aio-to-production | 2025-02-03 07:37:37 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmutualbengalmyjaeger-1-5d4b486d6wnzx.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:38:13 | es-from-aio-to-production | 2025-02-03 07:37:37 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmutualbengalmyjaeger-1-5d4b486d6wnzx.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:fe2439e5f9e14d828caace425214edfa3edc179d72d225c43f69e9e4c9a5837b" already present on machine kubelet logger.go:42: 07:38:13 | es-from-aio-to-production | 2025-02-03 07:37:37 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmutualbengalmyjaeger-1-5d4b486d6wnzx.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:38:13 | es-from-aio-to-production | 2025-02-03 07:37:37 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmutualbengalmyjaeger-1-5d4b486d6wnzx.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:38:13 | es-from-aio-to-production | 2025-02-03 07:37:39 +0000 UTC Normal Job.batch report-span Completed Job completed job-controller logger.go:42: 07:38:13 | es-from-aio-to-production | 2025-02-03 07:37:52 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestmutualbengalmyjaeger-1-5d4b486d6wnzx.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:38:13 | es-from-aio-to-production | 2025-02-03 07:38:03 +0000 UTC Normal Pod my-jaeger-6c4db85c5b-twgr5.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:38:13 | es-from-aio-to-production | 2025-02-03 07:38:03 +0000 UTC Normal Pod my-jaeger-6c4db85c5b-twgr5.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:38:13 | es-from-aio-to-production | 2025-02-03 07:38:03 +0000 UTC Normal Pod my-jaeger-collector-85d685857-rj9qx Binding Scheduled Successfully assigned kuttl-test-mutual-bengal/my-jaeger-collector-85d685857-rj9qx to ip-10-0-26-111.us-east-2.compute.internal default-scheduler logger.go:42: 07:38:13 | es-from-aio-to-production | 2025-02-03 07:38:03 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-85d685857 SuccessfulCreate Created pod: my-jaeger-collector-85d685857-rj9qx replicaset-controller logger.go:42: 07:38:13 | es-from-aio-to-production | 2025-02-03 07:38:03 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-85d685857 to 1 deployment-controller logger.go:42: 07:38:13 | es-from-aio-to-production | 2025-02-03 07:38:03 +0000 UTC Normal Pod my-jaeger-query-6b6d5c6585-5rlbw Binding Scheduled Successfully assigned kuttl-test-mutual-bengal/my-jaeger-query-6b6d5c6585-5rlbw to ip-10-0-26-111.us-east-2.compute.internal default-scheduler logger.go:42: 07:38:13 | es-from-aio-to-production | 2025-02-03 07:38:03 +0000 UTC Normal Pod my-jaeger-query-6b6d5c6585-5rlbw AddedInterface Add eth0 [10.128.2.51/23] from ovn-kubernetes multus logger.go:42: 07:38:13 | es-from-aio-to-production | 2025-02-03 07:38:03 +0000 UTC Normal Pod my-jaeger-query-6b6d5c6585-5rlbw.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:84020ce29bcb5a4bd018e6596188ae919c5cd600e08f78a546c0e76ea477685e" already present on machine kubelet logger.go:42: 07:38:13 | es-from-aio-to-production | 2025-02-03 07:38:03 +0000 UTC Normal Pod my-jaeger-query-6b6d5c6585-5rlbw.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:38:13 | es-from-aio-to-production | 2025-02-03 07:38:03 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-6b6d5c6585 SuccessfulCreate Created pod: my-jaeger-query-6b6d5c6585-5rlbw replicaset-controller logger.go:42: 07:38:13 | es-from-aio-to-production | 2025-02-03 07:38:03 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-6b6d5c6585 to 1 deployment-controller logger.go:42: 07:38:13 | es-from-aio-to-production | 2025-02-03 07:38:04 +0000 UTC Normal Pod my-jaeger-collector-85d685857-rj9qx AddedInterface Add eth0 [10.128.2.50/23] from ovn-kubernetes multus logger.go:42: 07:38:13 | es-from-aio-to-production | 2025-02-03 07:38:04 +0000 UTC Normal Pod my-jaeger-collector-85d685857-rj9qx.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:40a1c7fa23aa5ffa64e8e22aa38022f5d4d7ff644c46a3da7169b713d486c3c1" already present on machine kubelet logger.go:42: 07:38:13 | es-from-aio-to-production | 2025-02-03 07:38:04 +0000 UTC Normal Pod my-jaeger-collector-85d685857-rj9qx.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:38:13 | es-from-aio-to-production | 2025-02-03 07:38:04 +0000 UTC Normal Pod my-jaeger-collector-85d685857-rj9qx.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:38:13 | es-from-aio-to-production | 2025-02-03 07:38:04 +0000 UTC Normal Pod my-jaeger-query-6b6d5c6585-5rlbw.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:38:13 | es-from-aio-to-production | 2025-02-03 07:38:04 +0000 UTC Normal Pod my-jaeger-query-6b6d5c6585-5rlbw.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" already present on machine kubelet logger.go:42: 07:38:13 | es-from-aio-to-production | 2025-02-03 07:38:04 +0000 UTC Normal Pod my-jaeger-query-6b6d5c6585-5rlbw.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:38:13 | es-from-aio-to-production | 2025-02-03 07:38:04 +0000 UTC Normal Pod my-jaeger-query-6b6d5c6585-5rlbw.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:38:13 | es-from-aio-to-production | 2025-02-03 07:38:04 +0000 UTC Normal Pod my-jaeger-query-6b6d5c6585-5rlbw.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:65342c7d622f336741a8ae0b9dead79c7ecabc155d02cdd42b7b49ca36680e74" already present on machine kubelet logger.go:42: 07:38:13 | es-from-aio-to-production | 2025-02-03 07:38:04 +0000 UTC Normal Pod my-jaeger-query-6b6d5c6585-5rlbw.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:38:13 | es-from-aio-to-production | 2025-02-03 07:38:04 +0000 UTC Normal Pod my-jaeger-query-6b6d5c6585-5rlbw.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:38:13 | es-from-aio-to-production | Deleting namespace: kuttl-test-mutual-bengal === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- FAIL: kuttl (1325.89s) --- FAIL: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (7.15s) --- PASS: kuttl/harness/es-multiinstance (138.75s) --- PASS: kuttl/harness/es-simple-prod (6.39s) --- PASS: kuttl/harness/es-rollover-autoprov (284.61s) --- PASS: kuttl/harness/es-increasing-replicas (125.49s) --- FAIL: kuttl/harness/es-index-cleaner-autoprov (669.58s) --- PASS: kuttl/harness/es-from-aio-to-production (93.88s) FAIL + exit_code=1 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name elasticsearch --report --output /logs/artifacts/elasticsearch.xml ./artifacts/kuttl-report.xml time="2025-02-03T07:38:21Z" level=debug msg="Setting a new name for the test suites" time="2025-02-03T07:38:21Z" level=debug msg="Removing 'artifacts' TestCase" time="2025-02-03T07:38:21Z" level=debug msg="normalizing test case names" time="2025-02-03T07:38:21Z" level=debug msg="elasticsearch/artifacts -> elasticsearch_artifacts" time="2025-02-03T07:38:21Z" level=debug msg="elasticsearch/es-multiinstance -> elasticsearch_es_multiinstance" time="2025-02-03T07:38:21Z" level=debug msg="elasticsearch/es-simple-prod -> elasticsearch_es_simple_prod" time="2025-02-03T07:38:21Z" level=debug msg="elasticsearch/es-rollover-autoprov -> elasticsearch_es_rollover_autoprov" time="2025-02-03T07:38:21Z" level=debug msg="elasticsearch/es-increasing-replicas -> elasticsearch_es_increasing_replicas" time="2025-02-03T07:38:21Z" level=debug msg="elasticsearch/es-index-cleaner-autoprov -> elasticsearch_es_index_cleaner_autoprov" time="2025-02-03T07:38:21Z" level=debug msg="elasticsearch/es-from-aio-to-production -> elasticsearch_es_from_aio_to_production" +-----------------------------------------+--------+ | NAME | RESULT | +-----------------------------------------+--------+ | elasticsearch_artifacts | passed | | elasticsearch_es_multiinstance | passed | | elasticsearch_es_simple_prod | passed | | elasticsearch_es_rollover_autoprov | passed | | elasticsearch_es_increasing_replicas | passed | | elasticsearch_es_index_cleaner_autoprov | failed | | elasticsearch_es_from_aio_to_production | passed | +-----------------------------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh examples false true + '[' 3 -ne 3 ']' + test_suite_name=examples + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/examples.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-examples make[2]: Entering directory '/tmp/jaeger-tests' >>>> Elasticsearch image not loaded because SKIP_ES_EXTERNAL is true KAFKA_VERSION=3.6.0 \ SKIP_KAFKA=false \ VERTX_IMG=jaegertracing/vertx-create-span:operator-e2e-tests \ ./tests/e2e/examples/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.18.0-0.nightly-2025-02-01-185117 True False 29m Cluster version is 4.18.0-0.nightly-2025-02-01-185117' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.18.0-0.nightly-2025-02-01-185117 True False 29m Cluster version is 4.18.0-0.nightly-2025-02-01-185117' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/examples/render.sh ++ export SUITE_DIR=./tests/e2e/examples ++ SUITE_DIR=./tests/e2e/examples ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/examples ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ xargs -I '{}' cp -r '{}' _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + start_test examples-agent-with-priority-class + '[' 1 -ne 1 ']' + test_name=examples-agent-with-priority-class + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-agent-with-priority-class' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-agent-with-priority-class\e[0m' Rendering files for test examples-agent-with-priority-class + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build + '[' _build '!=' _build ']' + mkdir -p examples-agent-with-priority-class + cd examples-agent-with-priority-class + example_name=agent-with-priority-class + prepare_daemonset 00 + '[' 1 -ne 1 ']' + test_step=00 + '[' true = true ']' + cat /tmp/jaeger-tests/examples/openshift/hostport-scc-daemonset.yaml + echo --- + cat /tmp/jaeger-tests/examples/openshift/service_account_jaeger-agent-daemonset.yaml + '[' true '!=' true ']' + render_install_example agent-with-priority-class 02 + '[' 2 -ne 2 ']' + example_name=agent-with-priority-class + test_step=02 + install_file=./02-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/agent-with-priority-class.yaml -o ./02-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./02-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./02-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./02-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./02-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./02-install.yaml ++ jaeger_name=agent-as-daemonset ++ '[' -z agent-as-daemonset ']' ++ echo agent-as-daemonset ++ return 0 + JAEGER_NAME=agent-as-daemonset + local jaeger_strategy ++ get_jaeger_strategy ./02-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./02-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./02-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./02-install.yaml ++ strategy=DaemonSet ++ '[' DaemonSet = null ']' ++ echo DaemonSet ++ return 0 + jaeger_strategy=DaemonSet + '[' DaemonSet = DaemonSet ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./02-assert.yaml + render_smoke_test_example agent-with-priority-class 02 + '[' 2 -ne 2 ']' + example_name=agent-with-priority-class + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/agent-with-priority-class.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/agent-with-priority-class.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/agent-with-priority-class.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/agent-with-priority-class.yaml ++ jaeger_name=agent-as-daemonset ++ '[' -z agent-as-daemonset ']' ++ echo agent-as-daemonset ++ return 0 + jaeger_name=agent-as-daemonset + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test agent-as-daemonset true 02 + '[' 3 -ne 3 ']' + jaeger=agent-as-daemonset + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://agent-as-daemonset-query:443 + JAEGER_QUERY_ENDPOINT=https://agent-as-daemonset-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://agent-as-daemonset-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://agent-as-daemonset-collector-headless:14268 + export JAEGER_NAME=agent-as-daemonset + JAEGER_NAME=agent-as-daemonset + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-all-in-one-with-options + '[' 1 -ne 1 ']' + test_name=examples-all-in-one-with-options + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-all-in-one-with-options' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-all-in-one-with-options\e[0m' Rendering files for test examples-all-in-one-with-options + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-agent-with-priority-class + '[' examples-agent-with-priority-class '!=' _build ']' + cd .. + mkdir -p examples-all-in-one-with-options + cd examples-all-in-one-with-options + example_name=all-in-one-with-options + render_install_example all-in-one-with-options 00 + '[' 2 -ne 2 ']' + example_name=all-in-one-with-options + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/all-in-one-with-options.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=my-jaeger ++ '[' -z my-jaeger ']' ++ echo my-jaeger ++ return 0 + JAEGER_NAME=my-jaeger + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=allInOne ++ '[' allInOne = production ']' ++ '[' allInOne = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + /tmp/jaeger-tests/bin/yq e -i '.metadata.name="my-jaeger"' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i 'del(.spec.allInOne.image)' ./00-install.yaml + render_smoke_test_example all-in-one-with-options 01 + '[' 2 -ne 2 ']' + example_name=all-in-one-with-options + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/all-in-one-with-options.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/all-in-one-with-options.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/all-in-one-with-options.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/all-in-one-with-options.yaml ++ jaeger_name=my-jaeger ++ '[' -z my-jaeger ']' ++ echo my-jaeger ++ return 0 + jaeger_name=my-jaeger + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test my-jaeger true 01 + '[' 3 -ne 3 ']' + jaeger=my-jaeger + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + '[' true = true ']' + sed -i s~my-jaeger-query:443~my-jaeger-query:443/jaeger~gi ./01-smoke-test.yaml + start_test examples-business-application-injected-sidecar + '[' 1 -ne 1 ']' + test_name=examples-business-application-injected-sidecar + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-business-application-injected-sidecar' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-business-application-injected-sidecar\e[0m' Rendering files for test examples-business-application-injected-sidecar + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-all-in-one-with-options + '[' examples-all-in-one-with-options '!=' _build ']' + cd .. + mkdir -p examples-business-application-injected-sidecar + cd examples-business-application-injected-sidecar + example_name=simplest + cp /tmp/jaeger-tests/examples/business-application-injected-sidecar.yaml ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].image=strenv(VERTX_IMG)' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.httpGet.path="/"' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.httpGet.port=8080' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.initialDelaySeconds=1' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.failureThreshold=3' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.periodSeconds=10' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.successThreshold=1' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.timeoutSeconds=1' ./00-install.yaml + render_install_example simplest 01 + '[' 2 -ne 2 ']' + example_name=simplest + test_step=01 + install_file=./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/simplest.yaml -o ./01-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./01-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./01-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./01-install.yaml ++ jaeger_name=simplest ++ '[' -z simplest ']' ++ echo simplest ++ return 0 + JAEGER_NAME=simplest + local jaeger_strategy ++ get_jaeger_strategy ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./01-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./01-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./01-assert.yaml + render_smoke_test_example simplest 02 + '[' 2 -ne 2 ']' + example_name=simplest + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/simplest.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/simplest.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/simplest.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/simplest.yaml ++ jaeger_name=simplest ++ '[' -z simplest ']' ++ echo simplest ++ return 0 + jaeger_name=simplest + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test simplest true 02 + '[' 3 -ne 3 ']' + jaeger=simplest + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simplest-query:443 + JAEGER_QUERY_ENDPOINT=https://simplest-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 + export JAEGER_NAME=simplest + JAEGER_NAME=simplest + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-collector-with-priority-class + '[' 1 -ne 1 ']' + test_name=examples-collector-with-priority-class + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-collector-with-priority-class' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-collector-with-priority-class\e[0m' Rendering files for test examples-collector-with-priority-class + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-business-application-injected-sidecar + '[' examples-business-application-injected-sidecar '!=' _build ']' + cd .. + mkdir -p examples-collector-with-priority-class + cd examples-collector-with-priority-class + example_name=collector-with-priority-class + render_install_example collector-with-priority-class 00 + '[' 2 -ne 2 ']' + example_name=collector-with-priority-class + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/collector-with-priority-class.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=collector-with-high-priority ++ '[' -z collector-with-high-priority ']' ++ echo collector-with-high-priority ++ return 0 + JAEGER_NAME=collector-with-high-priority + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test_example collector-with-priority-class 01 + '[' 2 -ne 2 ']' + example_name=collector-with-priority-class + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/collector-with-priority-class.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/collector-with-priority-class.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/collector-with-priority-class.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/collector-with-priority-class.yaml ++ jaeger_name=collector-with-high-priority ++ '[' -z collector-with-high-priority ']' ++ echo collector-with-high-priority ++ return 0 + jaeger_name=collector-with-high-priority + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test collector-with-high-priority true 01 + '[' 3 -ne 3 ']' + jaeger=collector-with-high-priority + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://collector-with-high-priority-query:443 + JAEGER_QUERY_ENDPOINT=https://collector-with-high-priority-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://collector-with-high-priority-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://collector-with-high-priority-collector-headless:14268 + export JAEGER_NAME=collector-with-high-priority + JAEGER_NAME=collector-with-high-priority + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-service-types + '[' 1 -ne 1 ']' + test_name=examples-service-types + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-service-types' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-service-types\e[0m' Rendering files for test examples-service-types + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-collector-with-priority-class + '[' examples-collector-with-priority-class '!=' _build ']' + cd .. + mkdir -p examples-service-types + cd examples-service-types + example_name=service-types + render_install_example service-types 00 + '[' 2 -ne 2 ']' + example_name=service-types + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/service-types.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=service-types ++ '[' -z service-types ']' ++ echo service-types ++ return 0 + JAEGER_NAME=service-types + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test_example service-types 01 + '[' 2 -ne 2 ']' + example_name=service-types + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/service-types.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/service-types.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/service-types.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/service-types.yaml ++ jaeger_name=service-types ++ '[' -z service-types ']' ++ echo service-types ++ return 0 + jaeger_name=service-types + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test service-types true 01 + '[' 3 -ne 3 ']' + jaeger=service-types + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://service-types-query:443 + JAEGER_QUERY_ENDPOINT=https://service-types-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://service-types-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://service-types-collector-headless:14268 + export JAEGER_NAME=service-types + JAEGER_NAME=service-types + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-simple-prod + '[' 1 -ne 1 ']' + test_name=examples-simple-prod + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-simple-prod' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-simple-prod\e[0m' Rendering files for test examples-simple-prod + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-service-types + '[' examples-service-types '!=' _build ']' + cd .. + mkdir -p examples-simple-prod + cd examples-simple-prod + example_name=simple-prod + render_install_example simple-prod 01 + '[' 2 -ne 2 ']' + example_name=simple-prod + test_step=01 + install_file=./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/simple-prod.yaml -o ./01-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./01-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./01-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./01-install.yaml ++ jaeger_name=simple-prod ++ '[' -z simple-prod ']' ++ echo simple-prod ++ return 0 + JAEGER_NAME=simple-prod + local jaeger_strategy ++ get_jaeger_strategy ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./01-install.yaml ++ strategy=production ++ '[' production = production ']' ++ echo production ++ return 0 + jaeger_strategy=production + '[' production = DaemonSet ']' + '[' production = allInOne ']' + '[' production = production ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + [[ true = true ]] + [[ true = true ]] + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.options={}' ./01-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.elasticsearch={"nodeCount":1,"resources":{"limits":{"memory":"2Gi"}}}' ./01-install.yaml + render_smoke_test_example simple-prod 02 + '[' 2 -ne 2 ']' + example_name=simple-prod + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/simple-prod.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/simple-prod.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/simple-prod.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/simple-prod.yaml ++ jaeger_name=simple-prod ++ '[' -z simple-prod ']' ++ echo simple-prod ++ return 0 + jaeger_name=simple-prod + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test simple-prod true 02 + '[' 3 -ne 3 ']' + jaeger=simple-prod + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + export JAEGER_NAME=simple-prod + JAEGER_NAME=simple-prod + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-simple-prod-with-volumes + '[' 1 -ne 1 ']' + test_name=examples-simple-prod-with-volumes + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-simple-prod-with-volumes' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-simple-prod-with-volumes\e[0m' Rendering files for test examples-simple-prod-with-volumes + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-simple-prod + '[' examples-simple-prod '!=' _build ']' + cd .. + mkdir -p examples-simple-prod-with-volumes + cd examples-simple-prod-with-volumes + example_name=simple-prod-with-volumes + render_install_example simple-prod-with-volumes 01 + '[' 2 -ne 2 ']' + example_name=simple-prod-with-volumes + test_step=01 + install_file=./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/simple-prod-with-volumes.yaml -o ./01-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./01-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./01-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./01-install.yaml ++ jaeger_name=simple-prod ++ '[' -z simple-prod ']' ++ echo simple-prod ++ return 0 + JAEGER_NAME=simple-prod + local jaeger_strategy ++ get_jaeger_strategy ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./01-install.yaml ++ strategy=production ++ '[' production = production ']' ++ echo production ++ return 0 + jaeger_strategy=production + '[' production = DaemonSet ']' + '[' production = allInOne ']' + '[' production = production ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + [[ true = true ]] + [[ true = true ]] + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.options={}' ./01-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.elasticsearch={"nodeCount":1,"resources":{"limits":{"memory":"2Gi"}}}' ./01-install.yaml + render_smoke_test_example simple-prod-with-volumes 02 + '[' 2 -ne 2 ']' + example_name=simple-prod-with-volumes + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/simple-prod-with-volumes.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/simple-prod-with-volumes.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/simple-prod-with-volumes.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/simple-prod-with-volumes.yaml ++ jaeger_name=simple-prod ++ '[' -z simple-prod ']' ++ echo simple-prod ++ return 0 + jaeger_name=simple-prod + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test simple-prod true 02 + '[' 3 -ne 3 ']' + jaeger=simple-prod + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + export JAEGER_NAME=simple-prod + JAEGER_NAME=simple-prod + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + /tmp/jaeger-tests/bin/gomplate -f ./03-check-volume.yaml.template -o 03-check-volume.yaml + start_test examples-simplest + '[' 1 -ne 1 ']' + test_name=examples-simplest + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-simplest' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-simplest\e[0m' Rendering files for test examples-simplest + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-simple-prod-with-volumes + '[' examples-simple-prod-with-volumes '!=' _build ']' + cd .. + mkdir -p examples-simplest + cd examples-simplest + example_name=simplest + render_install_example simplest 00 + '[' 2 -ne 2 ']' + example_name=simplest + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/simplest.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=simplest ++ '[' -z simplest ']' ++ echo simplest ++ return 0 + JAEGER_NAME=simplest + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test_example simplest 01 + '[' 2 -ne 2 ']' + example_name=simplest + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/simplest.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/simplest.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/simplest.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/simplest.yaml ++ jaeger_name=simplest ++ '[' -z simplest ']' ++ echo simplest ++ return 0 + jaeger_name=simplest + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test simplest true 01 + '[' 3 -ne 3 ']' + jaeger=simplest + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simplest-query:443 + JAEGER_QUERY_ENDPOINT=https://simplest-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 + export JAEGER_NAME=simplest + JAEGER_NAME=simplest + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-with-badger + '[' 1 -ne 1 ']' + test_name=examples-with-badger + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-with-badger' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-with-badger\e[0m' Rendering files for test examples-with-badger + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-simplest + '[' examples-simplest '!=' _build ']' + cd .. + mkdir -p examples-with-badger + cd examples-with-badger + example_name=with-badger + render_install_example with-badger 00 + '[' 2 -ne 2 ']' + example_name=with-badger + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/with-badger.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=with-badger ++ '[' -z with-badger ']' ++ echo with-badger ++ return 0 + JAEGER_NAME=with-badger + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test_example with-badger 01 + '[' 2 -ne 2 ']' + example_name=with-badger + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/with-badger.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/with-badger.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/with-badger.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/with-badger.yaml ++ jaeger_name=with-badger ++ '[' -z with-badger ']' ++ echo with-badger ++ return 0 + jaeger_name=with-badger + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test with-badger true 01 + '[' 3 -ne 3 ']' + jaeger=with-badger + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://with-badger-query:443 + JAEGER_QUERY_ENDPOINT=https://with-badger-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://with-badger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://with-badger-collector-headless:14268 + export JAEGER_NAME=with-badger + JAEGER_NAME=with-badger + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-with-badger-and-volume + '[' 1 -ne 1 ']' + test_name=examples-with-badger-and-volume + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-with-badger-and-volume' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-with-badger-and-volume\e[0m' Rendering files for test examples-with-badger-and-volume + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-with-badger + '[' examples-with-badger '!=' _build ']' + cd .. + mkdir -p examples-with-badger-and-volume + cd examples-with-badger-and-volume + example_name=with-badger-and-volume + render_install_example with-badger-and-volume 00 + '[' 2 -ne 2 ']' + example_name=with-badger-and-volume + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/with-badger-and-volume.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=with-badger-and-volume ++ '[' -z with-badger-and-volume ']' ++ echo with-badger-and-volume ++ return 0 + JAEGER_NAME=with-badger-and-volume + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test_example with-badger-and-volume 01 + '[' 2 -ne 2 ']' + example_name=with-badger-and-volume + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/with-badger-and-volume.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/with-badger-and-volume.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/with-badger-and-volume.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/with-badger-and-volume.yaml ++ jaeger_name=with-badger-and-volume ++ '[' -z with-badger-and-volume ']' ++ echo with-badger-and-volume ++ return 0 + jaeger_name=with-badger-and-volume + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test with-badger-and-volume true 01 + '[' 3 -ne 3 ']' + jaeger=with-badger-and-volume + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://with-badger-and-volume-query:443 + JAEGER_QUERY_ENDPOINT=https://with-badger-and-volume-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://with-badger-and-volume-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://with-badger-and-volume-collector-headless:14268 + export JAEGER_NAME=with-badger-and-volume + JAEGER_NAME=with-badger-and-volume + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-with-cassandra + '[' 1 -ne 1 ']' + test_name=examples-with-cassandra + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-with-cassandra' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-with-cassandra\e[0m' Rendering files for test examples-with-cassandra + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-with-badger-and-volume + '[' examples-with-badger-and-volume '!=' _build ']' + cd .. + mkdir -p examples-with-cassandra + cd examples-with-cassandra + example_name=with-cassandra + render_install_cassandra 00 + '[' 1 -ne 1 ']' + test_step=00 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/cassandra-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/cassandra-assert.yaml.template -o ./00-assert.yaml + render_install_example with-cassandra 01 + '[' 2 -ne 2 ']' + example_name=with-cassandra + test_step=01 + install_file=./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/with-cassandra.yaml -o ./01-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./01-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./01-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./01-install.yaml ++ jaeger_name=with-cassandra ++ '[' -z with-cassandra ']' ++ echo with-cassandra ++ return 0 + JAEGER_NAME=with-cassandra + local jaeger_strategy ++ get_jaeger_strategy ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./01-install.yaml ++ strategy=allInOne ++ '[' allInOne = production ']' ++ '[' allInOne = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./01-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./01-assert.yaml + render_smoke_test_example with-cassandra 02 + '[' 2 -ne 2 ']' + example_name=with-cassandra + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/with-cassandra.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/with-cassandra.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/with-cassandra.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/with-cassandra.yaml ++ jaeger_name=with-cassandra ++ '[' -z with-cassandra ']' ++ echo with-cassandra ++ return 0 + jaeger_name=with-cassandra + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test with-cassandra true 02 + '[' 3 -ne 3 ']' + jaeger=with-cassandra + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://with-cassandra-query:443 + JAEGER_QUERY_ENDPOINT=https://with-cassandra-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://with-cassandra-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://with-cassandra-collector-headless:14268 + export JAEGER_NAME=with-cassandra + JAEGER_NAME=with-cassandra + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-with-sampling + '[' 1 -ne 1 ']' + test_name=examples-with-sampling + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-with-sampling' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-with-sampling\e[0m' Rendering files for test examples-with-sampling + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-with-cassandra + '[' examples-with-cassandra '!=' _build ']' + cd .. + mkdir -p examples-with-sampling + cd examples-with-sampling + export example_name=with-sampling + example_name=with-sampling + render_install_cassandra 00 + '[' 1 -ne 1 ']' + test_step=00 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/cassandra-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/cassandra-assert.yaml.template -o ./00-assert.yaml + render_install_example with-sampling 01 + '[' 2 -ne 2 ']' + example_name=with-sampling + test_step=01 + install_file=./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/with-sampling.yaml -o ./01-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./01-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./01-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./01-install.yaml ++ jaeger_name=with-sampling ++ '[' -z with-sampling ']' ++ echo with-sampling ++ return 0 + JAEGER_NAME=with-sampling + local jaeger_strategy ++ get_jaeger_strategy ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./01-install.yaml ++ strategy=allInOne ++ '[' allInOne = production ']' ++ '[' allInOne = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./01-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./01-assert.yaml + render_smoke_test_example with-sampling 02 + '[' 2 -ne 2 ']' + example_name=with-sampling + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/with-sampling.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/with-sampling.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/with-sampling.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/with-sampling.yaml ++ jaeger_name=with-sampling ++ '[' -z with-sampling ']' ++ echo with-sampling ++ return 0 + jaeger_name=with-sampling + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test with-sampling true 02 + '[' 3 -ne 3 ']' + jaeger=with-sampling + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://with-sampling-query:443 + JAEGER_QUERY_ENDPOINT=https://with-sampling-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://with-sampling-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://with-sampling-collector-headless:14268 + export JAEGER_NAME=with-sampling + JAEGER_NAME=with-sampling + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-agent-as-daemonset + '[' 1 -ne 1 ']' + test_name=examples-agent-as-daemonset + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-agent-as-daemonset' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-agent-as-daemonset\e[0m' Rendering files for test examples-agent-as-daemonset + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-with-sampling + '[' examples-with-sampling '!=' _build ']' + cd .. + mkdir -p examples-agent-as-daemonset + cd examples-agent-as-daemonset + '[' true = true ']' + prepare_daemonset 00 + '[' 1 -ne 1 ']' + test_step=00 + '[' true = true ']' + cat /tmp/jaeger-tests/examples/openshift/hostport-scc-daemonset.yaml + echo --- + cat /tmp/jaeger-tests/examples/openshift/service_account_jaeger-agent-daemonset.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/openshift/agent-as-daemonset.yaml -o 02-install.yaml + '[' true = true ']' + start_test examples-openshift-with-htpasswd + '[' 1 -ne 1 ']' + test_name=examples-openshift-with-htpasswd + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-openshift-with-htpasswd' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-openshift-with-htpasswd\e[0m' Rendering files for test examples-openshift-with-htpasswd + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-agent-as-daemonset + '[' examples-agent-as-daemonset '!=' _build ']' + cd .. + mkdir -p examples-openshift-with-htpasswd + cd examples-openshift-with-htpasswd + export JAEGER_NAME=with-htpasswd + JAEGER_NAME=with-htpasswd + export JAEGER_USERNAME=awesomeuser + JAEGER_USERNAME=awesomeuser + export JAEGER_PASSWORD=awesomepassword + JAEGER_PASSWORD=awesomepassword + export 'JAEGER_USER_PASSWORD_HASH=awesomeuser:{SHA}uUdqPVUyqNBmERU0Qxj3KFaZnjw=' + JAEGER_USER_PASSWORD_HASH='awesomeuser:{SHA}uUdqPVUyqNBmERU0Qxj3KFaZnjw=' ++ echo 'awesomeuser:{SHA}uUdqPVUyqNBmERU0Qxj3KFaZnjw=' ++ base64 + SECRET=YXdlc29tZXVzZXI6e1NIQX11VWRxUFZVeXFOQm1FUlUwUXhqM0tGYVpuanc9Cg== + /tmp/jaeger-tests/bin/gomplate -f ./00-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/openshift/with-htpasswd.yaml -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./01-assert.yaml + export 'GET_URL_COMMAND=kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE' + GET_URL_COMMAND='kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE' + export 'URL=https://$(kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE)/search' + URL='https://$(kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE)/search' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/ensure-ingress-host.sh.template -o ./ensure-ingress-host.sh + chmod +x ./ensure-ingress-host.sh + INSECURE=true + JAEGER_USERNAME= + JAEGER_PASSWORD= + EXPECTED_CODE=403 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./02-check-unsecured.yaml + JAEGER_USERNAME=wronguser + JAEGER_PASSWORD=wrongpassword + EXPECTED_CODE=403 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./03-check-unauthorized.yaml + EXPECTED_CODE=200 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./04-check-authorized.yaml + skip_test examples-agent-as-daemonset 'This test is flaky in Prow CI' + '[' 2 -ne 2 ']' + test_name=examples-agent-as-daemonset + message='This test is flaky in Prow CI' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-openshift-with-htpasswd + '[' examples-openshift-with-htpasswd '!=' _build ']' + cd .. + rm -rf examples-agent-as-daemonset + warning 'examples-agent-as-daemonset: This test is flaky in Prow CI' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: examples-agent-as-daemonset: This test is flaky in Prow CI\e[0m' WAR: examples-agent-as-daemonset: This test is flaky in Prow CI + skip_test examples-with-badger-and-volume 'This test is flaky in Prow CI' + '[' 2 -ne 2 ']' + test_name=examples-with-badger-and-volume + message='This test is flaky in Prow CI' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build + '[' _build '!=' _build ']' + rm -rf examples-with-badger-and-volume + warning 'examples-with-badger-and-volume: This test is flaky in Prow CI' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: examples-with-badger-and-volume: This test is flaky in Prow CI\e[0m' WAR: examples-with-badger-and-volume: This test is flaky in Prow CI + skip_test examples-collector-with-priority-class 'This test is flaky in Prow CI' + '[' 2 -ne 2 ']' + test_name=examples-collector-with-priority-class + message='This test is flaky in Prow CI' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build + '[' _build '!=' _build ']' + rm -rf examples-collector-with-priority-class + warning 'examples-collector-with-priority-class: This test is flaky in Prow CI' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: examples-collector-with-priority-class: This test is flaky in Prow CI\e[0m' WAR: examples-collector-with-priority-class: This test is flaky in Prow CI make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running examples E2E tests' Running examples E2E tests + cd tests/e2e/examples/_build + set +e + KUBECONFIG=/tmp/kubeconfig-2531963025 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-d102g9k0-f1a62.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 12 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/examples-agent-with-priority-class === PAUSE kuttl/harness/examples-agent-with-priority-class === RUN kuttl/harness/examples-all-in-one-with-options === PAUSE kuttl/harness/examples-all-in-one-with-options === RUN kuttl/harness/examples-business-application-injected-sidecar === PAUSE kuttl/harness/examples-business-application-injected-sidecar === RUN kuttl/harness/examples-openshift-with-htpasswd === PAUSE kuttl/harness/examples-openshift-with-htpasswd === RUN kuttl/harness/examples-service-types === PAUSE kuttl/harness/examples-service-types === RUN kuttl/harness/examples-simple-prod === PAUSE kuttl/harness/examples-simple-prod === RUN kuttl/harness/examples-simple-prod-with-volumes === PAUSE kuttl/harness/examples-simple-prod-with-volumes === RUN kuttl/harness/examples-simplest === PAUSE kuttl/harness/examples-simplest === RUN kuttl/harness/examples-with-badger === PAUSE kuttl/harness/examples-with-badger === RUN kuttl/harness/examples-with-cassandra === PAUSE kuttl/harness/examples-with-cassandra === RUN kuttl/harness/examples-with-sampling === PAUSE kuttl/harness/examples-with-sampling === CONT kuttl/harness/artifacts logger.go:42: 07:38:57 | artifacts | Creating namespace: kuttl-test-capital-impala logger.go:42: 07:38:57 | artifacts | artifacts events from ns kuttl-test-capital-impala: logger.go:42: 07:38:57 | artifacts | Deleting namespace: kuttl-test-capital-impala === CONT kuttl/harness/examples-simple-prod logger.go:42: 07:39:03 | examples-simple-prod | Creating namespace: kuttl-test-settling-kid logger.go:42: 07:39:03 | examples-simple-prod/1-install | starting test step 1-install logger.go:42: 07:39:03 | examples-simple-prod/1-install | Jaeger:kuttl-test-settling-kid/simple-prod created logger.go:42: 07:39:50 | examples-simple-prod/1-install | test step completed 1-install logger.go:42: 07:39:50 | examples-simple-prod/2-smoke-test | starting test step 2-smoke-test logger.go:42: 07:39:50 | examples-simple-prod/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE simple-prod /dev/null] logger.go:42: 07:39:51 | examples-simple-prod/2-smoke-test | Warning: resource jaegers/simple-prod is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:39:57 | examples-simple-prod/2-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:39:57 | examples-simple-prod/2-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:39:58 | examples-simple-prod/2-smoke-test | job.batch/report-span created logger.go:42: 07:39:58 | examples-simple-prod/2-smoke-test | job.batch/check-span created logger.go:42: 07:40:11 | examples-simple-prod/2-smoke-test | test step completed 2-smoke-test logger.go:42: 07:40:11 | examples-simple-prod | examples-simple-prod events from ns kuttl-test-settling-kid: logger.go:42: 07:40:11 | examples-simple-prod | 2025-02-03 07:39:17 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsettlingkidsimpleprod-1-c76b8bf52gbs Binding Scheduled Successfully assigned kuttl-test-settling-kid/elasticsearch-cdm-kuttltestsettlingkidsimpleprod-1-c76b8bf52gbs to ip-10-0-6-172.us-east-2.compute.internal default-scheduler logger.go:42: 07:40:11 | examples-simple-prod | 2025-02-03 07:39:17 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestsettlingkidsimpleprod-1-c76b8bf65 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestsettlingkidsimpleprod-1-c76b8bf52gbs replicaset-controller logger.go:42: 07:40:11 | examples-simple-prod | 2025-02-03 07:39:17 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestsettlingkidsimpleprod-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestsettlingkidsimpleprod-1-c76b8bf65 to 1 deployment-controller logger.go:42: 07:40:11 | examples-simple-prod | 2025-02-03 07:39:19 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestsettlingkidsimpleprod-1-c76b8bf52gbs FailedMount MountVolume.SetUp failed for volume "elasticsearch-metrics" : failed to sync secret cache: timed out waiting for the condition kubelet logger.go:42: 07:40:11 | examples-simple-prod | 2025-02-03 07:39:19 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestsettlingkidsimpleprod-1-c76b8bf52gbs FailedMount MountVolume.SetUp failed for volume "elasticsearch-config" : failed to sync configmap cache: timed out waiting for the condition kubelet logger.go:42: 07:40:11 | examples-simple-prod | 2025-02-03 07:39:20 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsettlingkidsimpleprod-1-c76b8bf52gbs AddedInterface Add eth0 [10.131.0.31/23] from ovn-kubernetes multus logger.go:42: 07:40:11 | examples-simple-prod | 2025-02-03 07:39:20 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsettlingkidsimpleprod-1-c76b8bf52gbs.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:410674bf2f77ff1d431f494cadfab3e81249b35695eb65aedfdb8498079b4d35" already present on machine kubelet logger.go:42: 07:40:11 | examples-simple-prod | 2025-02-03 07:39:20 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsettlingkidsimpleprod-1-c76b8bf52gbs.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:40:11 | examples-simple-prod | 2025-02-03 07:39:20 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsettlingkidsimpleprod-1-c76b8bf52gbs.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:40:11 | examples-simple-prod | 2025-02-03 07:39:20 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsettlingkidsimpleprod-1-c76b8bf52gbs.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:fe2439e5f9e14d828caace425214edfa3edc179d72d225c43f69e9e4c9a5837b" already present on machine kubelet logger.go:42: 07:40:11 | examples-simple-prod | 2025-02-03 07:39:20 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsettlingkidsimpleprod-1-c76b8bf52gbs.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:40:11 | examples-simple-prod | 2025-02-03 07:39:20 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsettlingkidsimpleprod-1-c76b8bf52gbs.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:40:11 | examples-simple-prod | 2025-02-03 07:39:35 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestsettlingkidsimpleprod-1-c76b8bf52gbs.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:40:11 | examples-simple-prod | 2025-02-03 07:39:46 +0000 UTC Normal Pod simple-prod-collector-5795c5f558-dctr7 Binding Scheduled Successfully assigned kuttl-test-settling-kid/simple-prod-collector-5795c5f558-dctr7 to ip-10-0-66-98.us-east-2.compute.internal default-scheduler logger.go:42: 07:40:11 | examples-simple-prod | 2025-02-03 07:39:46 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-5795c5f558 SuccessfulCreate Created pod: simple-prod-collector-5795c5f558-dctr7 replicaset-controller logger.go:42: 07:40:11 | examples-simple-prod | 2025-02-03 07:39:46 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-5795c5f558 to 1 deployment-controller logger.go:42: 07:40:11 | examples-simple-prod | 2025-02-03 07:39:46 +0000 UTC Normal Pod simple-prod-query-5cd5475b84-zkq4g Binding Scheduled Successfully assigned kuttl-test-settling-kid/simple-prod-query-5cd5475b84-zkq4g to ip-10-0-26-111.us-east-2.compute.internal default-scheduler logger.go:42: 07:40:11 | examples-simple-prod | 2025-02-03 07:39:46 +0000 UTC Normal ReplicaSet.apps simple-prod-query-5cd5475b84 SuccessfulCreate Created pod: simple-prod-query-5cd5475b84-zkq4g replicaset-controller logger.go:42: 07:40:11 | examples-simple-prod | 2025-02-03 07:39:46 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-5cd5475b84 to 1 deployment-controller logger.go:42: 07:40:11 | examples-simple-prod | 2025-02-03 07:39:47 +0000 UTC Normal Pod simple-prod-collector-5795c5f558-dctr7 AddedInterface Add eth0 [10.129.2.36/23] from ovn-kubernetes multus logger.go:42: 07:40:11 | examples-simple-prod | 2025-02-03 07:39:47 +0000 UTC Normal Pod simple-prod-collector-5795c5f558-dctr7.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:40a1c7fa23aa5ffa64e8e22aa38022f5d4d7ff644c46a3da7169b713d486c3c1" already present on machine kubelet logger.go:42: 07:40:11 | examples-simple-prod | 2025-02-03 07:39:47 +0000 UTC Normal Pod simple-prod-collector-5795c5f558-dctr7.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:40:11 | examples-simple-prod | 2025-02-03 07:39:47 +0000 UTC Normal Pod simple-prod-collector-5795c5f558-dctr7.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:40:11 | examples-simple-prod | 2025-02-03 07:39:47 +0000 UTC Normal Pod simple-prod-query-5cd5475b84-zkq4g AddedInterface Add eth0 [10.128.2.52/23] from ovn-kubernetes multus logger.go:42: 07:40:11 | examples-simple-prod | 2025-02-03 07:39:47 +0000 UTC Normal Pod simple-prod-query-5cd5475b84-zkq4g.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:84020ce29bcb5a4bd018e6596188ae919c5cd600e08f78a546c0e76ea477685e" already present on machine kubelet logger.go:42: 07:40:11 | examples-simple-prod | 2025-02-03 07:39:48 +0000 UTC Normal Pod simple-prod-query-5cd5475b84-zkq4g.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:40:11 | examples-simple-prod | 2025-02-03 07:39:48 +0000 UTC Normal Pod simple-prod-query-5cd5475b84-zkq4g.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:40:11 | examples-simple-prod | 2025-02-03 07:39:48 +0000 UTC Normal Pod simple-prod-query-5cd5475b84-zkq4g.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" already present on machine kubelet logger.go:42: 07:40:11 | examples-simple-prod | 2025-02-03 07:39:48 +0000 UTC Normal Pod simple-prod-query-5cd5475b84-zkq4g.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:40:11 | examples-simple-prod | 2025-02-03 07:39:48 +0000 UTC Normal Pod simple-prod-query-5cd5475b84-zkq4g.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:40:11 | examples-simple-prod | 2025-02-03 07:39:48 +0000 UTC Normal Pod simple-prod-query-5cd5475b84-zkq4g.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:65342c7d622f336741a8ae0b9dead79c7ecabc155d02cdd42b7b49ca36680e74" already present on machine kubelet logger.go:42: 07:40:11 | examples-simple-prod | 2025-02-03 07:39:48 +0000 UTC Normal Pod simple-prod-query-5cd5475b84-zkq4g.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:40:11 | examples-simple-prod | 2025-02-03 07:39:48 +0000 UTC Normal Pod simple-prod-query-5cd5475b84-zkq4g.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:40:11 | examples-simple-prod | 2025-02-03 07:39:52 +0000 UTC Normal ReplicaSet.apps simple-prod-query-5cd5475b84 SuccessfulDelete Deleted pod: simple-prod-query-5cd5475b84-zkq4g replicaset-controller logger.go:42: 07:40:11 | examples-simple-prod | 2025-02-03 07:39:52 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled down replica set simple-prod-query-5cd5475b84 to 0 from 1 deployment-controller logger.go:42: 07:40:11 | examples-simple-prod | 2025-02-03 07:39:53 +0000 UTC Normal Pod simple-prod-query-5cd5475b84-zkq4g.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:40:11 | examples-simple-prod | 2025-02-03 07:39:53 +0000 UTC Normal Pod simple-prod-query-5cd5475b84-zkq4g.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:40:11 | examples-simple-prod | 2025-02-03 07:39:53 +0000 UTC Normal Pod simple-prod-query-5cd5475b84-zkq4g.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:40:11 | examples-simple-prod | 2025-02-03 07:39:54 +0000 UTC Normal Pod simple-prod-query-d4478bc87-px4w2 Binding Scheduled Successfully assigned kuttl-test-settling-kid/simple-prod-query-d4478bc87-px4w2 to ip-10-0-26-111.us-east-2.compute.internal default-scheduler logger.go:42: 07:40:11 | examples-simple-prod | 2025-02-03 07:39:54 +0000 UTC Normal ReplicaSet.apps simple-prod-query-d4478bc87 SuccessfulCreate Created pod: simple-prod-query-d4478bc87-px4w2 replicaset-controller logger.go:42: 07:40:11 | examples-simple-prod | 2025-02-03 07:39:54 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-d4478bc87 to 1 deployment-controller logger.go:42: 07:40:11 | examples-simple-prod | 2025-02-03 07:39:55 +0000 UTC Normal Pod simple-prod-query-d4478bc87-px4w2 AddedInterface Add eth0 [10.128.2.53/23] from ovn-kubernetes multus logger.go:42: 07:40:11 | examples-simple-prod | 2025-02-03 07:39:55 +0000 UTC Normal Pod simple-prod-query-d4478bc87-px4w2.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:84020ce29bcb5a4bd018e6596188ae919c5cd600e08f78a546c0e76ea477685e" already present on machine kubelet logger.go:42: 07:40:11 | examples-simple-prod | 2025-02-03 07:39:55 +0000 UTC Normal Pod simple-prod-query-d4478bc87-px4w2.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:40:11 | examples-simple-prod | 2025-02-03 07:39:55 +0000 UTC Normal Pod simple-prod-query-d4478bc87-px4w2.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:40:11 | examples-simple-prod | 2025-02-03 07:39:55 +0000 UTC Normal Pod simple-prod-query-d4478bc87-px4w2.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" already present on machine kubelet logger.go:42: 07:40:11 | examples-simple-prod | 2025-02-03 07:39:55 +0000 UTC Normal Pod simple-prod-query-d4478bc87-px4w2.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:40:11 | examples-simple-prod | 2025-02-03 07:39:55 +0000 UTC Normal Pod simple-prod-query-d4478bc87-px4w2.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:40:11 | examples-simple-prod | 2025-02-03 07:39:55 +0000 UTC Normal Pod simple-prod-query-d4478bc87-px4w2.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:65342c7d622f336741a8ae0b9dead79c7ecabc155d02cdd42b7b49ca36680e74" already present on machine kubelet logger.go:42: 07:40:11 | examples-simple-prod | 2025-02-03 07:39:55 +0000 UTC Normal Pod simple-prod-query-d4478bc87-px4w2.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:40:11 | examples-simple-prod | 2025-02-03 07:39:55 +0000 UTC Normal Pod simple-prod-query-d4478bc87-px4w2.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:40:11 | examples-simple-prod | 2025-02-03 07:39:58 +0000 UTC Normal Pod check-span-58l9j Binding Scheduled Successfully assigned kuttl-test-settling-kid/check-span-58l9j to ip-10-0-66-98.us-east-2.compute.internal default-scheduler logger.go:42: 07:40:11 | examples-simple-prod | 2025-02-03 07:39:58 +0000 UTC Normal Pod check-span-58l9j AddedInterface Add eth0 [10.129.2.38/23] from ovn-kubernetes multus logger.go:42: 07:40:11 | examples-simple-prod | 2025-02-03 07:39:58 +0000 UTC Normal Pod check-span-58l9j.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:40:11 | examples-simple-prod | 2025-02-03 07:39:58 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-58l9j job-controller logger.go:42: 07:40:11 | examples-simple-prod | 2025-02-03 07:39:58 +0000 UTC Normal Pod report-span-9dhdd Binding Scheduled Successfully assigned kuttl-test-settling-kid/report-span-9dhdd to ip-10-0-66-98.us-east-2.compute.internal default-scheduler logger.go:42: 07:40:11 | examples-simple-prod | 2025-02-03 07:39:58 +0000 UTC Normal Pod report-span-9dhdd AddedInterface Add eth0 [10.129.2.37/23] from ovn-kubernetes multus logger.go:42: 07:40:11 | examples-simple-prod | 2025-02-03 07:39:58 +0000 UTC Normal Pod report-span-9dhdd.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:40:11 | examples-simple-prod | 2025-02-03 07:39:58 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-9dhdd job-controller logger.go:42: 07:40:11 | examples-simple-prod | 2025-02-03 07:39:59 +0000 UTC Normal Pod check-span-58l9j.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 664ms (664ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:40:11 | examples-simple-prod | 2025-02-03 07:39:59 +0000 UTC Normal Pod check-span-58l9j.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:40:11 | examples-simple-prod | 2025-02-03 07:39:59 +0000 UTC Normal Pod check-span-58l9j.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:40:11 | examples-simple-prod | 2025-02-03 07:39:59 +0000 UTC Normal Pod report-span-9dhdd.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 657ms (657ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:40:11 | examples-simple-prod | 2025-02-03 07:39:59 +0000 UTC Normal Pod report-span-9dhdd.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:40:11 | examples-simple-prod | 2025-02-03 07:39:59 +0000 UTC Normal Pod report-span-9dhdd.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:40:11 | examples-simple-prod | 2025-02-03 07:40:02 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:40:11 | examples-simple-prod | 2025-02-03 07:40:02 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:40:11 | examples-simple-prod | 2025-02-03 07:40:02 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:40:11 | examples-simple-prod | 2025-02-03 07:40:10 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:40:11 | examples-simple-prod | Deleting namespace: kuttl-test-settling-kid === CONT kuttl/harness/examples-with-sampling logger.go:42: 07:40:24 | examples-with-sampling | Creating namespace: kuttl-test-big-manatee logger.go:42: 07:40:24 | examples-with-sampling/0-install | starting test step 0-install logger.go:42: 07:40:24 | examples-with-sampling/0-install | running command: [sh -c cd /tmp/jaeger-tests && make cassandra STORAGE_NAMESPACE=$NAMESPACE] logger.go:42: 07:40:24 | examples-with-sampling/0-install | make[2]: Entering directory '/tmp/jaeger-tests' logger.go:42: 07:40:24 | examples-with-sampling/0-install | >>>> Creating namespace kuttl-test-big-manatee logger.go:42: 07:40:24 | examples-with-sampling/0-install | kubectl create namespace kuttl-test-big-manatee 2>&1 | grep -v "already exists" || true logger.go:42: 07:40:24 | examples-with-sampling/0-install | kubectl create -f ./tests/cassandra.yml --namespace kuttl-test-big-manatee 2>&1 | grep -v "already exists" || true logger.go:42: 07:40:24 | examples-with-sampling/0-install | service/cassandra created logger.go:42: 07:40:24 | examples-with-sampling/0-install | statefulset.apps/cassandra created logger.go:42: 07:40:24 | examples-with-sampling/0-install | make[2]: Leaving directory '/tmp/jaeger-tests' logger.go:42: 07:40:36 | examples-with-sampling/0-install | test step completed 0-install logger.go:42: 07:40:36 | examples-with-sampling/1-install | starting test step 1-install logger.go:42: 07:40:36 | examples-with-sampling/1-install | Jaeger:kuttl-test-big-manatee/with-sampling created logger.go:42: 07:40:43 | examples-with-sampling/1-install | test step completed 1-install logger.go:42: 07:40:43 | examples-with-sampling/2-smoke-test | starting test step 2-smoke-test logger.go:42: 07:40:43 | examples-with-sampling/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE with-sampling /dev/null] logger.go:42: 07:40:44 | examples-with-sampling/2-smoke-test | Warning: resource jaegers/with-sampling is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:40:50 | examples-with-sampling/2-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://with-sampling-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://with-sampling-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:40:50 | examples-with-sampling/2-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:40:51 | examples-with-sampling/2-smoke-test | job.batch/report-span created logger.go:42: 07:40:51 | examples-with-sampling/2-smoke-test | job.batch/check-span created logger.go:42: 07:41:04 | examples-with-sampling/2-smoke-test | test step completed 2-smoke-test logger.go:42: 07:41:04 | examples-with-sampling/3- | starting test step 3- logger.go:42: 07:41:04 | examples-with-sampling/3- | test step completed 3- logger.go:42: 07:41:04 | examples-with-sampling | examples-with-sampling events from ns kuttl-test-big-manatee: logger.go:42: 07:41:04 | examples-with-sampling | 2025-02-03 07:40:24 +0000 UTC Normal Pod cassandra-0 Binding Scheduled Successfully assigned kuttl-test-big-manatee/cassandra-0 to ip-10-0-6-172.us-east-2.compute.internal default-scheduler logger.go:42: 07:41:04 | examples-with-sampling | 2025-02-03 07:40:24 +0000 UTC Normal StatefulSet.apps cassandra SuccessfulCreate create Pod cassandra-0 in StatefulSet cassandra successful statefulset-controller logger.go:42: 07:41:04 | examples-with-sampling | 2025-02-03 07:40:25 +0000 UTC Normal Pod cassandra-0 AddedInterface Add eth0 [10.131.0.32/23] from ovn-kubernetes multus logger.go:42: 07:41:04 | examples-with-sampling | 2025-02-03 07:40:25 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Pulling Pulling image "cassandra:3.11" kubelet logger.go:42: 07:41:04 | examples-with-sampling | 2025-02-03 07:40:29 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Pulled Successfully pulled image "cassandra:3.11" in 4.166s (4.166s including waiting). Image size: 309833663 bytes. kubelet logger.go:42: 07:41:04 | examples-with-sampling | 2025-02-03 07:40:29 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Created Created container cassandra kubelet logger.go:42: 07:41:04 | examples-with-sampling | 2025-02-03 07:40:30 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Started Started container cassandra kubelet logger.go:42: 07:41:04 | examples-with-sampling | 2025-02-03 07:40:30 +0000 UTC Normal Pod cassandra-1 Binding Scheduled Successfully assigned kuttl-test-big-manatee/cassandra-1 to ip-10-0-66-98.us-east-2.compute.internal default-scheduler logger.go:42: 07:41:04 | examples-with-sampling | 2025-02-03 07:40:30 +0000 UTC Normal Pod cassandra-1 AddedInterface Add eth0 [10.129.2.39/23] from ovn-kubernetes multus logger.go:42: 07:41:04 | examples-with-sampling | 2025-02-03 07:40:30 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Pulling Pulling image "cassandra:3.11" kubelet logger.go:42: 07:41:04 | examples-with-sampling | 2025-02-03 07:40:30 +0000 UTC Normal StatefulSet.apps cassandra SuccessfulCreate create Pod cassandra-1 in StatefulSet cassandra successful statefulset-controller logger.go:42: 07:41:04 | examples-with-sampling | 2025-02-03 07:40:35 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Pulled Successfully pulled image "cassandra:3.11" in 4.053s (4.053s including waiting). Image size: 309833663 bytes. kubelet logger.go:42: 07:41:04 | examples-with-sampling | 2025-02-03 07:40:35 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Created Created container cassandra kubelet logger.go:42: 07:41:04 | examples-with-sampling | 2025-02-03 07:40:35 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Started Started container cassandra kubelet logger.go:42: 07:41:04 | examples-with-sampling | 2025-02-03 07:40:39 +0000 UTC Normal ReplicaSet.apps with-sampling-6d85f5d4f8 SuccessfulCreate Created pod: with-sampling-6d85f5d4f8-2zzm9 replicaset-controller logger.go:42: 07:41:04 | examples-with-sampling | 2025-02-03 07:40:39 +0000 UTC Normal Deployment.apps with-sampling ScalingReplicaSet Scaled up replica set with-sampling-6d85f5d4f8 to 1 deployment-controller logger.go:42: 07:41:04 | examples-with-sampling | 2025-02-03 07:40:40 +0000 UTC Normal Pod with-sampling-6d85f5d4f8-2zzm9 Binding Scheduled Successfully assigned kuttl-test-big-manatee/with-sampling-6d85f5d4f8-2zzm9 to ip-10-0-66-98.us-east-2.compute.internal default-scheduler logger.go:42: 07:41:04 | examples-with-sampling | 2025-02-03 07:40:40 +0000 UTC Warning Pod with-sampling-6d85f5d4f8-2zzm9 FailedMount MountVolume.SetUp failed for volume "with-sampling-ui-oauth-proxy-tls" : secret "with-sampling-ui-oauth-proxy-tls" not found kubelet logger.go:42: 07:41:04 | examples-with-sampling | 2025-02-03 07:40:40 +0000 UTC Warning Pod with-sampling-6d85f5d4f8-2zzm9 FailedMount MountVolume.SetUp failed for volume "with-sampling-collector-tls-config-volume" : secret "with-sampling-collector-headless-tls" not found kubelet logger.go:42: 07:41:04 | examples-with-sampling | 2025-02-03 07:40:41 +0000 UTC Normal Pod with-sampling-6d85f5d4f8-2zzm9 AddedInterface Add eth0 [10.129.2.40/23] from ovn-kubernetes multus logger.go:42: 07:41:04 | examples-with-sampling | 2025-02-03 07:40:41 +0000 UTC Normal Pod with-sampling-6d85f5d4f8-2zzm9.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:fa9aaf9ae026228265ce7f10fb451d5c52a4e0269ce1fa9024cbefddf765ae8d" already present on machine kubelet logger.go:42: 07:41:04 | examples-with-sampling | 2025-02-03 07:40:41 +0000 UTC Normal Pod with-sampling-6d85f5d4f8-2zzm9.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:41:04 | examples-with-sampling | 2025-02-03 07:40:41 +0000 UTC Normal Pod with-sampling-6d85f5d4f8-2zzm9.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:41:04 | examples-with-sampling | 2025-02-03 07:40:41 +0000 UTC Normal Pod with-sampling-6d85f5d4f8-2zzm9.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" already present on machine kubelet logger.go:42: 07:41:04 | examples-with-sampling | 2025-02-03 07:40:41 +0000 UTC Normal Pod with-sampling-6d85f5d4f8-2zzm9.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:41:04 | examples-with-sampling | 2025-02-03 07:40:41 +0000 UTC Normal Pod with-sampling-6d85f5d4f8-2zzm9.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:41:04 | examples-with-sampling | 2025-02-03 07:40:45 +0000 UTC Normal Pod with-sampling-6d85f5d4f8-2zzm9.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:41:04 | examples-with-sampling | 2025-02-03 07:40:45 +0000 UTC Normal Pod with-sampling-6d85f5d4f8-2zzm9.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:41:04 | examples-with-sampling | 2025-02-03 07:40:45 +0000 UTC Normal ReplicaSet.apps with-sampling-6d85f5d4f8 SuccessfulDelete Deleted pod: with-sampling-6d85f5d4f8-2zzm9 replicaset-controller logger.go:42: 07:41:04 | examples-with-sampling | 2025-02-03 07:40:45 +0000 UTC Normal Deployment.apps with-sampling ScalingReplicaSet Scaled down replica set with-sampling-6d85f5d4f8 to 0 from 1 deployment-controller logger.go:42: 07:41:04 | examples-with-sampling | 2025-02-03 07:40:46 +0000 UTC Normal Pod with-sampling-7d54596bbc-wtxsh Binding Scheduled Successfully assigned kuttl-test-big-manatee/with-sampling-7d54596bbc-wtxsh to ip-10-0-66-98.us-east-2.compute.internal default-scheduler logger.go:42: 07:41:04 | examples-with-sampling | 2025-02-03 07:40:46 +0000 UTC Normal ReplicaSet.apps with-sampling-7d54596bbc SuccessfulCreate Created pod: with-sampling-7d54596bbc-wtxsh replicaset-controller logger.go:42: 07:41:04 | examples-with-sampling | 2025-02-03 07:40:46 +0000 UTC Normal Deployment.apps with-sampling ScalingReplicaSet Scaled up replica set with-sampling-7d54596bbc to 1 deployment-controller logger.go:42: 07:41:04 | examples-with-sampling | 2025-02-03 07:40:47 +0000 UTC Normal Pod with-sampling-7d54596bbc-wtxsh AddedInterface Add eth0 [10.129.2.41/23] from ovn-kubernetes multus logger.go:42: 07:41:04 | examples-with-sampling | 2025-02-03 07:40:47 +0000 UTC Normal Pod with-sampling-7d54596bbc-wtxsh.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:fa9aaf9ae026228265ce7f10fb451d5c52a4e0269ce1fa9024cbefddf765ae8d" already present on machine kubelet logger.go:42: 07:41:04 | examples-with-sampling | 2025-02-03 07:40:47 +0000 UTC Normal Pod with-sampling-7d54596bbc-wtxsh.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:41:04 | examples-with-sampling | 2025-02-03 07:40:47 +0000 UTC Normal Pod with-sampling-7d54596bbc-wtxsh.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:41:04 | examples-with-sampling | 2025-02-03 07:40:47 +0000 UTC Normal Pod with-sampling-7d54596bbc-wtxsh.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" already present on machine kubelet logger.go:42: 07:41:04 | examples-with-sampling | 2025-02-03 07:40:47 +0000 UTC Normal Pod with-sampling-7d54596bbc-wtxsh.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:41:04 | examples-with-sampling | 2025-02-03 07:40:47 +0000 UTC Normal Pod with-sampling-7d54596bbc-wtxsh.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:41:04 | examples-with-sampling | 2025-02-03 07:40:51 +0000 UTC Normal Pod check-span-4ws2p Binding Scheduled Successfully assigned kuttl-test-big-manatee/check-span-4ws2p to ip-10-0-6-172.us-east-2.compute.internal default-scheduler logger.go:42: 07:41:04 | examples-with-sampling | 2025-02-03 07:40:51 +0000 UTC Normal Pod check-span-4ws2p AddedInterface Add eth0 [10.131.0.33/23] from ovn-kubernetes multus logger.go:42: 07:41:04 | examples-with-sampling | 2025-02-03 07:40:51 +0000 UTC Normal Pod check-span-4ws2p.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:41:04 | examples-with-sampling | 2025-02-03 07:40:51 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-4ws2p job-controller logger.go:42: 07:41:04 | examples-with-sampling | 2025-02-03 07:40:51 +0000 UTC Normal Pod report-span-lhjl5 Binding Scheduled Successfully assigned kuttl-test-big-manatee/report-span-lhjl5 to ip-10-0-26-111.us-east-2.compute.internal default-scheduler logger.go:42: 07:41:04 | examples-with-sampling | 2025-02-03 07:40:51 +0000 UTC Normal Pod report-span-lhjl5 AddedInterface Add eth0 [10.128.2.54/23] from ovn-kubernetes multus logger.go:42: 07:41:04 | examples-with-sampling | 2025-02-03 07:40:51 +0000 UTC Normal Pod report-span-lhjl5.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:41:04 | examples-with-sampling | 2025-02-03 07:40:51 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-lhjl5 job-controller logger.go:42: 07:41:04 | examples-with-sampling | 2025-02-03 07:40:52 +0000 UTC Normal Pod check-span-4ws2p.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 653ms (653ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:41:04 | examples-with-sampling | 2025-02-03 07:40:52 +0000 UTC Normal Pod check-span-4ws2p.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:41:04 | examples-with-sampling | 2025-02-03 07:40:52 +0000 UTC Normal Pod check-span-4ws2p.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:41:04 | examples-with-sampling | 2025-02-03 07:40:52 +0000 UTC Normal Pod report-span-lhjl5.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 576ms (576ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:41:04 | examples-with-sampling | 2025-02-03 07:40:52 +0000 UTC Normal Pod report-span-lhjl5.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:41:04 | examples-with-sampling | 2025-02-03 07:40:52 +0000 UTC Normal Pod report-span-lhjl5.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:41:04 | examples-with-sampling | 2025-02-03 07:41:03 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:41:04 | examples-with-sampling | Deleting namespace: kuttl-test-big-manatee === CONT kuttl/harness/examples-with-cassandra logger.go:42: 07:41:23 | examples-with-cassandra | Creating namespace: kuttl-test-pure-caribou logger.go:42: 07:41:23 | examples-with-cassandra/0-install | starting test step 0-install logger.go:42: 07:41:23 | examples-with-cassandra/0-install | running command: [sh -c cd /tmp/jaeger-tests && make cassandra STORAGE_NAMESPACE=$NAMESPACE] logger.go:42: 07:41:23 | examples-with-cassandra/0-install | make[2]: Entering directory '/tmp/jaeger-tests' logger.go:42: 07:41:23 | examples-with-cassandra/0-install | >>>> Creating namespace kuttl-test-pure-caribou logger.go:42: 07:41:23 | examples-with-cassandra/0-install | kubectl create namespace kuttl-test-pure-caribou 2>&1 | grep -v "already exists" || true logger.go:42: 07:41:23 | examples-with-cassandra/0-install | kubectl create -f ./tests/cassandra.yml --namespace kuttl-test-pure-caribou 2>&1 | grep -v "already exists" || true logger.go:42: 07:41:23 | examples-with-cassandra/0-install | service/cassandra created logger.go:42: 07:41:23 | examples-with-cassandra/0-install | statefulset.apps/cassandra created logger.go:42: 07:41:23 | examples-with-cassandra/0-install | make[2]: Leaving directory '/tmp/jaeger-tests' logger.go:42: 07:41:26 | examples-with-cassandra/0-install | test step completed 0-install logger.go:42: 07:41:26 | examples-with-cassandra/1-install | starting test step 1-install logger.go:42: 07:41:27 | examples-with-cassandra/1-install | Jaeger:kuttl-test-pure-caribou/with-cassandra created logger.go:42: 07:41:45 | examples-with-cassandra/1-install | test step completed 1-install logger.go:42: 07:41:45 | examples-with-cassandra/2-smoke-test | starting test step 2-smoke-test logger.go:42: 07:41:45 | examples-with-cassandra/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE with-cassandra /dev/null] logger.go:42: 07:41:46 | examples-with-cassandra/2-smoke-test | Warning: resource jaegers/with-cassandra is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:41:52 | examples-with-cassandra/2-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://with-cassandra-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://with-cassandra-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:41:52 | examples-with-cassandra/2-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:41:53 | examples-with-cassandra/2-smoke-test | job.batch/report-span created logger.go:42: 07:41:53 | examples-with-cassandra/2-smoke-test | job.batch/check-span created logger.go:42: 07:42:06 | examples-with-cassandra/2-smoke-test | test step completed 2-smoke-test logger.go:42: 07:42:06 | examples-with-cassandra | examples-with-cassandra events from ns kuttl-test-pure-caribou: logger.go:42: 07:42:06 | examples-with-cassandra | 2025-02-03 07:41:23 +0000 UTC Normal StatefulSet.apps cassandra SuccessfulCreate create Pod cassandra-0 in StatefulSet cassandra successful statefulset-controller logger.go:42: 07:42:06 | examples-with-cassandra | 2025-02-03 07:41:24 +0000 UTC Normal Pod cassandra-0 Binding Scheduled Successfully assigned kuttl-test-pure-caribou/cassandra-0 to ip-10-0-6-172.us-east-2.compute.internal default-scheduler logger.go:42: 07:42:06 | examples-with-cassandra | 2025-02-03 07:41:24 +0000 UTC Normal Pod cassandra-0 AddedInterface Add eth0 [10.131.0.34/23] from ovn-kubernetes multus logger.go:42: 07:42:06 | examples-with-cassandra | 2025-02-03 07:41:24 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Pulled Container image "cassandra:3.11" already present on machine kubelet logger.go:42: 07:42:06 | examples-with-cassandra | 2025-02-03 07:41:24 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Created Created container cassandra kubelet logger.go:42: 07:42:06 | examples-with-cassandra | 2025-02-03 07:41:24 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Started Started container cassandra kubelet logger.go:42: 07:42:06 | examples-with-cassandra | 2025-02-03 07:41:25 +0000 UTC Normal Pod cassandra-1 Binding Scheduled Successfully assigned kuttl-test-pure-caribou/cassandra-1 to ip-10-0-66-98.us-east-2.compute.internal default-scheduler logger.go:42: 07:42:06 | examples-with-cassandra | 2025-02-03 07:41:25 +0000 UTC Normal StatefulSet.apps cassandra SuccessfulCreate create Pod cassandra-1 in StatefulSet cassandra successful statefulset-controller logger.go:42: 07:42:06 | examples-with-cassandra | 2025-02-03 07:41:26 +0000 UTC Normal Pod cassandra-1 AddedInterface Add eth0 [10.129.2.42/23] from ovn-kubernetes multus logger.go:42: 07:42:06 | examples-with-cassandra | 2025-02-03 07:41:26 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Pulled Container image "cassandra:3.11" already present on machine kubelet logger.go:42: 07:42:06 | examples-with-cassandra | 2025-02-03 07:41:26 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Created Created container cassandra kubelet logger.go:42: 07:42:06 | examples-with-cassandra | 2025-02-03 07:41:26 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Started Started container cassandra kubelet logger.go:42: 07:42:06 | examples-with-cassandra | 2025-02-03 07:41:30 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-brlvr Binding Scheduled Successfully assigned kuttl-test-pure-caribou/with-cassandra-cassandra-schema-job-brlvr to ip-10-0-26-111.us-east-2.compute.internal default-scheduler logger.go:42: 07:42:06 | examples-with-cassandra | 2025-02-03 07:41:30 +0000 UTC Normal Job.batch with-cassandra-cassandra-schema-job SuccessfulCreate Created pod: with-cassandra-cassandra-schema-job-brlvr job-controller logger.go:42: 07:42:06 | examples-with-cassandra | 2025-02-03 07:41:31 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-brlvr AddedInterface Add eth0 [10.128.2.55/23] from ovn-kubernetes multus logger.go:42: 07:42:06 | examples-with-cassandra | 2025-02-03 07:41:31 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-brlvr.spec.containers{with-cassandra-cassandra-schema-job} Pulling Pulling image "jaegertracing/jaeger-cassandra-schema:1.62.0" kubelet logger.go:42: 07:42:06 | examples-with-cassandra | 2025-02-03 07:41:36 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-brlvr.spec.containers{with-cassandra-cassandra-schema-job} Pulled Successfully pulled image "jaegertracing/jaeger-cassandra-schema:1.62.0" in 5.158s (5.158s including waiting). Image size: 405965365 bytes. kubelet logger.go:42: 07:42:06 | examples-with-cassandra | 2025-02-03 07:41:36 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-brlvr.spec.containers{with-cassandra-cassandra-schema-job} Created Created container with-cassandra-cassandra-schema-job kubelet logger.go:42: 07:42:06 | examples-with-cassandra | 2025-02-03 07:41:36 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-brlvr.spec.containers{with-cassandra-cassandra-schema-job} Started Started container with-cassandra-cassandra-schema-job kubelet logger.go:42: 07:42:06 | examples-with-cassandra | 2025-02-03 07:41:42 +0000 UTC Normal Job.batch with-cassandra-cassandra-schema-job Completed Job completed job-controller logger.go:42: 07:42:06 | examples-with-cassandra | 2025-02-03 07:41:42 +0000 UTC Normal Deployment.apps with-cassandra ScalingReplicaSet Scaled up replica set with-cassandra-6dfbfb9f86 to 1 deployment-controller logger.go:42: 07:42:06 | examples-with-cassandra | 2025-02-03 07:41:43 +0000 UTC Normal Pod with-cassandra-6dfbfb9f86-5lwb6 Binding Scheduled Successfully assigned kuttl-test-pure-caribou/with-cassandra-6dfbfb9f86-5lwb6 to ip-10-0-66-98.us-east-2.compute.internal default-scheduler logger.go:42: 07:42:06 | examples-with-cassandra | 2025-02-03 07:41:43 +0000 UTC Normal Pod with-cassandra-6dfbfb9f86-5lwb6 AddedInterface Add eth0 [10.129.2.43/23] from ovn-kubernetes multus logger.go:42: 07:42:06 | examples-with-cassandra | 2025-02-03 07:41:43 +0000 UTC Normal Pod with-cassandra-6dfbfb9f86-5lwb6.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:fa9aaf9ae026228265ce7f10fb451d5c52a4e0269ce1fa9024cbefddf765ae8d" already present on machine kubelet logger.go:42: 07:42:06 | examples-with-cassandra | 2025-02-03 07:41:43 +0000 UTC Normal Pod with-cassandra-6dfbfb9f86-5lwb6.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:42:06 | examples-with-cassandra | 2025-02-03 07:41:43 +0000 UTC Normal Pod with-cassandra-6dfbfb9f86-5lwb6.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:42:06 | examples-with-cassandra | 2025-02-03 07:41:43 +0000 UTC Normal Pod with-cassandra-6dfbfb9f86-5lwb6.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" already present on machine kubelet logger.go:42: 07:42:06 | examples-with-cassandra | 2025-02-03 07:41:43 +0000 UTC Normal ReplicaSet.apps with-cassandra-6dfbfb9f86 SuccessfulCreate Created pod: with-cassandra-6dfbfb9f86-5lwb6 replicaset-controller logger.go:42: 07:42:06 | examples-with-cassandra | 2025-02-03 07:41:44 +0000 UTC Normal Pod with-cassandra-6dfbfb9f86-5lwb6.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:42:06 | examples-with-cassandra | 2025-02-03 07:41:44 +0000 UTC Normal Pod with-cassandra-6dfbfb9f86-5lwb6.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:42:06 | examples-with-cassandra | 2025-02-03 07:41:47 +0000 UTC Normal Pod with-cassandra-6dfbfb9f86-5lwb6.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:42:06 | examples-with-cassandra | 2025-02-03 07:41:47 +0000 UTC Normal Pod with-cassandra-6dfbfb9f86-5lwb6.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:42:06 | examples-with-cassandra | 2025-02-03 07:41:47 +0000 UTC Normal ReplicaSet.apps with-cassandra-6dfbfb9f86 SuccessfulDelete Deleted pod: with-cassandra-6dfbfb9f86-5lwb6 replicaset-controller logger.go:42: 07:42:06 | examples-with-cassandra | 2025-02-03 07:41:47 +0000 UTC Normal Deployment.apps with-cassandra ScalingReplicaSet Scaled down replica set with-cassandra-6dfbfb9f86 to 0 from 1 deployment-controller logger.go:42: 07:42:06 | examples-with-cassandra | 2025-02-03 07:41:48 +0000 UTC Normal Pod with-cassandra-596869ff8c-pkgmj Binding Scheduled Successfully assigned kuttl-test-pure-caribou/with-cassandra-596869ff8c-pkgmj to ip-10-0-66-98.us-east-2.compute.internal default-scheduler logger.go:42: 07:42:06 | examples-with-cassandra | 2025-02-03 07:41:48 +0000 UTC Normal ReplicaSet.apps with-cassandra-596869ff8c SuccessfulCreate Created pod: with-cassandra-596869ff8c-pkgmj replicaset-controller logger.go:42: 07:42:06 | examples-with-cassandra | 2025-02-03 07:41:48 +0000 UTC Normal Deployment.apps with-cassandra ScalingReplicaSet Scaled up replica set with-cassandra-596869ff8c to 1 deployment-controller logger.go:42: 07:42:06 | examples-with-cassandra | 2025-02-03 07:41:49 +0000 UTC Normal Pod with-cassandra-596869ff8c-pkgmj AddedInterface Add eth0 [10.129.2.44/23] from ovn-kubernetes multus logger.go:42: 07:42:06 | examples-with-cassandra | 2025-02-03 07:41:49 +0000 UTC Normal Pod with-cassandra-596869ff8c-pkgmj.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:fa9aaf9ae026228265ce7f10fb451d5c52a4e0269ce1fa9024cbefddf765ae8d" already present on machine kubelet logger.go:42: 07:42:06 | examples-with-cassandra | 2025-02-03 07:41:49 +0000 UTC Normal Pod with-cassandra-596869ff8c-pkgmj.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:42:06 | examples-with-cassandra | 2025-02-03 07:41:49 +0000 UTC Normal Pod with-cassandra-596869ff8c-pkgmj.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:42:06 | examples-with-cassandra | 2025-02-03 07:41:49 +0000 UTC Normal Pod with-cassandra-596869ff8c-pkgmj.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" already present on machine kubelet logger.go:42: 07:42:06 | examples-with-cassandra | 2025-02-03 07:41:49 +0000 UTC Normal Pod with-cassandra-596869ff8c-pkgmj.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:42:06 | examples-with-cassandra | 2025-02-03 07:41:49 +0000 UTC Normal Pod with-cassandra-596869ff8c-pkgmj.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:42:06 | examples-with-cassandra | 2025-02-03 07:41:53 +0000 UTC Normal Pod check-span-zw7bz Binding Scheduled Successfully assigned kuttl-test-pure-caribou/check-span-zw7bz to ip-10-0-6-172.us-east-2.compute.internal default-scheduler logger.go:42: 07:42:06 | examples-with-cassandra | 2025-02-03 07:41:53 +0000 UTC Normal Pod check-span-zw7bz AddedInterface Add eth0 [10.131.0.35/23] from ovn-kubernetes multus logger.go:42: 07:42:06 | examples-with-cassandra | 2025-02-03 07:41:53 +0000 UTC Normal Pod check-span-zw7bz.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:42:06 | examples-with-cassandra | 2025-02-03 07:41:53 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-zw7bz job-controller logger.go:42: 07:42:06 | examples-with-cassandra | 2025-02-03 07:41:53 +0000 UTC Normal Pod report-span-2lpfn Binding Scheduled Successfully assigned kuttl-test-pure-caribou/report-span-2lpfn to ip-10-0-26-111.us-east-2.compute.internal default-scheduler logger.go:42: 07:42:06 | examples-with-cassandra | 2025-02-03 07:41:53 +0000 UTC Normal Pod report-span-2lpfn AddedInterface Add eth0 [10.128.2.56/23] from ovn-kubernetes multus logger.go:42: 07:42:06 | examples-with-cassandra | 2025-02-03 07:41:53 +0000 UTC Normal Pod report-span-2lpfn.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:42:06 | examples-with-cassandra | 2025-02-03 07:41:53 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-2lpfn job-controller logger.go:42: 07:42:06 | examples-with-cassandra | 2025-02-03 07:41:54 +0000 UTC Normal Pod check-span-zw7bz.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 624ms (624ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:42:06 | examples-with-cassandra | 2025-02-03 07:41:54 +0000 UTC Normal Pod check-span-zw7bz.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:42:06 | examples-with-cassandra | 2025-02-03 07:41:54 +0000 UTC Normal Pod check-span-zw7bz.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:42:06 | examples-with-cassandra | 2025-02-03 07:41:54 +0000 UTC Normal Pod report-span-2lpfn.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 1.179s (1.179s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:42:06 | examples-with-cassandra | 2025-02-03 07:41:55 +0000 UTC Normal Pod report-span-2lpfn.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:42:06 | examples-with-cassandra | 2025-02-03 07:41:55 +0000 UTC Normal Pod report-span-2lpfn.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:42:06 | examples-with-cassandra | 2025-02-03 07:42:05 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:42:06 | examples-with-cassandra | Deleting namespace: kuttl-test-pure-caribou === CONT kuttl/harness/examples-with-badger logger.go:42: 07:42:19 | examples-with-badger | Creating namespace: kuttl-test-stable-lemming logger.go:42: 07:42:19 | examples-with-badger/0-install | starting test step 0-install logger.go:42: 07:42:19 | examples-with-badger/0-install | Jaeger:kuttl-test-stable-lemming/with-badger created logger.go:42: 07:42:27 | examples-with-badger/0-install | test step completed 0-install logger.go:42: 07:42:27 | examples-with-badger/1-smoke-test | starting test step 1-smoke-test logger.go:42: 07:42:27 | examples-with-badger/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE with-badger /dev/null] logger.go:42: 07:42:28 | examples-with-badger/1-smoke-test | Warning: resource jaegers/with-badger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:42:35 | examples-with-badger/1-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://with-badger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://with-badger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:42:36 | examples-with-badger/1-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:42:36 | examples-with-badger/1-smoke-test | job.batch/report-span created logger.go:42: 07:42:36 | examples-with-badger/1-smoke-test | job.batch/check-span created logger.go:42: 07:42:49 | examples-with-badger/1-smoke-test | test step completed 1-smoke-test logger.go:42: 07:42:49 | examples-with-badger | examples-with-badger events from ns kuttl-test-stable-lemming: logger.go:42: 07:42:49 | examples-with-badger | 2025-02-03 07:42:23 +0000 UTC Normal Pod with-badger-6c755fdbd8-j28ws Binding Scheduled Successfully assigned kuttl-test-stable-lemming/with-badger-6c755fdbd8-j28ws to ip-10-0-66-98.us-east-2.compute.internal default-scheduler logger.go:42: 07:42:49 | examples-with-badger | 2025-02-03 07:42:23 +0000 UTC Warning Pod with-badger-6c755fdbd8-j28ws FailedMount MountVolume.SetUp failed for volume "with-badger-ui-oauth-proxy-tls" : secret "with-badger-ui-oauth-proxy-tls" not found kubelet logger.go:42: 07:42:49 | examples-with-badger | 2025-02-03 07:42:23 +0000 UTC Normal ReplicaSet.apps with-badger-6c755fdbd8 SuccessfulCreate Created pod: with-badger-6c755fdbd8-j28ws replicaset-controller logger.go:42: 07:42:49 | examples-with-badger | 2025-02-03 07:42:23 +0000 UTC Normal Deployment.apps with-badger ScalingReplicaSet Scaled up replica set with-badger-6c755fdbd8 to 1 deployment-controller logger.go:42: 07:42:49 | examples-with-badger | 2025-02-03 07:42:24 +0000 UTC Normal Pod with-badger-6c755fdbd8-j28ws AddedInterface Add eth0 [10.129.2.45/23] from ovn-kubernetes multus logger.go:42: 07:42:49 | examples-with-badger | 2025-02-03 07:42:24 +0000 UTC Normal Pod with-badger-6c755fdbd8-j28ws.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:fa9aaf9ae026228265ce7f10fb451d5c52a4e0269ce1fa9024cbefddf765ae8d" already present on machine kubelet logger.go:42: 07:42:49 | examples-with-badger | 2025-02-03 07:42:25 +0000 UTC Normal Pod with-badger-6c755fdbd8-j28ws.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:42:49 | examples-with-badger | 2025-02-03 07:42:25 +0000 UTC Normal Pod with-badger-6c755fdbd8-j28ws.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:42:49 | examples-with-badger | 2025-02-03 07:42:25 +0000 UTC Normal Pod with-badger-6c755fdbd8-j28ws.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" already present on machine kubelet logger.go:42: 07:42:49 | examples-with-badger | 2025-02-03 07:42:25 +0000 UTC Normal Pod with-badger-6c755fdbd8-j28ws.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:42:49 | examples-with-badger | 2025-02-03 07:42:25 +0000 UTC Normal Pod with-badger-6c755fdbd8-j28ws.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:42:49 | examples-with-badger | 2025-02-03 07:42:31 +0000 UTC Normal Pod with-badger-6c755fdbd8-j28ws.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:42:49 | examples-with-badger | 2025-02-03 07:42:31 +0000 UTC Normal Pod with-badger-6c755fdbd8-j28ws.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:42:49 | examples-with-badger | 2025-02-03 07:42:31 +0000 UTC Normal ReplicaSet.apps with-badger-6c755fdbd8 SuccessfulDelete Deleted pod: with-badger-6c755fdbd8-j28ws replicaset-controller logger.go:42: 07:42:49 | examples-with-badger | 2025-02-03 07:42:31 +0000 UTC Normal Deployment.apps with-badger ScalingReplicaSet Scaled down replica set with-badger-6c755fdbd8 to 0 from 1 deployment-controller logger.go:42: 07:42:49 | examples-with-badger | 2025-02-03 07:42:32 +0000 UTC Normal Pod with-badger-8648c7565d-kfsv4 Binding Scheduled Successfully assigned kuttl-test-stable-lemming/with-badger-8648c7565d-kfsv4 to ip-10-0-66-98.us-east-2.compute.internal default-scheduler logger.go:42: 07:42:49 | examples-with-badger | 2025-02-03 07:42:32 +0000 UTC Normal ReplicaSet.apps with-badger-8648c7565d SuccessfulCreate Created pod: with-badger-8648c7565d-kfsv4 replicaset-controller logger.go:42: 07:42:49 | examples-with-badger | 2025-02-03 07:42:32 +0000 UTC Normal Deployment.apps with-badger ScalingReplicaSet Scaled up replica set with-badger-8648c7565d to 1 deployment-controller logger.go:42: 07:42:49 | examples-with-badger | 2025-02-03 07:42:33 +0000 UTC Normal Pod with-badger-8648c7565d-kfsv4 AddedInterface Add eth0 [10.129.2.46/23] from ovn-kubernetes multus logger.go:42: 07:42:49 | examples-with-badger | 2025-02-03 07:42:33 +0000 UTC Normal Pod with-badger-8648c7565d-kfsv4.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:fa9aaf9ae026228265ce7f10fb451d5c52a4e0269ce1fa9024cbefddf765ae8d" already present on machine kubelet logger.go:42: 07:42:49 | examples-with-badger | 2025-02-03 07:42:33 +0000 UTC Normal Pod with-badger-8648c7565d-kfsv4.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:42:49 | examples-with-badger | 2025-02-03 07:42:33 +0000 UTC Normal Pod with-badger-8648c7565d-kfsv4.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:42:49 | examples-with-badger | 2025-02-03 07:42:33 +0000 UTC Normal Pod with-badger-8648c7565d-kfsv4.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" already present on machine kubelet logger.go:42: 07:42:49 | examples-with-badger | 2025-02-03 07:42:33 +0000 UTC Normal Pod with-badger-8648c7565d-kfsv4.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:42:49 | examples-with-badger | 2025-02-03 07:42:33 +0000 UTC Normal Pod with-badger-8648c7565d-kfsv4.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:42:49 | examples-with-badger | 2025-02-03 07:42:36 +0000 UTC Normal Pod check-span-68r2g Binding Scheduled Successfully assigned kuttl-test-stable-lemming/check-span-68r2g to ip-10-0-26-111.us-east-2.compute.internal default-scheduler logger.go:42: 07:42:49 | examples-with-badger | 2025-02-03 07:42:36 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-68r2g job-controller logger.go:42: 07:42:49 | examples-with-badger | 2025-02-03 07:42:36 +0000 UTC Normal Pod report-span-8kvgq Binding Scheduled Successfully assigned kuttl-test-stable-lemming/report-span-8kvgq to ip-10-0-6-172.us-east-2.compute.internal default-scheduler logger.go:42: 07:42:49 | examples-with-badger | 2025-02-03 07:42:36 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-8kvgq job-controller logger.go:42: 07:42:49 | examples-with-badger | 2025-02-03 07:42:37 +0000 UTC Normal Pod check-span-68r2g AddedInterface Add eth0 [10.128.2.57/23] from ovn-kubernetes multus logger.go:42: 07:42:49 | examples-with-badger | 2025-02-03 07:42:37 +0000 UTC Normal Pod check-span-68r2g.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:42:49 | examples-with-badger | 2025-02-03 07:42:37 +0000 UTC Normal Pod report-span-8kvgq AddedInterface Add eth0 [10.131.0.36/23] from ovn-kubernetes multus logger.go:42: 07:42:49 | examples-with-badger | 2025-02-03 07:42:37 +0000 UTC Normal Pod report-span-8kvgq.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:42:49 | examples-with-badger | 2025-02-03 07:42:37 +0000 UTC Normal Pod report-span-8kvgq.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 573ms (573ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:42:49 | examples-with-badger | 2025-02-03 07:42:37 +0000 UTC Normal Pod report-span-8kvgq.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:42:49 | examples-with-badger | 2025-02-03 07:42:38 +0000 UTC Normal Pod check-span-68r2g.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 1.123s (1.123s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:42:49 | examples-with-badger | 2025-02-03 07:42:38 +0000 UTC Normal Pod check-span-68r2g.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:42:49 | examples-with-badger | 2025-02-03 07:42:38 +0000 UTC Normal Pod check-span-68r2g.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:42:49 | examples-with-badger | 2025-02-03 07:42:38 +0000 UTC Normal Pod report-span-8kvgq.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:42:49 | examples-with-badger | 2025-02-03 07:42:48 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:42:49 | examples-with-badger | Deleting namespace: kuttl-test-stable-lemming === CONT kuttl/harness/examples-simplest logger.go:42: 07:43:02 | examples-simplest | Creating namespace: kuttl-test-communal-kit logger.go:42: 07:43:02 | examples-simplest/0-install | starting test step 0-install logger.go:42: 07:43:02 | examples-simplest/0-install | Jaeger:kuttl-test-communal-kit/simplest created logger.go:42: 07:43:09 | examples-simplest/0-install | test step completed 0-install logger.go:42: 07:43:09 | examples-simplest/1-smoke-test | starting test step 1-smoke-test logger.go:42: 07:43:09 | examples-simplest/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE simplest /dev/null] logger.go:42: 07:43:10 | examples-simplest/1-smoke-test | Warning: resource jaegers/simplest is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:43:16 | examples-simplest/1-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://simplest-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:43:17 | examples-simplest/1-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:43:17 | examples-simplest/1-smoke-test | job.batch/report-span created logger.go:42: 07:43:17 | examples-simplest/1-smoke-test | job.batch/check-span created logger.go:42: 07:43:29 | examples-simplest/1-smoke-test | test step completed 1-smoke-test logger.go:42: 07:43:29 | examples-simplest | examples-simplest events from ns kuttl-test-communal-kit: logger.go:42: 07:43:29 | examples-simplest | 2025-02-03 07:43:06 +0000 UTC Normal Pod simplest-db8557cb7-gcxkd Binding Scheduled Successfully assigned kuttl-test-communal-kit/simplest-db8557cb7-gcxkd to ip-10-0-66-98.us-east-2.compute.internal default-scheduler logger.go:42: 07:43:29 | examples-simplest | 2025-02-03 07:43:06 +0000 UTC Normal Pod simplest-db8557cb7-gcxkd AddedInterface Add eth0 [10.129.2.47/23] from ovn-kubernetes multus logger.go:42: 07:43:29 | examples-simplest | 2025-02-03 07:43:06 +0000 UTC Normal Pod simplest-db8557cb7-gcxkd.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:fa9aaf9ae026228265ce7f10fb451d5c52a4e0269ce1fa9024cbefddf765ae8d" already present on machine kubelet logger.go:42: 07:43:29 | examples-simplest | 2025-02-03 07:43:06 +0000 UTC Normal ReplicaSet.apps simplest-db8557cb7 SuccessfulCreate Created pod: simplest-db8557cb7-gcxkd replicaset-controller logger.go:42: 07:43:29 | examples-simplest | 2025-02-03 07:43:06 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled up replica set simplest-db8557cb7 to 1 deployment-controller logger.go:42: 07:43:29 | examples-simplest | 2025-02-03 07:43:07 +0000 UTC Normal Pod simplest-db8557cb7-gcxkd.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:43:29 | examples-simplest | 2025-02-03 07:43:07 +0000 UTC Normal Pod simplest-db8557cb7-gcxkd.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:43:29 | examples-simplest | 2025-02-03 07:43:07 +0000 UTC Normal Pod simplest-db8557cb7-gcxkd.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" already present on machine kubelet logger.go:42: 07:43:29 | examples-simplest | 2025-02-03 07:43:07 +0000 UTC Normal Pod simplest-db8557cb7-gcxkd.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:43:29 | examples-simplest | 2025-02-03 07:43:07 +0000 UTC Normal Pod simplest-db8557cb7-gcxkd.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:43:29 | examples-simplest | 2025-02-03 07:43:12 +0000 UTC Normal Pod simplest-db8557cb7-gcxkd.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:43:29 | examples-simplest | 2025-02-03 07:43:12 +0000 UTC Normal Pod simplest-db8557cb7-gcxkd.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:43:29 | examples-simplest | 2025-02-03 07:43:12 +0000 UTC Normal ReplicaSet.apps simplest-db8557cb7 SuccessfulDelete Deleted pod: simplest-db8557cb7-gcxkd replicaset-controller logger.go:42: 07:43:29 | examples-simplest | 2025-02-03 07:43:12 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled down replica set simplest-db8557cb7 to 0 from 1 deployment-controller logger.go:42: 07:43:29 | examples-simplest | 2025-02-03 07:43:13 +0000 UTC Normal Pod simplest-c958bdc69-5jnd2 Binding Scheduled Successfully assigned kuttl-test-communal-kit/simplest-c958bdc69-5jnd2 to ip-10-0-66-98.us-east-2.compute.internal default-scheduler logger.go:42: 07:43:29 | examples-simplest | 2025-02-03 07:43:13 +0000 UTC Normal Pod simplest-c958bdc69-5jnd2 AddedInterface Add eth0 [10.129.2.48/23] from ovn-kubernetes multus logger.go:42: 07:43:29 | examples-simplest | 2025-02-03 07:43:13 +0000 UTC Normal Pod simplest-c958bdc69-5jnd2.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:fa9aaf9ae026228265ce7f10fb451d5c52a4e0269ce1fa9024cbefddf765ae8d" already present on machine kubelet logger.go:42: 07:43:29 | examples-simplest | 2025-02-03 07:43:13 +0000 UTC Normal Pod simplest-c958bdc69-5jnd2.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:43:29 | examples-simplest | 2025-02-03 07:43:13 +0000 UTC Normal Pod simplest-c958bdc69-5jnd2.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:43:29 | examples-simplest | 2025-02-03 07:43:13 +0000 UTC Normal Pod simplest-c958bdc69-5jnd2.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" already present on machine kubelet logger.go:42: 07:43:29 | examples-simplest | 2025-02-03 07:43:13 +0000 UTC Normal ReplicaSet.apps simplest-c958bdc69 SuccessfulCreate Created pod: simplest-c958bdc69-5jnd2 replicaset-controller logger.go:42: 07:43:29 | examples-simplest | 2025-02-03 07:43:13 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled up replica set simplest-c958bdc69 to 1 deployment-controller logger.go:42: 07:43:29 | examples-simplest | 2025-02-03 07:43:14 +0000 UTC Normal Pod simplest-c958bdc69-5jnd2.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:43:29 | examples-simplest | 2025-02-03 07:43:14 +0000 UTC Normal Pod simplest-c958bdc69-5jnd2.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:43:29 | examples-simplest | 2025-02-03 07:43:17 +0000 UTC Normal Pod check-span-8xdpk Binding Scheduled Successfully assigned kuttl-test-communal-kit/check-span-8xdpk to ip-10-0-26-111.us-east-2.compute.internal default-scheduler logger.go:42: 07:43:29 | examples-simplest | 2025-02-03 07:43:17 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-8xdpk job-controller logger.go:42: 07:43:29 | examples-simplest | 2025-02-03 07:43:17 +0000 UTC Normal Pod report-span-rvmn7 Binding Scheduled Successfully assigned kuttl-test-communal-kit/report-span-rvmn7 to ip-10-0-6-172.us-east-2.compute.internal default-scheduler logger.go:42: 07:43:29 | examples-simplest | 2025-02-03 07:43:17 +0000 UTC Normal Pod report-span-rvmn7 AddedInterface Add eth0 [10.131.0.37/23] from ovn-kubernetes multus logger.go:42: 07:43:29 | examples-simplest | 2025-02-03 07:43:17 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-rvmn7 job-controller logger.go:42: 07:43:29 | examples-simplest | 2025-02-03 07:43:18 +0000 UTC Normal Pod check-span-8xdpk AddedInterface Add eth0 [10.128.2.58/23] from ovn-kubernetes multus logger.go:42: 07:43:29 | examples-simplest | 2025-02-03 07:43:18 +0000 UTC Normal Pod check-span-8xdpk.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:43:29 | examples-simplest | 2025-02-03 07:43:18 +0000 UTC Normal Pod check-span-8xdpk.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 765ms (765ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:43:29 | examples-simplest | 2025-02-03 07:43:18 +0000 UTC Normal Pod check-span-8xdpk.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:43:29 | examples-simplest | 2025-02-03 07:43:18 +0000 UTC Normal Pod check-span-8xdpk.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:43:29 | examples-simplest | 2025-02-03 07:43:18 +0000 UTC Normal Pod report-span-rvmn7.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:43:29 | examples-simplest | 2025-02-03 07:43:18 +0000 UTC Normal Pod report-span-rvmn7.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 545ms (545ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:43:29 | examples-simplest | 2025-02-03 07:43:18 +0000 UTC Normal Pod report-span-rvmn7.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:43:29 | examples-simplest | 2025-02-03 07:43:18 +0000 UTC Normal Pod report-span-rvmn7.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:43:29 | examples-simplest | 2025-02-03 07:43:29 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:43:29 | examples-simplest | Deleting namespace: kuttl-test-communal-kit === CONT kuttl/harness/examples-simple-prod-with-volumes logger.go:42: 07:43:37 | examples-simple-prod-with-volumes | Ignoring 03-check-volume.yaml.template as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:43:37 | examples-simple-prod-with-volumes | Creating namespace: kuttl-test-picked-stingray logger.go:42: 07:43:37 | examples-simple-prod-with-volumes/1-install | starting test step 1-install logger.go:42: 07:43:37 | examples-simple-prod-with-volumes/1-install | Jaeger:kuttl-test-picked-stingray/simple-prod created logger.go:42: 07:44:38 | examples-simple-prod-with-volumes/1-install | test step completed 1-install logger.go:42: 07:44:38 | examples-simple-prod-with-volumes/2-smoke-test | starting test step 2-smoke-test logger.go:42: 07:44:38 | examples-simple-prod-with-volumes/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE simple-prod /dev/null] logger.go:42: 07:44:40 | examples-simple-prod-with-volumes/2-smoke-test | Warning: resource jaegers/simple-prod is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:44:46 | examples-simple-prod-with-volumes/2-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:44:46 | examples-simple-prod-with-volumes/2-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:44:46 | examples-simple-prod-with-volumes/2-smoke-test | job.batch/report-span created logger.go:42: 07:44:46 | examples-simple-prod-with-volumes/2-smoke-test | job.batch/check-span created logger.go:42: 07:44:59 | examples-simple-prod-with-volumes/2-smoke-test | test step completed 2-smoke-test logger.go:42: 07:44:59 | examples-simple-prod-with-volumes/3-check-volume | starting test step 3-check-volume logger.go:42: 07:44:59 | examples-simple-prod-with-volumes/3-check-volume | running command: [sh -c kubectl exec $(kubectl get pods -n $NAMESPACE -l app=jaeger -l app.kubernetes.io/component=collector -o yaml | /tmp/jaeger-tests/bin/yq e '.items[0].metadata.name') -n $NAMESPACE -- ls /usr/share/elasticsearch/data] logger.go:42: 07:45:00 | examples-simple-prod-with-volumes/3-check-volume | test step completed 3-check-volume logger.go:42: 07:45:00 | examples-simple-prod-with-volumes | examples-simple-prod-with-volumes events from ns kuttl-test-picked-stingray: logger.go:42: 07:45:00 | examples-simple-prod-with-volumes | 2025-02-03 07:44:07 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestpickedstingraysimpleprod-1-7dbc8fcd87 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestpickedstingraysimpleprod-1-7dbcrmvpr replicaset-controller logger.go:42: 07:45:00 | examples-simple-prod-with-volumes | 2025-02-03 07:44:07 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpickedstingraysimpleprod-1-7dbcrmvpr Binding Scheduled Successfully assigned kuttl-test-picked-stingray/elasticsearch-cdm-kuttltestpickedstingraysimpleprod-1-7dbcrmvpr to ip-10-0-66-98.us-east-2.compute.internal default-scheduler logger.go:42: 07:45:00 | examples-simple-prod-with-volumes | 2025-02-03 07:44:07 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestpickedstingraysimpleprod-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestpickedstingraysimpleprod-1-7dbc8fcd87 to 1 deployment-controller logger.go:42: 07:45:00 | examples-simple-prod-with-volumes | 2025-02-03 07:44:08 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpickedstingraysimpleprod-1-7dbcrmvpr AddedInterface Add eth0 [10.129.2.49/23] from ovn-kubernetes multus logger.go:42: 07:45:00 | examples-simple-prod-with-volumes | 2025-02-03 07:44:08 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpickedstingraysimpleprod-1-7dbcrmvpr.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:410674bf2f77ff1d431f494cadfab3e81249b35695eb65aedfdb8498079b4d35" already present on machine kubelet logger.go:42: 07:45:00 | examples-simple-prod-with-volumes | 2025-02-03 07:44:08 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpickedstingraysimpleprod-1-7dbcrmvpr.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:45:00 | examples-simple-prod-with-volumes | 2025-02-03 07:44:08 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpickedstingraysimpleprod-1-7dbcrmvpr.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:45:00 | examples-simple-prod-with-volumes | 2025-02-03 07:44:08 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpickedstingraysimpleprod-1-7dbcrmvpr.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:fe2439e5f9e14d828caace425214edfa3edc179d72d225c43f69e9e4c9a5837b" already present on machine kubelet logger.go:42: 07:45:00 | examples-simple-prod-with-volumes | 2025-02-03 07:44:08 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpickedstingraysimpleprod-1-7dbcrmvpr.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:45:00 | examples-simple-prod-with-volumes | 2025-02-03 07:44:08 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpickedstingraysimpleprod-1-7dbcrmvpr.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:45:00 | examples-simple-prod-with-volumes | 2025-02-03 07:44:23 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestpickedstingraysimpleprod-1-7dbcrmvpr.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:45:00 | examples-simple-prod-with-volumes | 2025-02-03 07:44:34 +0000 UTC Normal Pod simple-prod-collector-f86b556df-n7897 Binding Scheduled Successfully assigned kuttl-test-picked-stingray/simple-prod-collector-f86b556df-n7897 to ip-10-0-6-172.us-east-2.compute.internal default-scheduler logger.go:42: 07:45:00 | examples-simple-prod-with-volumes | 2025-02-03 07:44:34 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-f86b556df SuccessfulCreate Created pod: simple-prod-collector-f86b556df-n7897 replicaset-controller logger.go:42: 07:45:00 | examples-simple-prod-with-volumes | 2025-02-03 07:44:34 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-f86b556df to 1 deployment-controller logger.go:42: 07:45:00 | examples-simple-prod-with-volumes | 2025-02-03 07:44:34 +0000 UTC Normal Pod simple-prod-query-6cfc58ffcd-6f795 Binding Scheduled Successfully assigned kuttl-test-picked-stingray/simple-prod-query-6cfc58ffcd-6f795 to ip-10-0-26-111.us-east-2.compute.internal default-scheduler logger.go:42: 07:45:00 | examples-simple-prod-with-volumes | 2025-02-03 07:44:34 +0000 UTC Normal ReplicaSet.apps simple-prod-query-6cfc58ffcd SuccessfulCreate Created pod: simple-prod-query-6cfc58ffcd-6f795 replicaset-controller logger.go:42: 07:45:00 | examples-simple-prod-with-volumes | 2025-02-03 07:44:34 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-6cfc58ffcd to 1 deployment-controller logger.go:42: 07:45:00 | examples-simple-prod-with-volumes | 2025-02-03 07:44:35 +0000 UTC Normal Pod simple-prod-collector-f86b556df-n7897 AddedInterface Add eth0 [10.131.0.38/23] from ovn-kubernetes multus logger.go:42: 07:45:00 | examples-simple-prod-with-volumes | 2025-02-03 07:44:35 +0000 UTC Normal Pod simple-prod-collector-f86b556df-n7897.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:40a1c7fa23aa5ffa64e8e22aa38022f5d4d7ff644c46a3da7169b713d486c3c1" already present on machine kubelet logger.go:42: 07:45:00 | examples-simple-prod-with-volumes | 2025-02-03 07:44:35 +0000 UTC Normal Pod simple-prod-collector-f86b556df-n7897.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:45:00 | examples-simple-prod-with-volumes | 2025-02-03 07:44:35 +0000 UTC Normal Pod simple-prod-collector-f86b556df-n7897.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:45:00 | examples-simple-prod-with-volumes | 2025-02-03 07:44:35 +0000 UTC Normal Pod simple-prod-query-6cfc58ffcd-6f795 AddedInterface Add eth0 [10.128.2.59/23] from ovn-kubernetes multus logger.go:42: 07:45:00 | examples-simple-prod-with-volumes | 2025-02-03 07:44:35 +0000 UTC Normal Pod simple-prod-query-6cfc58ffcd-6f795.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:84020ce29bcb5a4bd018e6596188ae919c5cd600e08f78a546c0e76ea477685e" already present on machine kubelet logger.go:42: 07:45:00 | examples-simple-prod-with-volumes | 2025-02-03 07:44:36 +0000 UTC Normal Pod simple-prod-query-6cfc58ffcd-6f795.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:45:00 | examples-simple-prod-with-volumes | 2025-02-03 07:44:36 +0000 UTC Normal Pod simple-prod-query-6cfc58ffcd-6f795.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:45:00 | examples-simple-prod-with-volumes | 2025-02-03 07:44:36 +0000 UTC Normal Pod simple-prod-query-6cfc58ffcd-6f795.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" already present on machine kubelet logger.go:42: 07:45:00 | examples-simple-prod-with-volumes | 2025-02-03 07:44:36 +0000 UTC Normal Pod simple-prod-query-6cfc58ffcd-6f795.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:45:00 | examples-simple-prod-with-volumes | 2025-02-03 07:44:36 +0000 UTC Normal Pod simple-prod-query-6cfc58ffcd-6f795.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:45:00 | examples-simple-prod-with-volumes | 2025-02-03 07:44:36 +0000 UTC Normal Pod simple-prod-query-6cfc58ffcd-6f795.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:65342c7d622f336741a8ae0b9dead79c7ecabc155d02cdd42b7b49ca36680e74" already present on machine kubelet logger.go:42: 07:45:00 | examples-simple-prod-with-volumes | 2025-02-03 07:44:36 +0000 UTC Normal Pod simple-prod-query-6cfc58ffcd-6f795.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:45:00 | examples-simple-prod-with-volumes | 2025-02-03 07:44:36 +0000 UTC Normal Pod simple-prod-query-6cfc58ffcd-6f795.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:45:00 | examples-simple-prod-with-volumes | 2025-02-03 07:44:41 +0000 UTC Normal Pod simple-prod-query-6cfc58ffcd-6f795.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:45:00 | examples-simple-prod-with-volumes | 2025-02-03 07:44:41 +0000 UTC Normal Pod simple-prod-query-6cfc58ffcd-6f795.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:45:00 | examples-simple-prod-with-volumes | 2025-02-03 07:44:41 +0000 UTC Normal Pod simple-prod-query-6cfc58ffcd-6f795.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:45:00 | examples-simple-prod-with-volumes | 2025-02-03 07:44:41 +0000 UTC Normal ReplicaSet.apps simple-prod-query-6cfc58ffcd SuccessfulDelete Deleted pod: simple-prod-query-6cfc58ffcd-6f795 replicaset-controller logger.go:42: 07:45:00 | examples-simple-prod-with-volumes | 2025-02-03 07:44:41 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled down replica set simple-prod-query-6cfc58ffcd to 0 from 1 deployment-controller logger.go:42: 07:45:00 | examples-simple-prod-with-volumes | 2025-02-03 07:44:43 +0000 UTC Normal Pod simple-prod-query-86dc4864cb-gg8mz Binding Scheduled Successfully assigned kuttl-test-picked-stingray/simple-prod-query-86dc4864cb-gg8mz to ip-10-0-26-111.us-east-2.compute.internal default-scheduler logger.go:42: 07:45:00 | examples-simple-prod-with-volumes | 2025-02-03 07:44:43 +0000 UTC Normal Pod simple-prod-query-86dc4864cb-gg8mz AddedInterface Add eth0 [10.128.2.60/23] from ovn-kubernetes multus logger.go:42: 07:45:00 | examples-simple-prod-with-volumes | 2025-02-03 07:44:43 +0000 UTC Normal Pod simple-prod-query-86dc4864cb-gg8mz.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:84020ce29bcb5a4bd018e6596188ae919c5cd600e08f78a546c0e76ea477685e" already present on machine kubelet logger.go:42: 07:45:00 | examples-simple-prod-with-volumes | 2025-02-03 07:44:43 +0000 UTC Normal ReplicaSet.apps simple-prod-query-86dc4864cb SuccessfulCreate Created pod: simple-prod-query-86dc4864cb-gg8mz replicaset-controller logger.go:42: 07:45:00 | examples-simple-prod-with-volumes | 2025-02-03 07:44:43 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-86dc4864cb to 1 deployment-controller logger.go:42: 07:45:00 | examples-simple-prod-with-volumes | 2025-02-03 07:44:44 +0000 UTC Normal Pod simple-prod-query-86dc4864cb-gg8mz.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:45:00 | examples-simple-prod-with-volumes | 2025-02-03 07:44:44 +0000 UTC Normal Pod simple-prod-query-86dc4864cb-gg8mz.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:45:00 | examples-simple-prod-with-volumes | 2025-02-03 07:44:44 +0000 UTC Normal Pod simple-prod-query-86dc4864cb-gg8mz.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" already present on machine kubelet logger.go:42: 07:45:00 | examples-simple-prod-with-volumes | 2025-02-03 07:44:44 +0000 UTC Normal Pod simple-prod-query-86dc4864cb-gg8mz.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:45:00 | examples-simple-prod-with-volumes | 2025-02-03 07:44:44 +0000 UTC Normal Pod simple-prod-query-86dc4864cb-gg8mz.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:45:00 | examples-simple-prod-with-volumes | 2025-02-03 07:44:44 +0000 UTC Normal Pod simple-prod-query-86dc4864cb-gg8mz.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:65342c7d622f336741a8ae0b9dead79c7ecabc155d02cdd42b7b49ca36680e74" already present on machine kubelet logger.go:42: 07:45:00 | examples-simple-prod-with-volumes | 2025-02-03 07:44:44 +0000 UTC Normal Pod simple-prod-query-86dc4864cb-gg8mz.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:45:00 | examples-simple-prod-with-volumes | 2025-02-03 07:44:44 +0000 UTC Normal Pod simple-prod-query-86dc4864cb-gg8mz.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:45:00 | examples-simple-prod-with-volumes | 2025-02-03 07:44:46 +0000 UTC Normal Pod check-span-p6mhb Binding Scheduled Successfully assigned kuttl-test-picked-stingray/check-span-p6mhb to ip-10-0-6-172.us-east-2.compute.internal default-scheduler logger.go:42: 07:45:00 | examples-simple-prod-with-volumes | 2025-02-03 07:44:46 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-p6mhb job-controller logger.go:42: 07:45:00 | examples-simple-prod-with-volumes | 2025-02-03 07:44:46 +0000 UTC Normal Pod report-span-fq6qp Binding Scheduled Successfully assigned kuttl-test-picked-stingray/report-span-fq6qp to ip-10-0-6-172.us-east-2.compute.internal default-scheduler logger.go:42: 07:45:00 | examples-simple-prod-with-volumes | 2025-02-03 07:44:46 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-fq6qp job-controller logger.go:42: 07:45:00 | examples-simple-prod-with-volumes | 2025-02-03 07:44:47 +0000 UTC Normal Pod check-span-p6mhb AddedInterface Add eth0 [10.131.0.40/23] from ovn-kubernetes multus logger.go:42: 07:45:00 | examples-simple-prod-with-volumes | 2025-02-03 07:44:47 +0000 UTC Normal Pod check-span-p6mhb.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:45:00 | examples-simple-prod-with-volumes | 2025-02-03 07:44:47 +0000 UTC Normal Pod report-span-fq6qp AddedInterface Add eth0 [10.131.0.39/23] from ovn-kubernetes multus logger.go:42: 07:45:00 | examples-simple-prod-with-volumes | 2025-02-03 07:44:47 +0000 UTC Normal Pod report-span-fq6qp.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:45:00 | examples-simple-prod-with-volumes | 2025-02-03 07:44:48 +0000 UTC Normal Pod check-span-p6mhb.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 608ms (608ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:45:00 | examples-simple-prod-with-volumes | 2025-02-03 07:44:48 +0000 UTC Normal Pod check-span-p6mhb.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:45:00 | examples-simple-prod-with-volumes | 2025-02-03 07:44:48 +0000 UTC Normal Pod check-span-p6mhb.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:45:00 | examples-simple-prod-with-volumes | 2025-02-03 07:44:48 +0000 UTC Normal Pod report-span-fq6qp.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 702ms (702ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:45:00 | examples-simple-prod-with-volumes | 2025-02-03 07:44:48 +0000 UTC Normal Pod report-span-fq6qp.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:45:00 | examples-simple-prod-with-volumes | 2025-02-03 07:44:48 +0000 UTC Normal Pod report-span-fq6qp.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:45:00 | examples-simple-prod-with-volumes | 2025-02-03 07:44:50 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:45:00 | examples-simple-prod-with-volumes | 2025-02-03 07:44:50 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedGetResourceMetric failed to get memory utilization: missing request for memory in container jaeger-collector of Pod simple-prod-collector-f86b556df-n7897 horizontal-pod-autoscaler logger.go:42: 07:45:00 | examples-simple-prod-with-volumes | 2025-02-03 07:44:50 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:45:00 | examples-simple-prod-with-volumes | 2025-02-03 07:44:58 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:45:00 | examples-simple-prod-with-volumes | Deleting namespace: kuttl-test-picked-stingray === CONT kuttl/harness/examples-business-application-injected-sidecar logger.go:42: 07:45:13 | examples-business-application-injected-sidecar | Creating namespace: kuttl-test-wealthy-tiger logger.go:42: 07:45:13 | examples-business-application-injected-sidecar/0-install | starting test step 0-install logger.go:42: 07:45:13 | examples-business-application-injected-sidecar/0-install | Deployment:kuttl-test-wealthy-tiger/myapp created logger.go:42: 07:45:13 | examples-business-application-injected-sidecar/0-install | test step completed 0-install logger.go:42: 07:45:13 | examples-business-application-injected-sidecar/1-install | starting test step 1-install logger.go:42: 07:45:13 | examples-business-application-injected-sidecar/1-install | Jaeger:kuttl-test-wealthy-tiger/simplest created logger.go:42: 07:45:25 | examples-business-application-injected-sidecar/1-install | test step completed 1-install logger.go:42: 07:45:25 | examples-business-application-injected-sidecar/2-smoke-test | starting test step 2-smoke-test logger.go:42: 07:45:25 | examples-business-application-injected-sidecar/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE simplest /dev/null] logger.go:42: 07:45:27 | examples-business-application-injected-sidecar/2-smoke-test | Warning: resource jaegers/simplest is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:45:33 | examples-business-application-injected-sidecar/2-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://simplest-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:45:33 | examples-business-application-injected-sidecar/2-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:45:33 | examples-business-application-injected-sidecar/2-smoke-test | job.batch/report-span created logger.go:42: 07:45:33 | examples-business-application-injected-sidecar/2-smoke-test | job.batch/check-span created logger.go:42: 07:45:46 | examples-business-application-injected-sidecar/2-smoke-test | test step completed 2-smoke-test logger.go:42: 07:45:46 | examples-business-application-injected-sidecar | examples-business-application-injected-sidecar events from ns kuttl-test-wealthy-tiger: logger.go:42: 07:45:46 | examples-business-application-injected-sidecar | 2025-02-03 07:45:13 +0000 UTC Normal Pod myapp-6b45cc496c-n62xr Binding Scheduled Successfully assigned kuttl-test-wealthy-tiger/myapp-6b45cc496c-n62xr to ip-10-0-6-172.us-east-2.compute.internal default-scheduler logger.go:42: 07:45:46 | examples-business-application-injected-sidecar | 2025-02-03 07:45:13 +0000 UTC Normal ReplicaSet.apps myapp-6b45cc496c SuccessfulCreate Created pod: myapp-6b45cc496c-n62xr replicaset-controller logger.go:42: 07:45:46 | examples-business-application-injected-sidecar | 2025-02-03 07:45:13 +0000 UTC Normal Pod myapp-fc54f78d8-jnwsh Binding Scheduled Successfully assigned kuttl-test-wealthy-tiger/myapp-fc54f78d8-jnwsh to ip-10-0-66-98.us-east-2.compute.internal default-scheduler logger.go:42: 07:45:46 | examples-business-application-injected-sidecar | 2025-02-03 07:45:13 +0000 UTC Normal ReplicaSet.apps myapp-fc54f78d8 SuccessfulCreate Created pod: myapp-fc54f78d8-jnwsh replicaset-controller logger.go:42: 07:45:46 | examples-business-application-injected-sidecar | 2025-02-03 07:45:13 +0000 UTC Normal Deployment.apps myapp ScalingReplicaSet Scaled up replica set myapp-fc54f78d8 to 1 deployment-controller logger.go:42: 07:45:46 | examples-business-application-injected-sidecar | 2025-02-03 07:45:13 +0000 UTC Normal Deployment.apps myapp ScalingReplicaSet Scaled up replica set myapp-6b45cc496c to 1 deployment-controller logger.go:42: 07:45:46 | examples-business-application-injected-sidecar | 2025-02-03 07:45:14 +0000 UTC Warning Pod myapp-6b45cc496c-n62xr FailedMount MountVolume.SetUp failed for volume "simplest-service-ca" : configmap "simplest-service-ca" not found kubelet logger.go:42: 07:45:46 | examples-business-application-injected-sidecar | 2025-02-03 07:45:14 +0000 UTC Warning Pod myapp-6b45cc496c-n62xr FailedMount MountVolume.SetUp failed for volume "simplest-trusted-ca" : configmap "simplest-trusted-ca" not found kubelet logger.go:42: 07:45:46 | examples-business-application-injected-sidecar | 2025-02-03 07:45:14 +0000 UTC Normal Pod myapp-fc54f78d8-jnwsh AddedInterface Add eth0 [10.129.2.50/23] from ovn-kubernetes multus logger.go:42: 07:45:46 | examples-business-application-injected-sidecar | 2025-02-03 07:45:14 +0000 UTC Normal Pod myapp-fc54f78d8-jnwsh.spec.containers{myapp} Pulling Pulling image "jaegertracing/vertx-create-span:operator-e2e-tests" kubelet logger.go:42: 07:45:46 | examples-business-application-injected-sidecar | 2025-02-03 07:45:18 +0000 UTC Normal Pod myapp-fc54f78d8-jnwsh.spec.containers{myapp} Pulled Successfully pulled image "jaegertracing/vertx-create-span:operator-e2e-tests" in 3.807s (3.807s including waiting). Image size: 282912835 bytes. kubelet logger.go:42: 07:45:46 | examples-business-application-injected-sidecar | 2025-02-03 07:45:18 +0000 UTC Normal Pod myapp-fc54f78d8-jnwsh.spec.containers{myapp} Created Created container myapp kubelet logger.go:42: 07:45:46 | examples-business-application-injected-sidecar | 2025-02-03 07:45:18 +0000 UTC Normal Pod myapp-fc54f78d8-jnwsh.spec.containers{myapp} Started Started container myapp kubelet logger.go:42: 07:45:46 | examples-business-application-injected-sidecar | 2025-02-03 07:45:23 +0000 UTC Normal Pod simplest-557787598c-cjwnf Binding Scheduled Successfully assigned kuttl-test-wealthy-tiger/simplest-557787598c-cjwnf to ip-10-0-66-98.us-east-2.compute.internal default-scheduler logger.go:42: 07:45:46 | examples-business-application-injected-sidecar | 2025-02-03 07:45:23 +0000 UTC Warning Pod simplest-557787598c-cjwnf FailedMount MountVolume.SetUp failed for volume "simplest-ui-oauth-proxy-tls" : secret "simplest-ui-oauth-proxy-tls" not found kubelet logger.go:42: 07:45:46 | examples-business-application-injected-sidecar | 2025-02-03 07:45:23 +0000 UTC Warning Pod simplest-557787598c-cjwnf FailedMount MountVolume.SetUp failed for volume "simplest-collector-tls-config-volume" : secret "simplest-collector-headless-tls" not found kubelet logger.go:42: 07:45:46 | examples-business-application-injected-sidecar | 2025-02-03 07:45:23 +0000 UTC Normal ReplicaSet.apps simplest-557787598c SuccessfulCreate Created pod: simplest-557787598c-cjwnf replicaset-controller logger.go:42: 07:45:46 | examples-business-application-injected-sidecar | 2025-02-03 07:45:23 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled up replica set simplest-557787598c to 1 deployment-controller logger.go:42: 07:45:46 | examples-business-application-injected-sidecar | 2025-02-03 07:45:24 +0000 UTC Normal Pod simplest-557787598c-cjwnf AddedInterface Add eth0 [10.129.2.51/23] from ovn-kubernetes multus logger.go:42: 07:45:46 | examples-business-application-injected-sidecar | 2025-02-03 07:45:24 +0000 UTC Normal Pod simplest-557787598c-cjwnf.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:fa9aaf9ae026228265ce7f10fb451d5c52a4e0269ce1fa9024cbefddf765ae8d" already present on machine kubelet logger.go:42: 07:45:46 | examples-business-application-injected-sidecar | 2025-02-03 07:45:24 +0000 UTC Normal Pod simplest-557787598c-cjwnf.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:45:46 | examples-business-application-injected-sidecar | 2025-02-03 07:45:24 +0000 UTC Normal Pod simplest-557787598c-cjwnf.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:45:46 | examples-business-application-injected-sidecar | 2025-02-03 07:45:24 +0000 UTC Normal Pod simplest-557787598c-cjwnf.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" already present on machine kubelet logger.go:42: 07:45:46 | examples-business-application-injected-sidecar | 2025-02-03 07:45:24 +0000 UTC Normal Pod simplest-557787598c-cjwnf.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:45:46 | examples-business-application-injected-sidecar | 2025-02-03 07:45:24 +0000 UTC Normal Pod simplest-557787598c-cjwnf.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:45:46 | examples-business-application-injected-sidecar | 2025-02-03 07:45:25 +0000 UTC Warning Pod myapp-fc54f78d8-jnwsh.spec.containers{myapp} Unhealthy Liveness probe failed: Get "http://10.129.2.50:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:45:46 | examples-business-application-injected-sidecar | 2025-02-03 07:45:28 +0000 UTC Normal Pod simplest-557787598c-cjwnf.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:45:46 | examples-business-application-injected-sidecar | 2025-02-03 07:45:28 +0000 UTC Normal Pod simplest-557787598c-cjwnf.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:45:46 | examples-business-application-injected-sidecar | 2025-02-03 07:45:28 +0000 UTC Normal ReplicaSet.apps simplest-557787598c SuccessfulDelete Deleted pod: simplest-557787598c-cjwnf replicaset-controller logger.go:42: 07:45:46 | examples-business-application-injected-sidecar | 2025-02-03 07:45:28 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled down replica set simplest-557787598c to 0 from 1 deployment-controller logger.go:42: 07:45:46 | examples-business-application-injected-sidecar | 2025-02-03 07:45:29 +0000 UTC Normal Pod simplest-956997c48-6hf46 Binding Scheduled Successfully assigned kuttl-test-wealthy-tiger/simplest-956997c48-6hf46 to ip-10-0-66-98.us-east-2.compute.internal default-scheduler logger.go:42: 07:45:46 | examples-business-application-injected-sidecar | 2025-02-03 07:45:29 +0000 UTC Normal ReplicaSet.apps simplest-956997c48 SuccessfulCreate Created pod: simplest-956997c48-6hf46 replicaset-controller logger.go:42: 07:45:46 | examples-business-application-injected-sidecar | 2025-02-03 07:45:29 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled up replica set simplest-956997c48 to 1 deployment-controller logger.go:42: 07:45:46 | examples-business-application-injected-sidecar | 2025-02-03 07:45:30 +0000 UTC Normal Pod myapp-6b45cc496c-n62xr AddedInterface Add eth0 [10.131.0.42/23] from ovn-kubernetes multus logger.go:42: 07:45:46 | examples-business-application-injected-sidecar | 2025-02-03 07:45:30 +0000 UTC Normal Pod myapp-6b45cc496c-n62xr.spec.containers{myapp} Pulling Pulling image "jaegertracing/vertx-create-span:operator-e2e-tests" kubelet logger.go:42: 07:45:46 | examples-business-application-injected-sidecar | 2025-02-03 07:45:30 +0000 UTC Normal Pod simplest-956997c48-6hf46 AddedInterface Add eth0 [10.129.2.52/23] from ovn-kubernetes multus logger.go:42: 07:45:46 | examples-business-application-injected-sidecar | 2025-02-03 07:45:30 +0000 UTC Normal Pod simplest-956997c48-6hf46.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:fa9aaf9ae026228265ce7f10fb451d5c52a4e0269ce1fa9024cbefddf765ae8d" already present on machine kubelet logger.go:42: 07:45:46 | examples-business-application-injected-sidecar | 2025-02-03 07:45:30 +0000 UTC Normal Pod simplest-956997c48-6hf46.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:45:46 | examples-business-application-injected-sidecar | 2025-02-03 07:45:30 +0000 UTC Normal Pod simplest-956997c48-6hf46.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:45:46 | examples-business-application-injected-sidecar | 2025-02-03 07:45:30 +0000 UTC Normal Pod simplest-956997c48-6hf46.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" already present on machine kubelet logger.go:42: 07:45:46 | examples-business-application-injected-sidecar | 2025-02-03 07:45:30 +0000 UTC Normal Pod simplest-956997c48-6hf46.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:45:46 | examples-business-application-injected-sidecar | 2025-02-03 07:45:30 +0000 UTC Normal Pod simplest-956997c48-6hf46.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:45:46 | examples-business-application-injected-sidecar | 2025-02-03 07:45:33 +0000 UTC Normal Pod check-span-mchc9 Binding Scheduled Successfully assigned kuttl-test-wealthy-tiger/check-span-mchc9 to ip-10-0-6-172.us-east-2.compute.internal default-scheduler logger.go:42: 07:45:46 | examples-business-application-injected-sidecar | 2025-02-03 07:45:33 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-mchc9 job-controller logger.go:42: 07:45:46 | examples-business-application-injected-sidecar | 2025-02-03 07:45:33 +0000 UTC Normal Pod report-span-gx8jl Binding Scheduled Successfully assigned kuttl-test-wealthy-tiger/report-span-gx8jl to ip-10-0-26-111.us-east-2.compute.internal default-scheduler logger.go:42: 07:45:46 | examples-business-application-injected-sidecar | 2025-02-03 07:45:33 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-gx8jl job-controller logger.go:42: 07:45:46 | examples-business-application-injected-sidecar | 2025-02-03 07:45:34 +0000 UTC Normal Pod check-span-mchc9 AddedInterface Add eth0 [10.131.0.43/23] from ovn-kubernetes multus logger.go:42: 07:45:46 | examples-business-application-injected-sidecar | 2025-02-03 07:45:34 +0000 UTC Normal Pod myapp-6b45cc496c-n62xr.spec.containers{myapp} Pulled Successfully pulled image "jaegertracing/vertx-create-span:operator-e2e-tests" in 4.384s (4.384s including waiting). Image size: 282912835 bytes. kubelet logger.go:42: 07:45:46 | examples-business-application-injected-sidecar | 2025-02-03 07:45:34 +0000 UTC Normal Pod myapp-6b45cc496c-n62xr.spec.containers{myapp} Created Created container myapp kubelet logger.go:42: 07:45:46 | examples-business-application-injected-sidecar | 2025-02-03 07:45:34 +0000 UTC Normal Pod myapp-6b45cc496c-n62xr.spec.containers{myapp} Started Started container myapp kubelet logger.go:42: 07:45:46 | examples-business-application-injected-sidecar | 2025-02-03 07:45:34 +0000 UTC Normal Pod myapp-6b45cc496c-n62xr.spec.containers{jaeger-agent} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:65342c7d622f336741a8ae0b9dead79c7ecabc155d02cdd42b7b49ca36680e74" kubelet logger.go:42: 07:45:46 | examples-business-application-injected-sidecar | 2025-02-03 07:45:34 +0000 UTC Normal Pod report-span-gx8jl AddedInterface Add eth0 [10.128.2.61/23] from ovn-kubernetes multus logger.go:42: 07:45:46 | examples-business-application-injected-sidecar | 2025-02-03 07:45:34 +0000 UTC Normal Pod report-span-gx8jl.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:45:46 | examples-business-application-injected-sidecar | 2025-02-03 07:45:35 +0000 UTC Normal Pod check-span-mchc9.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:45:46 | examples-business-application-injected-sidecar | 2025-02-03 07:45:35 +0000 UTC Normal Pod check-span-mchc9.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 556ms (556ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:45:46 | examples-business-application-injected-sidecar | 2025-02-03 07:45:35 +0000 UTC Normal Pod check-span-mchc9.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:45:46 | examples-business-application-injected-sidecar | 2025-02-03 07:45:35 +0000 UTC Normal Pod check-span-mchc9.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:45:46 | examples-business-application-injected-sidecar | 2025-02-03 07:45:35 +0000 UTC Normal Pod report-span-gx8jl.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 725ms (725ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:45:46 | examples-business-application-injected-sidecar | 2025-02-03 07:45:35 +0000 UTC Normal Pod report-span-gx8jl.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:45:46 | examples-business-application-injected-sidecar | 2025-02-03 07:45:35 +0000 UTC Normal Pod report-span-gx8jl.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:45:46 | examples-business-application-injected-sidecar | 2025-02-03 07:45:36 +0000 UTC Normal Pod myapp-6b45cc496c-n62xr.spec.containers{jaeger-agent} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:65342c7d622f336741a8ae0b9dead79c7ecabc155d02cdd42b7b49ca36680e74" in 2.083s (2.083s including waiting). Image size: 112614125 bytes. kubelet logger.go:42: 07:45:46 | examples-business-application-injected-sidecar | 2025-02-03 07:45:37 +0000 UTC Normal Pod myapp-6b45cc496c-n62xr.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:45:46 | examples-business-application-injected-sidecar | 2025-02-03 07:45:37 +0000 UTC Normal Pod myapp-6b45cc496c-n62xr.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:45:46 | examples-business-application-injected-sidecar | 2025-02-03 07:45:38 +0000 UTC Normal Pod myapp-fc54f78d8-jnwsh.spec.containers{myapp} Killing Stopping container myapp kubelet logger.go:42: 07:45:46 | examples-business-application-injected-sidecar | 2025-02-03 07:45:38 +0000 UTC Normal ReplicaSet.apps myapp-fc54f78d8 SuccessfulDelete Deleted pod: myapp-fc54f78d8-jnwsh replicaset-controller logger.go:42: 07:45:46 | examples-business-application-injected-sidecar | 2025-02-03 07:45:38 +0000 UTC Normal Deployment.apps myapp ScalingReplicaSet Scaled down replica set myapp-fc54f78d8 to 0 from 1 deployment-controller logger.go:42: 07:45:46 | examples-business-application-injected-sidecar | 2025-02-03 07:45:40 +0000 UTC Warning Pod myapp-6b45cc496c-n62xr.spec.containers{myapp} Unhealthy Liveness probe failed: Get "http://10.131.0.42:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:45:46 | examples-business-application-injected-sidecar | 2025-02-03 07:45:46 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:45:46 | examples-business-application-injected-sidecar | Deleting namespace: kuttl-test-wealthy-tiger === CONT kuttl/harness/examples-service-types logger.go:42: 07:46:00 | examples-service-types | Creating namespace: kuttl-test-holy-midge logger.go:42: 07:46:00 | examples-service-types/0-install | starting test step 0-install logger.go:42: 07:46:00 | examples-service-types/0-install | Jaeger:kuttl-test-holy-midge/service-types created logger.go:42: 07:46:07 | examples-service-types/0-install | test step completed 0-install logger.go:42: 07:46:07 | examples-service-types/1-smoke-test | starting test step 1-smoke-test logger.go:42: 07:46:07 | examples-service-types/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE service-types /dev/null] logger.go:42: 07:46:08 | examples-service-types/1-smoke-test | Warning: resource jaegers/service-types is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:46:14 | examples-service-types/1-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://service-types-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://service-types-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:46:15 | examples-service-types/1-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:46:15 | examples-service-types/1-smoke-test | job.batch/report-span created logger.go:42: 07:46:15 | examples-service-types/1-smoke-test | job.batch/check-span created logger.go:42: 07:46:28 | examples-service-types/1-smoke-test | test step completed 1-smoke-test logger.go:42: 07:46:28 | examples-service-types/2- | starting test step 2- logger.go:42: 07:46:28 | examples-service-types/2- | test step completed 2- logger.go:42: 07:46:28 | examples-service-types | examples-service-types events from ns kuttl-test-holy-midge: logger.go:42: 07:46:28 | examples-service-types | 2025-02-03 07:46:04 +0000 UTC Normal Pod service-types-665ffcc57c-q5m29 Binding Scheduled Successfully assigned kuttl-test-holy-midge/service-types-665ffcc57c-q5m29 to ip-10-0-66-98.us-east-2.compute.internal default-scheduler logger.go:42: 07:46:28 | examples-service-types | 2025-02-03 07:46:04 +0000 UTC Normal ReplicaSet.apps service-types-665ffcc57c SuccessfulCreate Created pod: service-types-665ffcc57c-q5m29 replicaset-controller logger.go:42: 07:46:28 | examples-service-types | 2025-02-03 07:46:04 +0000 UTC Normal Service service-types-collector EnsuringLoadBalancer Ensuring load balancer service-controller logger.go:42: 07:46:28 | examples-service-types | 2025-02-03 07:46:04 +0000 UTC Normal Deployment.apps service-types ScalingReplicaSet Scaled up replica set service-types-665ffcc57c to 1 deployment-controller logger.go:42: 07:46:28 | examples-service-types | 2025-02-03 07:46:05 +0000 UTC Normal Pod service-types-665ffcc57c-q5m29 AddedInterface Add eth0 [10.129.2.53/23] from ovn-kubernetes multus logger.go:42: 07:46:28 | examples-service-types | 2025-02-03 07:46:05 +0000 UTC Normal Pod service-types-665ffcc57c-q5m29.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:fa9aaf9ae026228265ce7f10fb451d5c52a4e0269ce1fa9024cbefddf765ae8d" already present on machine kubelet logger.go:42: 07:46:28 | examples-service-types | 2025-02-03 07:46:05 +0000 UTC Normal Pod service-types-665ffcc57c-q5m29.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:46:28 | examples-service-types | 2025-02-03 07:46:05 +0000 UTC Normal Pod service-types-665ffcc57c-q5m29.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:46:28 | examples-service-types | 2025-02-03 07:46:05 +0000 UTC Normal Pod service-types-665ffcc57c-q5m29.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" already present on machine kubelet logger.go:42: 07:46:28 | examples-service-types | 2025-02-03 07:46:05 +0000 UTC Normal Pod service-types-665ffcc57c-q5m29.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:46:28 | examples-service-types | 2025-02-03 07:46:05 +0000 UTC Normal Pod service-types-665ffcc57c-q5m29.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:46:28 | examples-service-types | 2025-02-03 07:46:07 +0000 UTC Normal Service service-types-collector EnsuredLoadBalancer Ensured load balancer service-controller logger.go:42: 07:46:28 | examples-service-types | 2025-02-03 07:46:07 +0000 UTC Normal Service service-types-query EnsuringLoadBalancer Ensuring load balancer service-controller logger.go:42: 07:46:28 | examples-service-types | 2025-02-03 07:46:10 +0000 UTC Normal Pod service-types-665ffcc57c-q5m29.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:46:28 | examples-service-types | 2025-02-03 07:46:10 +0000 UTC Normal Pod service-types-665ffcc57c-q5m29.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:46:28 | examples-service-types | 2025-02-03 07:46:10 +0000 UTC Normal ReplicaSet.apps service-types-665ffcc57c SuccessfulDelete Deleted pod: service-types-665ffcc57c-q5m29 replicaset-controller logger.go:42: 07:46:28 | examples-service-types | 2025-02-03 07:46:10 +0000 UTC Normal Service service-types-query EnsuredLoadBalancer Ensured load balancer service-controller logger.go:42: 07:46:28 | examples-service-types | 2025-02-03 07:46:10 +0000 UTC Normal Deployment.apps service-types ScalingReplicaSet Scaled down replica set service-types-665ffcc57c to 0 from 1 deployment-controller logger.go:42: 07:46:28 | examples-service-types | 2025-02-03 07:46:11 +0000 UTC Normal Pod service-types-666bd79f9c-9p5ww Binding Scheduled Successfully assigned kuttl-test-holy-midge/service-types-666bd79f9c-9p5ww to ip-10-0-66-98.us-east-2.compute.internal default-scheduler logger.go:42: 07:46:28 | examples-service-types | 2025-02-03 07:46:11 +0000 UTC Normal ReplicaSet.apps service-types-666bd79f9c SuccessfulCreate Created pod: service-types-666bd79f9c-9p5ww replicaset-controller logger.go:42: 07:46:28 | examples-service-types | 2025-02-03 07:46:11 +0000 UTC Normal Deployment.apps service-types ScalingReplicaSet Scaled up replica set service-types-666bd79f9c to 1 deployment-controller logger.go:42: 07:46:28 | examples-service-types | 2025-02-03 07:46:12 +0000 UTC Normal Pod service-types-666bd79f9c-9p5ww AddedInterface Add eth0 [10.129.2.54/23] from ovn-kubernetes multus logger.go:42: 07:46:28 | examples-service-types | 2025-02-03 07:46:12 +0000 UTC Normal Pod service-types-666bd79f9c-9p5ww.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:fa9aaf9ae026228265ce7f10fb451d5c52a4e0269ce1fa9024cbefddf765ae8d" already present on machine kubelet logger.go:42: 07:46:28 | examples-service-types | 2025-02-03 07:46:12 +0000 UTC Normal Pod service-types-666bd79f9c-9p5ww.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:46:28 | examples-service-types | 2025-02-03 07:46:12 +0000 UTC Normal Pod service-types-666bd79f9c-9p5ww.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:46:28 | examples-service-types | 2025-02-03 07:46:12 +0000 UTC Normal Pod service-types-666bd79f9c-9p5ww.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" already present on machine kubelet logger.go:42: 07:46:28 | examples-service-types | 2025-02-03 07:46:12 +0000 UTC Normal Pod service-types-666bd79f9c-9p5ww.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:46:28 | examples-service-types | 2025-02-03 07:46:12 +0000 UTC Normal Pod service-types-666bd79f9c-9p5ww.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:46:28 | examples-service-types | 2025-02-03 07:46:15 +0000 UTC Normal Pod check-span-pbfx9 Binding Scheduled Successfully assigned kuttl-test-holy-midge/check-span-pbfx9 to ip-10-0-26-111.us-east-2.compute.internal default-scheduler logger.go:42: 07:46:28 | examples-service-types | 2025-02-03 07:46:15 +0000 UTC Normal Pod check-span-pbfx9 AddedInterface Add eth0 [10.128.2.62/23] from ovn-kubernetes multus logger.go:42: 07:46:28 | examples-service-types | 2025-02-03 07:46:15 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-pbfx9 job-controller logger.go:42: 07:46:28 | examples-service-types | 2025-02-03 07:46:15 +0000 UTC Normal Pod report-span-mjkvc Binding Scheduled Successfully assigned kuttl-test-holy-midge/report-span-mjkvc to ip-10-0-6-172.us-east-2.compute.internal default-scheduler logger.go:42: 07:46:28 | examples-service-types | 2025-02-03 07:46:15 +0000 UTC Normal Pod report-span-mjkvc AddedInterface Add eth0 [10.131.0.44/23] from ovn-kubernetes multus logger.go:42: 07:46:28 | examples-service-types | 2025-02-03 07:46:15 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-mjkvc job-controller logger.go:42: 07:46:28 | examples-service-types | 2025-02-03 07:46:16 +0000 UTC Normal Pod check-span-pbfx9.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:46:28 | examples-service-types | 2025-02-03 07:46:16 +0000 UTC Normal Pod check-span-pbfx9.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 794ms (794ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:46:28 | examples-service-types | 2025-02-03 07:46:16 +0000 UTC Normal Pod check-span-pbfx9.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:46:28 | examples-service-types | 2025-02-03 07:46:16 +0000 UTC Normal Pod check-span-pbfx9.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:46:28 | examples-service-types | 2025-02-03 07:46:16 +0000 UTC Normal Pod report-span-mjkvc.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:46:28 | examples-service-types | 2025-02-03 07:46:16 +0000 UTC Normal Pod report-span-mjkvc.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 618ms (618ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:46:28 | examples-service-types | 2025-02-03 07:46:16 +0000 UTC Normal Pod report-span-mjkvc.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:46:28 | examples-service-types | 2025-02-03 07:46:16 +0000 UTC Normal Pod report-span-mjkvc.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:46:28 | examples-service-types | 2025-02-03 07:46:27 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:46:28 | examples-service-types | Deleting namespace: kuttl-test-holy-midge === CONT kuttl/harness/examples-openshift-with-htpasswd logger.go:42: 07:46:56 | examples-openshift-with-htpasswd | Ignoring 00-install.yaml.template as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:46:56 | examples-openshift-with-htpasswd | Ignoring ensure-ingress-host.sh as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:46:56 | examples-openshift-with-htpasswd | Creating namespace: kuttl-test-apt-wombat logger.go:42: 07:46:56 | examples-openshift-with-htpasswd/0-install | starting test step 0-install logger.go:42: 07:46:56 | examples-openshift-with-htpasswd/0-install | Secret:kuttl-test-apt-wombat/htpasswd created logger.go:42: 07:46:56 | examples-openshift-with-htpasswd/0-install | test step completed 0-install logger.go:42: 07:46:56 | examples-openshift-with-htpasswd/1-install | starting test step 1-install logger.go:42: 07:46:56 | examples-openshift-with-htpasswd/1-install | Jaeger:kuttl-test-apt-wombat/with-htpasswd created logger.go:42: 07:47:03 | examples-openshift-with-htpasswd/1-install | test step completed 1-install logger.go:42: 07:47:03 | examples-openshift-with-htpasswd/2-check-unsecured | starting test step 2-check-unsecured logger.go:42: 07:47:03 | examples-openshift-with-htpasswd/2-check-unsecured | running command: [./ensure-ingress-host.sh] logger.go:42: 07:47:03 | examples-openshift-with-htpasswd/2-check-unsecured | Checking the Ingress host value was populated logger.go:42: 07:47:03 | examples-openshift-with-htpasswd/2-check-unsecured | Try number 0 logger.go:42: 07:47:03 | examples-openshift-with-htpasswd/2-check-unsecured | Hostname is with-htpasswd-kuttl-test-apt-wombat.apps.ci-op-d102g9k0-f1a62.cspilp.interop.ccitredhat.com logger.go:42: 07:47:03 | examples-openshift-with-htpasswd/2-check-unsecured | running command: [sh -c INSECURE=true ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 403 true $NAMESPACE with-htpasswd] logger.go:42: 07:47:03 | examples-openshift-with-htpasswd/2-check-unsecured | Checking an expected HTTP response logger.go:42: 07:47:03 | examples-openshift-with-htpasswd/2-check-unsecured | Running in OpenShift logger.go:42: 07:47:03 | examples-openshift-with-htpasswd/2-check-unsecured | Not using any secret logger.go:42: 07:47:03 | examples-openshift-with-htpasswd/2-check-unsecured | Try number 1/30 the https://with-htpasswd-kuttl-test-apt-wombat.apps.ci-op-d102g9k0-f1a62.cspilp.interop.ccitredhat.com/search logger.go:42: 07:47:04 | examples-openshift-with-htpasswd/2-check-unsecured | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 07:47:04 | examples-openshift-with-htpasswd/2-check-unsecured | Try number 2/30 the https://with-htpasswd-kuttl-test-apt-wombat.apps.ci-op-d102g9k0-f1a62.cspilp.interop.ccitredhat.com/search logger.go:42: 07:47:04 | examples-openshift-with-htpasswd/2-check-unsecured | curl response asserted properly logger.go:42: 07:47:04 | examples-openshift-with-htpasswd/2-check-unsecured | test step completed 2-check-unsecured logger.go:42: 07:47:04 | examples-openshift-with-htpasswd/3-check-unauthorized | starting test step 3-check-unauthorized logger.go:42: 07:47:04 | examples-openshift-with-htpasswd/3-check-unauthorized | running command: [./ensure-ingress-host.sh] logger.go:42: 07:47:04 | examples-openshift-with-htpasswd/3-check-unauthorized | Checking the Ingress host value was populated logger.go:42: 07:47:04 | examples-openshift-with-htpasswd/3-check-unauthorized | Try number 0 logger.go:42: 07:47:04 | examples-openshift-with-htpasswd/3-check-unauthorized | Hostname is with-htpasswd-kuttl-test-apt-wombat.apps.ci-op-d102g9k0-f1a62.cspilp.interop.ccitredhat.com logger.go:42: 07:47:04 | examples-openshift-with-htpasswd/3-check-unauthorized | running command: [sh -c JAEGER_USERNAME=wronguser JAEGER_PASSWORD=wrongpassword ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 403 true $NAMESPACE with-htpasswd] logger.go:42: 07:47:04 | examples-openshift-with-htpasswd/3-check-unauthorized | Checking an expected HTTP response logger.go:42: 07:47:04 | examples-openshift-with-htpasswd/3-check-unauthorized | Running in OpenShift logger.go:42: 07:47:04 | examples-openshift-with-htpasswd/3-check-unauthorized | Using Jaeger basic authentication logger.go:42: 07:47:04 | examples-openshift-with-htpasswd/3-check-unauthorized | Try number 1/30 the https://with-htpasswd-kuttl-test-apt-wombat.apps.ci-op-d102g9k0-f1a62.cspilp.interop.ccitredhat.com/search logger.go:42: 07:47:04 | examples-openshift-with-htpasswd/3-check-unauthorized | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 07:47:04 | examples-openshift-with-htpasswd/3-check-unauthorized | Try number 2/30 the https://with-htpasswd-kuttl-test-apt-wombat.apps.ci-op-d102g9k0-f1a62.cspilp.interop.ccitredhat.com/search logger.go:42: 07:47:04 | examples-openshift-with-htpasswd/3-check-unauthorized | curl response asserted properly logger.go:42: 07:47:04 | examples-openshift-with-htpasswd/3-check-unauthorized | test step completed 3-check-unauthorized logger.go:42: 07:47:04 | examples-openshift-with-htpasswd/4-check-authorized | starting test step 4-check-authorized logger.go:42: 07:47:04 | examples-openshift-with-htpasswd/4-check-authorized | running command: [./ensure-ingress-host.sh] logger.go:42: 07:47:04 | examples-openshift-with-htpasswd/4-check-authorized | Checking the Ingress host value was populated logger.go:42: 07:47:04 | examples-openshift-with-htpasswd/4-check-authorized | Try number 0 logger.go:42: 07:47:04 | examples-openshift-with-htpasswd/4-check-authorized | Hostname is with-htpasswd-kuttl-test-apt-wombat.apps.ci-op-d102g9k0-f1a62.cspilp.interop.ccitredhat.com logger.go:42: 07:47:04 | examples-openshift-with-htpasswd/4-check-authorized | running command: [sh -c JAEGER_USERNAME=awesomeuser JAEGER_PASSWORD=awesomepassword ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 200 true $NAMESPACE with-htpasswd] logger.go:42: 07:47:04 | examples-openshift-with-htpasswd/4-check-authorized | Checking an expected HTTP response logger.go:42: 07:47:04 | examples-openshift-with-htpasswd/4-check-authorized | Running in OpenShift logger.go:42: 07:47:04 | examples-openshift-with-htpasswd/4-check-authorized | Using Jaeger basic authentication logger.go:42: 07:47:04 | examples-openshift-with-htpasswd/4-check-authorized | Try number 1/30 the https://with-htpasswd-kuttl-test-apt-wombat.apps.ci-op-d102g9k0-f1a62.cspilp.interop.ccitredhat.com/search logger.go:42: 07:47:04 | examples-openshift-with-htpasswd/4-check-authorized | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 07:47:04 | examples-openshift-with-htpasswd/4-check-authorized | Try number 2/30 the https://with-htpasswd-kuttl-test-apt-wombat.apps.ci-op-d102g9k0-f1a62.cspilp.interop.ccitredhat.com/search logger.go:42: 07:47:04 | examples-openshift-with-htpasswd/4-check-authorized | curl response asserted properly logger.go:42: 07:47:04 | examples-openshift-with-htpasswd/4-check-authorized | test step completed 4-check-authorized logger.go:42: 07:47:04 | examples-openshift-with-htpasswd | examples-openshift-with-htpasswd events from ns kuttl-test-apt-wombat: logger.go:42: 07:47:04 | examples-openshift-with-htpasswd | 2025-02-03 07:47:00 +0000 UTC Normal Pod with-htpasswd-757bbb788-vg4q9 Binding Scheduled Successfully assigned kuttl-test-apt-wombat/with-htpasswd-757bbb788-vg4q9 to ip-10-0-66-98.us-east-2.compute.internal default-scheduler logger.go:42: 07:47:04 | examples-openshift-with-htpasswd | 2025-02-03 07:47:00 +0000 UTC Normal ReplicaSet.apps with-htpasswd-757bbb788 SuccessfulCreate Created pod: with-htpasswd-757bbb788-vg4q9 replicaset-controller logger.go:42: 07:47:04 | examples-openshift-with-htpasswd | 2025-02-03 07:47:00 +0000 UTC Normal Deployment.apps with-htpasswd ScalingReplicaSet Scaled up replica set with-htpasswd-757bbb788 to 1 deployment-controller logger.go:42: 07:47:04 | examples-openshift-with-htpasswd | 2025-02-03 07:47:01 +0000 UTC Normal Pod with-htpasswd-757bbb788-vg4q9 AddedInterface Add eth0 [10.129.2.55/23] from ovn-kubernetes multus logger.go:42: 07:47:04 | examples-openshift-with-htpasswd | 2025-02-03 07:47:01 +0000 UTC Normal Pod with-htpasswd-757bbb788-vg4q9.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:fa9aaf9ae026228265ce7f10fb451d5c52a4e0269ce1fa9024cbefddf765ae8d" already present on machine kubelet logger.go:42: 07:47:04 | examples-openshift-with-htpasswd | 2025-02-03 07:47:01 +0000 UTC Normal Pod with-htpasswd-757bbb788-vg4q9.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:47:04 | examples-openshift-with-htpasswd | 2025-02-03 07:47:01 +0000 UTC Normal Pod with-htpasswd-757bbb788-vg4q9.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:47:04 | examples-openshift-with-htpasswd | 2025-02-03 07:47:01 +0000 UTC Normal Pod with-htpasswd-757bbb788-vg4q9.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" already present on machine kubelet logger.go:42: 07:47:04 | examples-openshift-with-htpasswd | 2025-02-03 07:47:01 +0000 UTC Normal Pod with-htpasswd-757bbb788-vg4q9.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:47:04 | examples-openshift-with-htpasswd | 2025-02-03 07:47:01 +0000 UTC Normal Pod with-htpasswd-757bbb788-vg4q9.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:47:04 | examples-openshift-with-htpasswd | Deleting namespace: kuttl-test-apt-wombat === CONT kuttl/harness/examples-all-in-one-with-options logger.go:42: 07:47:11 | examples-all-in-one-with-options | Creating namespace: kuttl-test-proven-troll logger.go:42: 07:47:11 | examples-all-in-one-with-options/0-install | starting test step 0-install logger.go:42: 07:47:11 | examples-all-in-one-with-options/0-install | Jaeger:kuttl-test-proven-troll/my-jaeger created logger.go:42: 07:47:18 | examples-all-in-one-with-options/0-install | test step completed 0-install logger.go:42: 07:47:18 | examples-all-in-one-with-options/1-smoke-test | starting test step 1-smoke-test logger.go:42: 07:47:18 | examples-all-in-one-with-options/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:47:19 | examples-all-in-one-with-options/1-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:47:26 | examples-all-in-one-with-options/1-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443/jaeger MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:47:26 | examples-all-in-one-with-options/1-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:47:27 | examples-all-in-one-with-options/1-smoke-test | job.batch/report-span created logger.go:42: 07:47:27 | examples-all-in-one-with-options/1-smoke-test | job.batch/check-span created logger.go:42: 07:47:39 | examples-all-in-one-with-options/1-smoke-test | test step completed 1-smoke-test logger.go:42: 07:47:39 | examples-all-in-one-with-options | examples-all-in-one-with-options events from ns kuttl-test-proven-troll: logger.go:42: 07:47:39 | examples-all-in-one-with-options | 2025-02-03 07:47:15 +0000 UTC Normal Pod my-jaeger-586669bf7b-44fv8 Binding Scheduled Successfully assigned kuttl-test-proven-troll/my-jaeger-586669bf7b-44fv8 to ip-10-0-66-98.us-east-2.compute.internal default-scheduler logger.go:42: 07:47:39 | examples-all-in-one-with-options | 2025-02-03 07:47:15 +0000 UTC Normal ReplicaSet.apps my-jaeger-586669bf7b SuccessfulCreate Created pod: my-jaeger-586669bf7b-44fv8 replicaset-controller logger.go:42: 07:47:39 | examples-all-in-one-with-options | 2025-02-03 07:47:15 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-586669bf7b to 1 deployment-controller logger.go:42: 07:47:39 | examples-all-in-one-with-options | 2025-02-03 07:47:16 +0000 UTC Normal Pod my-jaeger-586669bf7b-44fv8 AddedInterface Add eth0 [10.129.2.56/23] from ovn-kubernetes multus logger.go:42: 07:47:39 | examples-all-in-one-with-options | 2025-02-03 07:47:16 +0000 UTC Normal Pod my-jaeger-586669bf7b-44fv8.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:fa9aaf9ae026228265ce7f10fb451d5c52a4e0269ce1fa9024cbefddf765ae8d" already present on machine kubelet logger.go:42: 07:47:39 | examples-all-in-one-with-options | 2025-02-03 07:47:16 +0000 UTC Normal Pod my-jaeger-586669bf7b-44fv8.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:47:39 | examples-all-in-one-with-options | 2025-02-03 07:47:16 +0000 UTC Normal Pod my-jaeger-586669bf7b-44fv8.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:47:39 | examples-all-in-one-with-options | 2025-02-03 07:47:16 +0000 UTC Normal Pod my-jaeger-586669bf7b-44fv8.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" already present on machine kubelet logger.go:42: 07:47:39 | examples-all-in-one-with-options | 2025-02-03 07:47:16 +0000 UTC Normal Pod my-jaeger-586669bf7b-44fv8.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:47:39 | examples-all-in-one-with-options | 2025-02-03 07:47:16 +0000 UTC Normal Pod my-jaeger-586669bf7b-44fv8.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:47:39 | examples-all-in-one-with-options | 2025-02-03 07:47:21 +0000 UTC Normal Pod my-jaeger-586669bf7b-44fv8.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:47:39 | examples-all-in-one-with-options | 2025-02-03 07:47:21 +0000 UTC Normal Pod my-jaeger-586669bf7b-44fv8.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:47:39 | examples-all-in-one-with-options | 2025-02-03 07:47:21 +0000 UTC Normal ReplicaSet.apps my-jaeger-586669bf7b SuccessfulDelete Deleted pod: my-jaeger-586669bf7b-44fv8 replicaset-controller logger.go:42: 07:47:39 | examples-all-in-one-with-options | 2025-02-03 07:47:21 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled down replica set my-jaeger-586669bf7b to 0 from 1 deployment-controller logger.go:42: 07:47:39 | examples-all-in-one-with-options | 2025-02-03 07:47:22 +0000 UTC Normal Pod my-jaeger-79fb4585dc-jnvw5 Binding Scheduled Successfully assigned kuttl-test-proven-troll/my-jaeger-79fb4585dc-jnvw5 to ip-10-0-66-98.us-east-2.compute.internal default-scheduler logger.go:42: 07:47:39 | examples-all-in-one-with-options | 2025-02-03 07:47:22 +0000 UTC Normal ReplicaSet.apps my-jaeger-79fb4585dc SuccessfulCreate Created pod: my-jaeger-79fb4585dc-jnvw5 replicaset-controller logger.go:42: 07:47:39 | examples-all-in-one-with-options | 2025-02-03 07:47:22 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-79fb4585dc to 1 deployment-controller logger.go:42: 07:47:39 | examples-all-in-one-with-options | 2025-02-03 07:47:23 +0000 UTC Normal Pod my-jaeger-79fb4585dc-jnvw5 AddedInterface Add eth0 [10.129.2.57/23] from ovn-kubernetes multus logger.go:42: 07:47:39 | examples-all-in-one-with-options | 2025-02-03 07:47:23 +0000 UTC Normal Pod my-jaeger-79fb4585dc-jnvw5.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:fa9aaf9ae026228265ce7f10fb451d5c52a4e0269ce1fa9024cbefddf765ae8d" already present on machine kubelet logger.go:42: 07:47:39 | examples-all-in-one-with-options | 2025-02-03 07:47:23 +0000 UTC Normal Pod my-jaeger-79fb4585dc-jnvw5.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:47:39 | examples-all-in-one-with-options | 2025-02-03 07:47:23 +0000 UTC Normal Pod my-jaeger-79fb4585dc-jnvw5.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:47:39 | examples-all-in-one-with-options | 2025-02-03 07:47:23 +0000 UTC Normal Pod my-jaeger-79fb4585dc-jnvw5.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" already present on machine kubelet logger.go:42: 07:47:39 | examples-all-in-one-with-options | 2025-02-03 07:47:23 +0000 UTC Normal Pod my-jaeger-79fb4585dc-jnvw5.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:47:39 | examples-all-in-one-with-options | 2025-02-03 07:47:23 +0000 UTC Normal Pod my-jaeger-79fb4585dc-jnvw5.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:47:39 | examples-all-in-one-with-options | 2025-02-03 07:47:27 +0000 UTC Normal Pod check-span-ffbhj Binding Scheduled Successfully assigned kuttl-test-proven-troll/check-span-ffbhj to ip-10-0-26-111.us-east-2.compute.internal default-scheduler logger.go:42: 07:47:39 | examples-all-in-one-with-options | 2025-02-03 07:47:27 +0000 UTC Normal Pod check-span-ffbhj AddedInterface Add eth0 [10.128.2.63/23] from ovn-kubernetes multus logger.go:42: 07:47:39 | examples-all-in-one-with-options | 2025-02-03 07:47:27 +0000 UTC Normal Pod check-span-ffbhj.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:47:39 | examples-all-in-one-with-options | 2025-02-03 07:47:27 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-ffbhj job-controller logger.go:42: 07:47:39 | examples-all-in-one-with-options | 2025-02-03 07:47:27 +0000 UTC Normal Pod report-span-x69rd Binding Scheduled Successfully assigned kuttl-test-proven-troll/report-span-x69rd to ip-10-0-6-172.us-east-2.compute.internal default-scheduler logger.go:42: 07:47:39 | examples-all-in-one-with-options | 2025-02-03 07:47:27 +0000 UTC Normal Pod report-span-x69rd AddedInterface Add eth0 [10.131.0.45/23] from ovn-kubernetes multus logger.go:42: 07:47:39 | examples-all-in-one-with-options | 2025-02-03 07:47:27 +0000 UTC Normal Pod report-span-x69rd.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:47:39 | examples-all-in-one-with-options | 2025-02-03 07:47:27 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-x69rd job-controller logger.go:42: 07:47:39 | examples-all-in-one-with-options | 2025-02-03 07:47:28 +0000 UTC Normal Pod check-span-ffbhj.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 612ms (612ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:47:39 | examples-all-in-one-with-options | 2025-02-03 07:47:28 +0000 UTC Normal Pod check-span-ffbhj.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:47:39 | examples-all-in-one-with-options | 2025-02-03 07:47:28 +0000 UTC Normal Pod check-span-ffbhj.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:47:39 | examples-all-in-one-with-options | 2025-02-03 07:47:28 +0000 UTC Normal Pod report-span-x69rd.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 663ms (663ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:47:39 | examples-all-in-one-with-options | 2025-02-03 07:47:28 +0000 UTC Normal Pod report-span-x69rd.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:47:39 | examples-all-in-one-with-options | 2025-02-03 07:47:28 +0000 UTC Normal Pod report-span-x69rd.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:47:39 | examples-all-in-one-with-options | 2025-02-03 07:47:38 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:47:39 | examples-all-in-one-with-options | Deleting namespace: kuttl-test-proven-troll === CONT kuttl/harness/examples-agent-with-priority-class logger.go:42: 07:47:52 | examples-agent-with-priority-class | Creating namespace: kuttl-test-perfect-lynx logger.go:42: 07:47:52 | examples-agent-with-priority-class/0-install | starting test step 0-install logger.go:42: 07:47:52 | examples-agent-with-priority-class/0-install | SecurityContextConstraints:/daemonset-with-hostport created logger.go:42: 07:47:52 | examples-agent-with-priority-class/0-install | ServiceAccount:kuttl-test-perfect-lynx/jaeger-agent-daemonset created logger.go:42: 07:47:52 | examples-agent-with-priority-class/0-install | test step completed 0-install logger.go:42: 07:47:52 | examples-agent-with-priority-class/1-add-policy | starting test step 1-add-policy logger.go:42: 07:47:52 | examples-agent-with-priority-class/1-add-policy | running command: [sh -c oc adm policy --namespace $NAMESPACE add-scc-to-user daemonset-with-hostport -z jaeger-agent-daemonset] logger.go:42: 07:47:52 | examples-agent-with-priority-class/1-add-policy | clusterrole.rbac.authorization.k8s.io/system:openshift:scc:daemonset-with-hostport added: "jaeger-agent-daemonset" logger.go:42: 07:47:52 | examples-agent-with-priority-class/1-add-policy | running command: [sh -c sleep 5] logger.go:42: 07:47:57 | examples-agent-with-priority-class/1-add-policy | test step completed 1-add-policy logger.go:42: 07:47:57 | examples-agent-with-priority-class/2-install | starting test step 2-install logger.go:42: 07:47:57 | examples-agent-with-priority-class/2-install | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE agent-as-daemonset /dev/null] logger.go:42: 07:48:05 | examples-agent-with-priority-class/2-install | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://agent-as-daemonset-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://agent-as-daemonset-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:48:06 | examples-agent-with-priority-class/2-install | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:48:06 | examples-agent-with-priority-class/2-install | job.batch/report-span created logger.go:42: 07:48:06 | examples-agent-with-priority-class/2-install | job.batch/check-span created logger.go:42: 07:48:06 | examples-agent-with-priority-class/2-install | PriorityClass:/high-priority created logger.go:42: 07:48:06 | examples-agent-with-priority-class/2-install | Jaeger:kuttl-test-perfect-lynx/agent-as-daemonset updated logger.go:42: 07:48:19 | examples-agent-with-priority-class/2-install | test step completed 2-install logger.go:42: 07:48:19 | examples-agent-with-priority-class | examples-agent-with-priority-class events from ns kuttl-test-perfect-lynx: logger.go:42: 07:48:19 | examples-agent-with-priority-class | 2025-02-03 07:48:03 +0000 UTC Normal Pod agent-as-daemonset-6d474c7f5b-4x9f7 Binding Scheduled Successfully assigned kuttl-test-perfect-lynx/agent-as-daemonset-6d474c7f5b-4x9f7 to ip-10-0-66-98.us-east-2.compute.internal default-scheduler logger.go:42: 07:48:19 | examples-agent-with-priority-class | 2025-02-03 07:48:03 +0000 UTC Warning Pod agent-as-daemonset-6d474c7f5b-4x9f7 FailedMount MountVolume.SetUp failed for volume "agent-as-daemonset-collector-tls-config-volume" : secret "agent-as-daemonset-collector-headless-tls" not found kubelet logger.go:42: 07:48:19 | examples-agent-with-priority-class | 2025-02-03 07:48:03 +0000 UTC Normal ReplicaSet.apps agent-as-daemonset-6d474c7f5b SuccessfulCreate Created pod: agent-as-daemonset-6d474c7f5b-4x9f7 replicaset-controller logger.go:42: 07:48:19 | examples-agent-with-priority-class | 2025-02-03 07:48:03 +0000 UTC Normal Deployment.apps agent-as-daemonset ScalingReplicaSet Scaled up replica set agent-as-daemonset-6d474c7f5b to 1 deployment-controller logger.go:42: 07:48:19 | examples-agent-with-priority-class | 2025-02-03 07:48:04 +0000 UTC Normal Pod agent-as-daemonset-6d474c7f5b-4x9f7 AddedInterface Add eth0 [10.129.2.58/23] from ovn-kubernetes multus logger.go:42: 07:48:19 | examples-agent-with-priority-class | 2025-02-03 07:48:04 +0000 UTC Normal Pod agent-as-daemonset-6d474c7f5b-4x9f7.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:fa9aaf9ae026228265ce7f10fb451d5c52a4e0269ce1fa9024cbefddf765ae8d" already present on machine kubelet logger.go:42: 07:48:19 | examples-agent-with-priority-class | 2025-02-03 07:48:04 +0000 UTC Normal Pod agent-as-daemonset-6d474c7f5b-4x9f7.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:48:19 | examples-agent-with-priority-class | 2025-02-03 07:48:04 +0000 UTC Normal Pod agent-as-daemonset-6d474c7f5b-4x9f7.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:48:19 | examples-agent-with-priority-class | 2025-02-03 07:48:04 +0000 UTC Normal Pod agent-as-daemonset-6d474c7f5b-4x9f7.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" already present on machine kubelet logger.go:42: 07:48:19 | examples-agent-with-priority-class | 2025-02-03 07:48:04 +0000 UTC Normal Pod agent-as-daemonset-6d474c7f5b-4x9f7.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:48:19 | examples-agent-with-priority-class | 2025-02-03 07:48:04 +0000 UTC Normal Pod agent-as-daemonset-6d474c7f5b-4x9f7.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:48:19 | examples-agent-with-priority-class | 2025-02-03 07:48:06 +0000 UTC Normal Pod check-span-9r796 Binding Scheduled Successfully assigned kuttl-test-perfect-lynx/check-span-9r796 to ip-10-0-26-111.us-east-2.compute.internal default-scheduler logger.go:42: 07:48:19 | examples-agent-with-priority-class | 2025-02-03 07:48:06 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-9r796 job-controller logger.go:42: 07:48:19 | examples-agent-with-priority-class | 2025-02-03 07:48:06 +0000 UTC Normal Pod report-span-2qb5w Binding Scheduled Successfully assigned kuttl-test-perfect-lynx/report-span-2qb5w to ip-10-0-6-172.us-east-2.compute.internal default-scheduler logger.go:42: 07:48:19 | examples-agent-with-priority-class | 2025-02-03 07:48:06 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-2qb5w job-controller logger.go:42: 07:48:19 | examples-agent-with-priority-class | 2025-02-03 07:48:07 +0000 UTC Normal Pod check-span-9r796 AddedInterface Add eth0 [10.128.2.64/23] from ovn-kubernetes multus logger.go:42: 07:48:19 | examples-agent-with-priority-class | 2025-02-03 07:48:07 +0000 UTC Normal Pod check-span-9r796.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:48:19 | examples-agent-with-priority-class | 2025-02-03 07:48:07 +0000 UTC Normal Pod report-span-2qb5w AddedInterface Add eth0 [10.131.0.46/23] from ovn-kubernetes multus logger.go:42: 07:48:19 | examples-agent-with-priority-class | 2025-02-03 07:48:07 +0000 UTC Normal Pod report-span-2qb5w.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:48:19 | examples-agent-with-priority-class | 2025-02-03 07:48:07 +0000 UTC Normal Pod report-span-2qb5w.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 576ms (576ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:48:19 | examples-agent-with-priority-class | 2025-02-03 07:48:08 +0000 UTC Warning DaemonSet.apps agent-as-daemonset-agent-daemonset FailedCreate Error creating: pods "agent-as-daemonset-agent-daemonset-" is forbidden: unable to validate against any security context constraint: [provider "anyuid": Forbidden: not usable by user or serviceaccount, provider restricted-v2: .containers[0].containers[0].hostPort: Invalid value: 5775: Host ports are not allowed to be used, provider restricted-v2: .containers[0].containers[0].hostPort: Invalid value: 5778: Host ports are not allowed to be used, provider restricted-v2: .containers[0].containers[0].hostPort: Invalid value: 6831: Host ports are not allowed to be used, provider restricted-v2: .containers[0].containers[0].hostPort: Invalid value: 6832: Host ports are not allowed to be used, provider restricted-v2: .containers[0].containers[0].hostPort: Invalid value: 14271: Host ports are not allowed to be used, provider "restricted": Forbidden: not usable by user or serviceaccount, provider "nonroot-v2": Forbidden: not usable by user or serviceaccount, provider "nonroot": Forbidden: not usable by user or serviceaccount, provider "hostmount-anyuid": Forbidden: not usable by user or serviceaccount, provider "elasticsearch-scc": Forbidden: not usable by user or serviceaccount, provider "machine-api-termination-handler": Forbidden: not usable by user or serviceaccount, provider "daemonset-with-hostport": Forbidden: not usable by user or serviceaccount, provider "hostnetwork-v2": Forbidden: not usable by user or serviceaccount, provider "hostnetwork": Forbidden: not usable by user or serviceaccount, provider "hostaccess": Forbidden: not usable by user or serviceaccount, provider "node-exporter": Forbidden: not usable by user or serviceaccount, provider "privileged": Forbidden: not usable by user or serviceaccount] daemonset-controller logger.go:42: 07:48:19 | examples-agent-with-priority-class | 2025-02-03 07:48:08 +0000 UTC Normal Pod check-span-9r796.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 658ms (658ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:48:19 | examples-agent-with-priority-class | 2025-02-03 07:48:08 +0000 UTC Normal Pod check-span-9r796.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:48:19 | examples-agent-with-priority-class | 2025-02-03 07:48:08 +0000 UTC Normal Pod check-span-9r796.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:48:19 | examples-agent-with-priority-class | 2025-02-03 07:48:08 +0000 UTC Normal Pod report-span-2qb5w.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:48:19 | examples-agent-with-priority-class | 2025-02-03 07:48:08 +0000 UTC Normal Pod report-span-2qb5w.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:48:19 | examples-agent-with-priority-class | 2025-02-03 07:48:18 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:48:19 | examples-agent-with-priority-class | Deleting namespace: kuttl-test-perfect-lynx === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (575.84s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (6.26s) --- PASS: kuttl/harness/examples-simple-prod (81.02s) --- PASS: kuttl/harness/examples-with-sampling (59.00s) --- PASS: kuttl/harness/examples-with-cassandra (55.91s) --- PASS: kuttl/harness/examples-with-badger (42.71s) --- PASS: kuttl/harness/examples-simplest (35.09s) --- PASS: kuttl/harness/examples-simple-prod-with-volumes (96.41s) --- PASS: kuttl/harness/examples-business-application-injected-sidecar (46.62s) --- PASS: kuttl/harness/examples-service-types (56.31s) --- PASS: kuttl/harness/examples-openshift-with-htpasswd (14.61s) --- PASS: kuttl/harness/examples-all-in-one-with-options (41.45s) --- PASS: kuttl/harness/examples-agent-with-priority-class (40.39s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name examples --report --output /logs/artifacts/examples.xml ./artifacts/kuttl-report.xml time="2025-02-03T07:48:33Z" level=debug msg="Setting a new name for the test suites" time="2025-02-03T07:48:33Z" level=debug msg="Removing 'artifacts' TestCase" time="2025-02-03T07:48:33Z" level=debug msg="normalizing test case names" time="2025-02-03T07:48:33Z" level=debug msg="examples/artifacts -> examples_artifacts" time="2025-02-03T07:48:33Z" level=debug msg="examples/examples-simple-prod -> examples_examples_simple_prod" time="2025-02-03T07:48:33Z" level=debug msg="examples/examples-with-sampling -> examples_examples_with_sampling" time="2025-02-03T07:48:33Z" level=debug msg="examples/examples-with-cassandra -> examples_examples_with_cassandra" time="2025-02-03T07:48:33Z" level=debug msg="examples/examples-with-badger -> examples_examples_with_badger" time="2025-02-03T07:48:33Z" level=debug msg="examples/examples-simplest -> examples_examples_simplest" time="2025-02-03T07:48:33Z" level=debug msg="examples/examples-simple-prod-with-volumes -> examples_examples_simple_prod_with_volumes" time="2025-02-03T07:48:33Z" level=debug msg="examples/examples-business-application-injected-sidecar -> examples_examples_business_application_injected_sidecar" time="2025-02-03T07:48:33Z" level=debug msg="examples/examples-service-types -> examples_examples_service_types" time="2025-02-03T07:48:33Z" level=debug msg="examples/examples-openshift-with-htpasswd -> examples_examples_openshift_with_htpasswd" time="2025-02-03T07:48:33Z" level=debug msg="examples/examples-all-in-one-with-options -> examples_examples_all_in_one_with_options" time="2025-02-03T07:48:33Z" level=debug msg="examples/examples-agent-with-priority-class -> examples_examples_agent_with_priority_class" +---------------------------------------------------------+--------+ | NAME | RESULT | +---------------------------------------------------------+--------+ | examples_artifacts | passed | | examples_examples_simple_prod | passed | | examples_examples_with_sampling | passed | | examples_examples_with_cassandra | passed | | examples_examples_with_badger | passed | | examples_examples_simplest | passed | | examples_examples_simple_prod_with_volumes | passed | | examples_examples_business_application_injected_sidecar | passed | | examples_examples_service_types | passed | | examples_examples_openshift_with_htpasswd | passed | | examples_examples_all_in_one_with_options | passed | | examples_examples_agent_with_priority_class | passed | +---------------------------------------------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh generate false true + '[' 3 -ne 3 ']' + test_suite_name=generate + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/generate.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-generate make[2]: Entering directory '/tmp/jaeger-tests' test -s /tmp/jaeger-tests/bin/operator-sdk || curl -sLo /tmp/jaeger-tests/bin/operator-sdk https://github.com/operator-framework/operator-sdk/releases/download/v1.32.0/operator-sdk_`go env GOOS`_`go env GOARCH` ./hack/install/install-golangci-lint.sh Installing golangci-lint Try 0... go install github.com/golangci/golangci-lint/cmd/golangci-lint@v1.55.2 ./hack/install/install-goimports.sh Installing goimports Try 0... go install golang.org/x/tools/cmd/goimports@v0.1.12 >>>> Formatting code... ./.ci/format.sh >>>> Building... ./hack/install/install-dependencies.sh Installing go dependencies Try 0... go mod download GOOS= GOARCH= CGO_ENABLED=0 GO111MODULE=on go build -ldflags "-X "github.com/jaegertracing/jaeger-operator/pkg/version".version="1.62.0" -X "github.com/jaegertracing/jaeger-operator/pkg/version".buildDate=2025-02-03T07:48:38Z -X "github.com/jaegertracing/jaeger-operator/pkg/version".defaultJaeger="1.62.0"" -o "bin/jaeger-operator" main.go JAEGER_VERSION="1.62.0" ./tests/e2e/generate/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.18.0-0.nightly-2025-02-01-185117 True False 40m Cluster version is 4.18.0-0.nightly-2025-02-01-185117' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.18.0-0.nightly-2025-02-01-185117 True False 40m Cluster version is 4.18.0-0.nightly-2025-02-01-185117' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/generate/render.sh ++ export SUITE_DIR=./tests/e2e/generate ++ SUITE_DIR=./tests/e2e/generate ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/generate ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + '[' true = true ']' + skip_test generate 'This test was skipped until https://github.com/jaegertracing/jaeger-operator/issues/2145 is fixed' + '[' 2 -ne 2 ']' + test_name=generate + message='This test was skipped until https://github.com/jaegertracing/jaeger-operator/issues/2145 is fixed' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/generate/_build + '[' _build '!=' _build ']' + rm -rf generate + warning 'generate: This test was skipped until https://github.com/jaegertracing/jaeger-operator/issues/2145 is fixed' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: generate: This test was skipped until https://github.com/jaegertracing/jaeger-operator/issues/2145 is fixed\e[0m' WAR: generate: This test was skipped until https://github.com/jaegertracing/jaeger-operator/issues/2145 is fixed make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running generate E2E tests' Running generate E2E tests + cd tests/e2e/generate/_build + set +e + KUBECONFIG=/tmp/kubeconfig-2531963025 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-d102g9k0-f1a62.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 1 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === CONT kuttl/harness/artifacts logger.go:42: 07:48:53 | artifacts | Creating namespace: kuttl-test-noted-jaguar logger.go:42: 07:48:53 | artifacts | artifacts events from ns kuttl-test-noted-jaguar: logger.go:42: 07:48:53 | artifacts | Deleting namespace: kuttl-test-noted-jaguar === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (6.09s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (6.04s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name generate --report --output /logs/artifacts/generate.xml ./artifacts/kuttl-report.xml time="2025-02-03T07:49:00Z" level=debug msg="Setting a new name for the test suites" time="2025-02-03T07:49:00Z" level=debug msg="Removing 'artifacts' TestCase" time="2025-02-03T07:49:00Z" level=debug msg="normalizing test case names" time="2025-02-03T07:49:00Z" level=debug msg="generate/artifacts -> generate_artifacts" +--------------------+--------+ | NAME | RESULT | +--------------------+--------+ | generate_artifacts | passed | +--------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh miscellaneous false true + '[' 3 -ne 3 ']' + test_suite_name=miscellaneous + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/miscellaneous.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-miscellaneous make[2]: Entering directory '/tmp/jaeger-tests' SKIP_ES_EXTERNAL=true ./tests/e2e/miscellaneous/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.18.0-0.nightly-2025-02-01-185117 True False 40m Cluster version is 4.18.0-0.nightly-2025-02-01-185117' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.18.0-0.nightly-2025-02-01-185117 True False 40m Cluster version is 4.18.0-0.nightly-2025-02-01-185117' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/miscellaneous/render.sh ++ export SUITE_DIR=./tests/e2e/miscellaneous ++ SUITE_DIR=./tests/e2e/miscellaneous ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/miscellaneous ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + start_test collector-autoscale + '[' 1 -ne 1 ']' + test_name=collector-autoscale + echo =========================================================================== =========================================================================== + info 'Rendering files for test collector-autoscale' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test collector-autoscale\e[0m' Rendering files for test collector-autoscale + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build + '[' _build '!=' _build ']' + mkdir -p collector-autoscale + cd collector-autoscale + jaeger_name=simple-prod + jaeger_deploy_mode=production + [[ true = true ]] + [[ true = true ]] + jaeger_deploy_mode=production_autoprovisioned + ELASTICSEARCH_NODECOUNT=1 + render_install_jaeger simple-prod production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=simple-prod + JAEGER_NAME=simple-prod + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.collector.resources.requests.memory="200m"' 01-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.collector.autoscale=true 01-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.collector.minReplicas=1 01-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.collector.maxReplicas=2 01-install.yaml + version_lt 1.30 1.23 ++ echo 1.30 1.23 ++ tr ' ' '\n' ++ sort -rV ++ head -n 1 + test 1.30 '!=' 1.30 + rm ./03-assert.yaml + generate_otlp_e2e_tests http + test_protocol=http + is_secured=false + '[' true = true ']' + is_secured=true + start_test collector-otlp-allinone-http + '[' 1 -ne 1 ']' + test_name=collector-otlp-allinone-http + echo =========================================================================== =========================================================================== + info 'Rendering files for test collector-otlp-allinone-http' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test collector-otlp-allinone-http\e[0m' Rendering files for test collector-otlp-allinone-http + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/collector-autoscale + '[' collector-autoscale '!=' _build ']' + cd .. + mkdir -p collector-otlp-allinone-http + cd collector-otlp-allinone-http + render_install_jaeger my-jaeger allInOne 00 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=allInOne + test_step=00 + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_otlp_smoke_test my-jaeger http true 01 + '[' 4 -ne 4 ']' + jaeger=my-jaeger + reporting_protocol=http + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template + '[' http = grpc ']' + reporting_port=:4318 + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 + OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + REPORTING_PROTOCOL=http + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset OTEL_EXPORTER_OTLP_ENDPOINT + start_test collector-otlp-production-http + '[' 1 -ne 1 ']' + test_name=collector-otlp-production-http + echo =========================================================================== =========================================================================== + info 'Rendering files for test collector-otlp-production-http' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test collector-otlp-production-http\e[0m' Rendering files for test collector-otlp-production-http + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/collector-otlp-allinone-http + '[' collector-otlp-allinone-http '!=' _build ']' + cd .. + mkdir -p collector-otlp-production-http + cd collector-otlp-production-http + jaeger_deploy_mode=production + [[ true = true ]] + [[ true = true ]] + jaeger_deploy_mode=production_autoprovisioned + render_install_jaeger my-jaeger production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + render_otlp_smoke_test my-jaeger http true 02 + '[' 4 -ne 4 ']' + jaeger=my-jaeger + reporting_protocol=http + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template + '[' http = grpc ']' + reporting_port=:4318 + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 + OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + REPORTING_PROTOCOL=http + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset OTEL_EXPORTER_OTLP_ENDPOINT + generate_otlp_e2e_tests grpc + test_protocol=grpc + is_secured=false + '[' true = true ']' + is_secured=true + start_test collector-otlp-allinone-grpc + '[' 1 -ne 1 ']' + test_name=collector-otlp-allinone-grpc + echo =========================================================================== =========================================================================== + info 'Rendering files for test collector-otlp-allinone-grpc' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test collector-otlp-allinone-grpc\e[0m' Rendering files for test collector-otlp-allinone-grpc + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/collector-otlp-production-http + '[' collector-otlp-production-http '!=' _build ']' + cd .. + mkdir -p collector-otlp-allinone-grpc + cd collector-otlp-allinone-grpc + render_install_jaeger my-jaeger allInOne 00 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=allInOne + test_step=00 + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_otlp_smoke_test my-jaeger grpc true 01 + '[' 4 -ne 4 ']' + jaeger=my-jaeger + reporting_protocol=grpc + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template + '[' grpc = grpc ']' + reporting_port=:4317 + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 + OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + REPORTING_PROTOCOL=grpc + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset OTEL_EXPORTER_OTLP_ENDPOINT + start_test collector-otlp-production-grpc + '[' 1 -ne 1 ']' + test_name=collector-otlp-production-grpc + echo =========================================================================== =========================================================================== + info 'Rendering files for test collector-otlp-production-grpc' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test collector-otlp-production-grpc\e[0m' Rendering files for test collector-otlp-production-grpc + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/collector-otlp-allinone-grpc + '[' collector-otlp-allinone-grpc '!=' _build ']' + cd .. + mkdir -p collector-otlp-production-grpc + cd collector-otlp-production-grpc + jaeger_deploy_mode=production + [[ true = true ]] + [[ true = true ]] + jaeger_deploy_mode=production_autoprovisioned + render_install_jaeger my-jaeger production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + render_otlp_smoke_test my-jaeger grpc true 02 + '[' 4 -ne 4 ']' + jaeger=my-jaeger + reporting_protocol=grpc + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template + '[' grpc = grpc ']' + reporting_port=:4317 + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 + OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + REPORTING_PROTOCOL=grpc + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset OTEL_EXPORTER_OTLP_ENDPOINT + '[' true = true ']' + skip_test istio 'Test not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=istio + message='Test not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/collector-otlp-production-grpc + '[' collector-otlp-production-grpc '!=' _build ']' + cd .. + rm -rf istio + warning 'istio: Test not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: istio: Test not supported in OpenShift\e[0m' WAR: istio: Test not supported in OpenShift + '[' true = true ']' + skip_test outside-cluster 'Test not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=outside-cluster + message='Test not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build + '[' _build '!=' _build ']' + rm -rf outside-cluster + warning 'outside-cluster: Test not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: outside-cluster: Test not supported in OpenShift\e[0m' WAR: outside-cluster: Test not supported in OpenShift + start_test set-custom-img + '[' 1 -ne 1 ']' + test_name=set-custom-img + echo =========================================================================== =========================================================================== + info 'Rendering files for test set-custom-img' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test set-custom-img\e[0m' Rendering files for test set-custom-img + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build + '[' _build '!=' _build ']' + mkdir -p set-custom-img + cd set-custom-img + jaeger_name=my-jaeger + jaeger_deploy_mode=production + [[ true = true ]] + [[ true = true ]] + jaeger_deploy_mode=production_autoprovisioned + render_install_jaeger my-jaeger production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + cp ./01-install.yaml ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.collector.image="test"' ./02-install.yaml + '[' true = true ']' + skip_test non-cluster-wide 'Test not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=non-cluster-wide + message='Test not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/set-custom-img + '[' set-custom-img '!=' _build ']' + cd .. + rm -rf non-cluster-wide + warning 'non-cluster-wide: Test not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: non-cluster-wide: Test not supported in OpenShift\e[0m' WAR: non-cluster-wide: Test not supported in OpenShift make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running miscellaneous E2E tests' Running miscellaneous E2E tests + cd tests/e2e/miscellaneous/_build + set +e + KUBECONFIG=/tmp/kubeconfig-2531963025 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-d102g9k0-f1a62.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 8 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/cassandra-spark === PAUSE kuttl/harness/cassandra-spark === RUN kuttl/harness/collector-autoscale === PAUSE kuttl/harness/collector-autoscale === RUN kuttl/harness/collector-otlp-allinone-grpc === PAUSE kuttl/harness/collector-otlp-allinone-grpc === RUN kuttl/harness/collector-otlp-allinone-http === PAUSE kuttl/harness/collector-otlp-allinone-http === RUN kuttl/harness/collector-otlp-production-grpc === PAUSE kuttl/harness/collector-otlp-production-grpc === RUN kuttl/harness/collector-otlp-production-http === PAUSE kuttl/harness/collector-otlp-production-http === RUN kuttl/harness/set-custom-img === PAUSE kuttl/harness/set-custom-img === CONT kuttl/harness/collector-otlp-production-grpc logger.go:42: 07:49:11 | collector-otlp-production-grpc | Creating namespace: kuttl-test-dear-treefrog logger.go:42: 07:49:11 | collector-otlp-production-grpc/1-install | starting test step 1-install logger.go:42: 07:49:11 | collector-otlp-production-grpc/1-install | Jaeger:kuttl-test-dear-treefrog/my-jaeger created logger.go:42: 07:50:02 | collector-otlp-production-grpc/1-install | test step completed 1-install logger.go:42: 07:50:02 | collector-otlp-production-grpc/2-smoke-test | starting test step 2-smoke-test logger.go:42: 07:50:02 | collector-otlp-production-grpc/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:50:04 | collector-otlp-production-grpc/2-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:50:11 | collector-otlp-production-grpc/2-smoke-test | running command: [sh -c REPORTING_PROTOCOL=grpc ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/otlp-smoke-test.yaml.template -o otlp-smoke-test-job.yaml] logger.go:42: 07:50:11 | collector-otlp-production-grpc/2-smoke-test | running command: [sh -c kubectl create -f otlp-smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:50:11 | collector-otlp-production-grpc/2-smoke-test | job.batch/report-span created logger.go:42: 07:50:11 | collector-otlp-production-grpc/2-smoke-test | job.batch/check-span created logger.go:42: 07:50:32 | collector-otlp-production-grpc/2-smoke-test | test step completed 2-smoke-test logger.go:42: 07:50:32 | collector-otlp-production-grpc | collector-otlp-production-grpc events from ns kuttl-test-dear-treefrog: logger.go:42: 07:50:32 | collector-otlp-production-grpc | 2025-02-03 07:49:32 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestdeartreefrogmyjaeger-1-65468cd75j266 Binding Scheduled Successfully assigned kuttl-test-dear-treefrog/elasticsearch-cdm-kuttltestdeartreefrogmyjaeger-1-65468cd75j266 to ip-10-0-66-98.us-east-2.compute.internal default-scheduler logger.go:42: 07:50:32 | collector-otlp-production-grpc | 2025-02-03 07:49:32 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestdeartreefrogmyjaeger-1-65468cd7fc SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestdeartreefrogmyjaeger-1-65468cd75j266 replicaset-controller logger.go:42: 07:50:32 | collector-otlp-production-grpc | 2025-02-03 07:49:32 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestdeartreefrogmyjaeger-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestdeartreefrogmyjaeger-1-65468cd7fc to 1 deployment-controller logger.go:42: 07:50:32 | collector-otlp-production-grpc | 2025-02-03 07:49:33 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestdeartreefrogmyjaeger-1-65468cd75j266 AddedInterface Add eth0 [10.129.2.59/23] from ovn-kubernetes multus logger.go:42: 07:50:32 | collector-otlp-production-grpc | 2025-02-03 07:49:33 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestdeartreefrogmyjaeger-1-65468cd75j266.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:410674bf2f77ff1d431f494cadfab3e81249b35695eb65aedfdb8498079b4d35" already present on machine kubelet logger.go:42: 07:50:32 | collector-otlp-production-grpc | 2025-02-03 07:49:33 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestdeartreefrogmyjaeger-1-65468cd75j266.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:50:32 | collector-otlp-production-grpc | 2025-02-03 07:49:33 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestdeartreefrogmyjaeger-1-65468cd75j266.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:50:32 | collector-otlp-production-grpc | 2025-02-03 07:49:33 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestdeartreefrogmyjaeger-1-65468cd75j266.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:fe2439e5f9e14d828caace425214edfa3edc179d72d225c43f69e9e4c9a5837b" already present on machine kubelet logger.go:42: 07:50:32 | collector-otlp-production-grpc | 2025-02-03 07:49:33 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestdeartreefrogmyjaeger-1-65468cd75j266.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:50:32 | collector-otlp-production-grpc | 2025-02-03 07:49:33 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestdeartreefrogmyjaeger-1-65468cd75j266.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:50:32 | collector-otlp-production-grpc | 2025-02-03 07:49:43 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestdeartreefrogmyjaeger-1-65468cd75j266.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 07:50:32 | collector-otlp-production-grpc | 2025-02-03 07:49:48 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestdeartreefrogmyjaeger-1-65468cd75j266.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:50:32 | collector-otlp-production-grpc | 2025-02-03 07:49:59 +0000 UTC Normal Pod my-jaeger-collector-99ffcdd54-trk5d Binding Scheduled Successfully assigned kuttl-test-dear-treefrog/my-jaeger-collector-99ffcdd54-trk5d to ip-10-0-6-172.us-east-2.compute.internal default-scheduler logger.go:42: 07:50:32 | collector-otlp-production-grpc | 2025-02-03 07:49:59 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-99ffcdd54 SuccessfulCreate Created pod: my-jaeger-collector-99ffcdd54-trk5d replicaset-controller logger.go:42: 07:50:32 | collector-otlp-production-grpc | 2025-02-03 07:49:59 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-99ffcdd54 to 1 deployment-controller logger.go:42: 07:50:32 | collector-otlp-production-grpc | 2025-02-03 07:49:59 +0000 UTC Normal Pod my-jaeger-query-546ffb4ddd-r7n2p Binding Scheduled Successfully assigned kuttl-test-dear-treefrog/my-jaeger-query-546ffb4ddd-r7n2p to ip-10-0-26-111.us-east-2.compute.internal default-scheduler logger.go:42: 07:50:32 | collector-otlp-production-grpc | 2025-02-03 07:49:59 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-546ffb4ddd SuccessfulCreate Created pod: my-jaeger-query-546ffb4ddd-r7n2p replicaset-controller logger.go:42: 07:50:32 | collector-otlp-production-grpc | 2025-02-03 07:49:59 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-546ffb4ddd to 1 deployment-controller logger.go:42: 07:50:32 | collector-otlp-production-grpc | 2025-02-03 07:50:00 +0000 UTC Normal Pod my-jaeger-collector-99ffcdd54-trk5d AddedInterface Add eth0 [10.131.0.47/23] from ovn-kubernetes multus logger.go:42: 07:50:32 | collector-otlp-production-grpc | 2025-02-03 07:50:00 +0000 UTC Normal Pod my-jaeger-collector-99ffcdd54-trk5d.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:40a1c7fa23aa5ffa64e8e22aa38022f5d4d7ff644c46a3da7169b713d486c3c1" already present on machine kubelet logger.go:42: 07:50:32 | collector-otlp-production-grpc | 2025-02-03 07:50:00 +0000 UTC Normal Pod my-jaeger-collector-99ffcdd54-trk5d.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:50:32 | collector-otlp-production-grpc | 2025-02-03 07:50:00 +0000 UTC Normal Pod my-jaeger-collector-99ffcdd54-trk5d.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:50:32 | collector-otlp-production-grpc | 2025-02-03 07:50:00 +0000 UTC Normal Pod my-jaeger-query-546ffb4ddd-r7n2p AddedInterface Add eth0 [10.128.2.65/23] from ovn-kubernetes multus logger.go:42: 07:50:32 | collector-otlp-production-grpc | 2025-02-03 07:50:00 +0000 UTC Normal Pod my-jaeger-query-546ffb4ddd-r7n2p.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:84020ce29bcb5a4bd018e6596188ae919c5cd600e08f78a546c0e76ea477685e" already present on machine kubelet logger.go:42: 07:50:32 | collector-otlp-production-grpc | 2025-02-03 07:50:00 +0000 UTC Normal Pod my-jaeger-query-546ffb4ddd-r7n2p.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:50:32 | collector-otlp-production-grpc | 2025-02-03 07:50:00 +0000 UTC Normal Pod my-jaeger-query-546ffb4ddd-r7n2p.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:50:32 | collector-otlp-production-grpc | 2025-02-03 07:50:00 +0000 UTC Normal Pod my-jaeger-query-546ffb4ddd-r7n2p.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" already present on machine kubelet logger.go:42: 07:50:32 | collector-otlp-production-grpc | 2025-02-03 07:50:00 +0000 UTC Normal Pod my-jaeger-query-546ffb4ddd-r7n2p.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:50:32 | collector-otlp-production-grpc | 2025-02-03 07:50:00 +0000 UTC Normal Pod my-jaeger-query-546ffb4ddd-r7n2p.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:50:32 | collector-otlp-production-grpc | 2025-02-03 07:50:00 +0000 UTC Normal Pod my-jaeger-query-546ffb4ddd-r7n2p.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:65342c7d622f336741a8ae0b9dead79c7ecabc155d02cdd42b7b49ca36680e74" already present on machine kubelet logger.go:42: 07:50:32 | collector-otlp-production-grpc | 2025-02-03 07:50:00 +0000 UTC Normal Pod my-jaeger-query-546ffb4ddd-r7n2p.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:50:32 | collector-otlp-production-grpc | 2025-02-03 07:50:00 +0000 UTC Normal Pod my-jaeger-query-546ffb4ddd-r7n2p.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:50:32 | collector-otlp-production-grpc | 2025-02-03 07:50:07 +0000 UTC Normal Pod my-jaeger-query-546ffb4ddd-r7n2p.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:50:32 | collector-otlp-production-grpc | 2025-02-03 07:50:07 +0000 UTC Normal Pod my-jaeger-query-546ffb4ddd-r7n2p.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:50:32 | collector-otlp-production-grpc | 2025-02-03 07:50:07 +0000 UTC Normal Pod my-jaeger-query-546ffb4ddd-r7n2p.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:50:32 | collector-otlp-production-grpc | 2025-02-03 07:50:07 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-546ffb4ddd SuccessfulDelete Deleted pod: my-jaeger-query-546ffb4ddd-r7n2p replicaset-controller logger.go:42: 07:50:32 | collector-otlp-production-grpc | 2025-02-03 07:50:07 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled down replica set my-jaeger-query-546ffb4ddd to 0 from 1 deployment-controller logger.go:42: 07:50:32 | collector-otlp-production-grpc | 2025-02-03 07:50:08 +0000 UTC Normal Pod my-jaeger-query-69749fc56c-g5c7g Binding Scheduled Successfully assigned kuttl-test-dear-treefrog/my-jaeger-query-69749fc56c-g5c7g to ip-10-0-26-111.us-east-2.compute.internal default-scheduler logger.go:42: 07:50:32 | collector-otlp-production-grpc | 2025-02-03 07:50:08 +0000 UTC Normal Pod my-jaeger-query-69749fc56c-g5c7g AddedInterface Add eth0 [10.128.2.66/23] from ovn-kubernetes multus logger.go:42: 07:50:32 | collector-otlp-production-grpc | 2025-02-03 07:50:08 +0000 UTC Normal Pod my-jaeger-query-69749fc56c-g5c7g.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:84020ce29bcb5a4bd018e6596188ae919c5cd600e08f78a546c0e76ea477685e" already present on machine kubelet logger.go:42: 07:50:32 | collector-otlp-production-grpc | 2025-02-03 07:50:08 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-69749fc56c SuccessfulCreate Created pod: my-jaeger-query-69749fc56c-g5c7g replicaset-controller logger.go:42: 07:50:32 | collector-otlp-production-grpc | 2025-02-03 07:50:08 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-69749fc56c to 1 deployment-controller logger.go:42: 07:50:32 | collector-otlp-production-grpc | 2025-02-03 07:50:09 +0000 UTC Normal Pod my-jaeger-query-69749fc56c-g5c7g.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:50:32 | collector-otlp-production-grpc | 2025-02-03 07:50:09 +0000 UTC Normal Pod my-jaeger-query-69749fc56c-g5c7g.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:50:32 | collector-otlp-production-grpc | 2025-02-03 07:50:09 +0000 UTC Normal Pod my-jaeger-query-69749fc56c-g5c7g.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" already present on machine kubelet logger.go:42: 07:50:32 | collector-otlp-production-grpc | 2025-02-03 07:50:09 +0000 UTC Normal Pod my-jaeger-query-69749fc56c-g5c7g.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:50:32 | collector-otlp-production-grpc | 2025-02-03 07:50:09 +0000 UTC Normal Pod my-jaeger-query-69749fc56c-g5c7g.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:50:32 | collector-otlp-production-grpc | 2025-02-03 07:50:09 +0000 UTC Normal Pod my-jaeger-query-69749fc56c-g5c7g.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:65342c7d622f336741a8ae0b9dead79c7ecabc155d02cdd42b7b49ca36680e74" already present on machine kubelet logger.go:42: 07:50:32 | collector-otlp-production-grpc | 2025-02-03 07:50:09 +0000 UTC Normal Pod my-jaeger-query-69749fc56c-g5c7g.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:50:32 | collector-otlp-production-grpc | 2025-02-03 07:50:09 +0000 UTC Normal Pod my-jaeger-query-69749fc56c-g5c7g.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:50:32 | collector-otlp-production-grpc | 2025-02-03 07:50:11 +0000 UTC Normal Pod check-span-hwz9b Binding Scheduled Successfully assigned kuttl-test-dear-treefrog/check-span-hwz9b to ip-10-0-6-172.us-east-2.compute.internal default-scheduler logger.go:42: 07:50:32 | collector-otlp-production-grpc | 2025-02-03 07:50:11 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-hwz9b job-controller logger.go:42: 07:50:32 | collector-otlp-production-grpc | 2025-02-03 07:50:11 +0000 UTC Normal Pod report-span-q5sbp Binding Scheduled Successfully assigned kuttl-test-dear-treefrog/report-span-q5sbp to ip-10-0-6-172.us-east-2.compute.internal default-scheduler logger.go:42: 07:50:32 | collector-otlp-production-grpc | 2025-02-03 07:50:11 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-q5sbp job-controller logger.go:42: 07:50:32 | collector-otlp-production-grpc | 2025-02-03 07:50:12 +0000 UTC Normal Pod check-span-hwz9b AddedInterface Add eth0 [10.131.0.49/23] from ovn-kubernetes multus logger.go:42: 07:50:32 | collector-otlp-production-grpc | 2025-02-03 07:50:12 +0000 UTC Normal Pod check-span-hwz9b.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:50:32 | collector-otlp-production-grpc | 2025-02-03 07:50:12 +0000 UTC Normal Pod report-span-q5sbp AddedInterface Add eth0 [10.131.0.48/23] from ovn-kubernetes multus logger.go:42: 07:50:32 | collector-otlp-production-grpc | 2025-02-03 07:50:12 +0000 UTC Normal Pod report-span-q5sbp.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:50:32 | collector-otlp-production-grpc | 2025-02-03 07:50:13 +0000 UTC Normal Pod check-span-hwz9b.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 589ms (589ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:50:32 | collector-otlp-production-grpc | 2025-02-03 07:50:13 +0000 UTC Normal Pod check-span-hwz9b.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:50:32 | collector-otlp-production-grpc | 2025-02-03 07:50:13 +0000 UTC Normal Pod check-span-hwz9b.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:50:32 | collector-otlp-production-grpc | 2025-02-03 07:50:13 +0000 UTC Normal Pod report-span-q5sbp.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 979ms (979ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:50:32 | collector-otlp-production-grpc | 2025-02-03 07:50:13 +0000 UTC Normal Pod report-span-q5sbp.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:50:32 | collector-otlp-production-grpc | 2025-02-03 07:50:13 +0000 UTC Normal Pod report-span-q5sbp.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:50:32 | collector-otlp-production-grpc | 2025-02-03 07:50:14 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:50:32 | collector-otlp-production-grpc | 2025-02-03 07:50:14 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:50:32 | collector-otlp-production-grpc | 2025-02-03 07:50:14 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:50:32 | collector-otlp-production-grpc | 2025-02-03 07:50:29 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 07:50:32 | collector-otlp-production-grpc | 2025-02-03 07:50:29 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get memory utilization: missing request for memory in container jaeger-collector of Pod my-jaeger-collector-99ffcdd54-trk5d horizontal-pod-autoscaler logger.go:42: 07:50:32 | collector-otlp-production-grpc | 2025-02-03 07:50:29 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 07:50:32 | collector-otlp-production-grpc | 2025-02-03 07:50:31 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:50:32 | collector-otlp-production-grpc | Deleting namespace: kuttl-test-dear-treefrog === CONT kuttl/harness/artifacts logger.go:42: 07:50:45 | artifacts | Creating namespace: kuttl-test-wealthy-bison logger.go:42: 07:50:45 | artifacts | artifacts events from ns kuttl-test-wealthy-bison: logger.go:42: 07:50:45 | artifacts | Deleting namespace: kuttl-test-wealthy-bison === CONT kuttl/harness/collector-otlp-allinone-http logger.go:42: 07:50:51 | collector-otlp-allinone-http | Creating namespace: kuttl-test-fit-minnow logger.go:42: 07:50:51 | collector-otlp-allinone-http/0-install | starting test step 0-install logger.go:42: 07:50:51 | collector-otlp-allinone-http/0-install | Jaeger:kuttl-test-fit-minnow/my-jaeger created logger.go:42: 07:50:58 | collector-otlp-allinone-http/0-install | test step completed 0-install logger.go:42: 07:50:58 | collector-otlp-allinone-http/1-smoke-test | starting test step 1-smoke-test logger.go:42: 07:50:58 | collector-otlp-allinone-http/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:51:00 | collector-otlp-allinone-http/1-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:51:06 | collector-otlp-allinone-http/1-smoke-test | running command: [sh -c REPORTING_PROTOCOL=http ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/otlp-smoke-test.yaml.template -o otlp-smoke-test-job.yaml] logger.go:42: 07:51:06 | collector-otlp-allinone-http/1-smoke-test | running command: [sh -c kubectl create -f otlp-smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:51:07 | collector-otlp-allinone-http/1-smoke-test | job.batch/report-span created logger.go:42: 07:51:07 | collector-otlp-allinone-http/1-smoke-test | job.batch/check-span created logger.go:42: 07:51:20 | collector-otlp-allinone-http/1-smoke-test | test step completed 1-smoke-test logger.go:42: 07:51:20 | collector-otlp-allinone-http | collector-otlp-allinone-http events from ns kuttl-test-fit-minnow: logger.go:42: 07:51:20 | collector-otlp-allinone-http | 2025-02-03 07:50:55 +0000 UTC Normal Pod my-jaeger-657755c55-sz9ld Binding Scheduled Successfully assigned kuttl-test-fit-minnow/my-jaeger-657755c55-sz9ld to ip-10-0-66-98.us-east-2.compute.internal default-scheduler logger.go:42: 07:51:20 | collector-otlp-allinone-http | 2025-02-03 07:50:55 +0000 UTC Warning Pod my-jaeger-657755c55-sz9ld FailedMount MountVolume.SetUp failed for volume "my-jaeger-ui-oauth-proxy-tls" : secret "my-jaeger-ui-oauth-proxy-tls" not found kubelet logger.go:42: 07:51:20 | collector-otlp-allinone-http | 2025-02-03 07:50:55 +0000 UTC Normal ReplicaSet.apps my-jaeger-657755c55 SuccessfulCreate Created pod: my-jaeger-657755c55-sz9ld replicaset-controller logger.go:42: 07:51:20 | collector-otlp-allinone-http | 2025-02-03 07:50:55 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-657755c55 to 1 deployment-controller logger.go:42: 07:51:20 | collector-otlp-allinone-http | 2025-02-03 07:50:57 +0000 UTC Normal Pod my-jaeger-657755c55-sz9ld AddedInterface Add eth0 [10.129.2.60/23] from ovn-kubernetes multus logger.go:42: 07:51:20 | collector-otlp-allinone-http | 2025-02-03 07:50:57 +0000 UTC Normal Pod my-jaeger-657755c55-sz9ld.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:fa9aaf9ae026228265ce7f10fb451d5c52a4e0269ce1fa9024cbefddf765ae8d" already present on machine kubelet logger.go:42: 07:51:20 | collector-otlp-allinone-http | 2025-02-03 07:50:57 +0000 UTC Normal Pod my-jaeger-657755c55-sz9ld.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:51:20 | collector-otlp-allinone-http | 2025-02-03 07:50:57 +0000 UTC Normal Pod my-jaeger-657755c55-sz9ld.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:51:20 | collector-otlp-allinone-http | 2025-02-03 07:50:57 +0000 UTC Normal Pod my-jaeger-657755c55-sz9ld.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" already present on machine kubelet logger.go:42: 07:51:20 | collector-otlp-allinone-http | 2025-02-03 07:50:57 +0000 UTC Normal Pod my-jaeger-657755c55-sz9ld.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:51:20 | collector-otlp-allinone-http | 2025-02-03 07:50:57 +0000 UTC Normal Pod my-jaeger-657755c55-sz9ld.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:51:20 | collector-otlp-allinone-http | 2025-02-03 07:51:01 +0000 UTC Normal Pod my-jaeger-657755c55-sz9ld.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:51:20 | collector-otlp-allinone-http | 2025-02-03 07:51:01 +0000 UTC Normal Pod my-jaeger-657755c55-sz9ld.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:51:20 | collector-otlp-allinone-http | 2025-02-03 07:51:01 +0000 UTC Normal ReplicaSet.apps my-jaeger-657755c55 SuccessfulDelete Deleted pod: my-jaeger-657755c55-sz9ld replicaset-controller logger.go:42: 07:51:20 | collector-otlp-allinone-http | 2025-02-03 07:51:01 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled down replica set my-jaeger-657755c55 to 0 from 1 deployment-controller logger.go:42: 07:51:20 | collector-otlp-allinone-http | 2025-02-03 07:51:02 +0000 UTC Normal Pod my-jaeger-7457496958-m7897 Binding Scheduled Successfully assigned kuttl-test-fit-minnow/my-jaeger-7457496958-m7897 to ip-10-0-66-98.us-east-2.compute.internal default-scheduler logger.go:42: 07:51:20 | collector-otlp-allinone-http | 2025-02-03 07:51:02 +0000 UTC Normal ReplicaSet.apps my-jaeger-7457496958 SuccessfulCreate Created pod: my-jaeger-7457496958-m7897 replicaset-controller logger.go:42: 07:51:20 | collector-otlp-allinone-http | 2025-02-03 07:51:02 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-7457496958 to 1 deployment-controller logger.go:42: 07:51:20 | collector-otlp-allinone-http | 2025-02-03 07:51:03 +0000 UTC Normal Pod my-jaeger-7457496958-m7897 AddedInterface Add eth0 [10.129.2.61/23] from ovn-kubernetes multus logger.go:42: 07:51:20 | collector-otlp-allinone-http | 2025-02-03 07:51:03 +0000 UTC Normal Pod my-jaeger-7457496958-m7897.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:fa9aaf9ae026228265ce7f10fb451d5c52a4e0269ce1fa9024cbefddf765ae8d" already present on machine kubelet logger.go:42: 07:51:20 | collector-otlp-allinone-http | 2025-02-03 07:51:03 +0000 UTC Normal Pod my-jaeger-7457496958-m7897.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:51:20 | collector-otlp-allinone-http | 2025-02-03 07:51:03 +0000 UTC Normal Pod my-jaeger-7457496958-m7897.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:51:20 | collector-otlp-allinone-http | 2025-02-03 07:51:03 +0000 UTC Normal Pod my-jaeger-7457496958-m7897.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" already present on machine kubelet logger.go:42: 07:51:20 | collector-otlp-allinone-http | 2025-02-03 07:51:03 +0000 UTC Normal Pod my-jaeger-7457496958-m7897.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:51:20 | collector-otlp-allinone-http | 2025-02-03 07:51:03 +0000 UTC Normal Pod my-jaeger-7457496958-m7897.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:51:20 | collector-otlp-allinone-http | 2025-02-03 07:51:07 +0000 UTC Normal Pod check-span-f7nxz Binding Scheduled Successfully assigned kuttl-test-fit-minnow/check-span-f7nxz to ip-10-0-26-111.us-east-2.compute.internal default-scheduler logger.go:42: 07:51:20 | collector-otlp-allinone-http | 2025-02-03 07:51:07 +0000 UTC Normal Pod check-span-f7nxz AddedInterface Add eth0 [10.128.2.67/23] from ovn-kubernetes multus logger.go:42: 07:51:20 | collector-otlp-allinone-http | 2025-02-03 07:51:07 +0000 UTC Normal Pod check-span-f7nxz.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:51:20 | collector-otlp-allinone-http | 2025-02-03 07:51:07 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-f7nxz job-controller logger.go:42: 07:51:20 | collector-otlp-allinone-http | 2025-02-03 07:51:07 +0000 UTC Normal Pod report-span-vp5vc Binding Scheduled Successfully assigned kuttl-test-fit-minnow/report-span-vp5vc to ip-10-0-6-172.us-east-2.compute.internal default-scheduler logger.go:42: 07:51:20 | collector-otlp-allinone-http | 2025-02-03 07:51:07 +0000 UTC Normal Pod report-span-vp5vc AddedInterface Add eth0 [10.131.0.50/23] from ovn-kubernetes multus logger.go:42: 07:51:20 | collector-otlp-allinone-http | 2025-02-03 07:51:07 +0000 UTC Normal Pod report-span-vp5vc.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:51:20 | collector-otlp-allinone-http | 2025-02-03 07:51:07 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-vp5vc job-controller logger.go:42: 07:51:20 | collector-otlp-allinone-http | 2025-02-03 07:51:08 +0000 UTC Normal Pod check-span-f7nxz.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 538ms (538ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:51:20 | collector-otlp-allinone-http | 2025-02-03 07:51:08 +0000 UTC Normal Pod check-span-f7nxz.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:51:20 | collector-otlp-allinone-http | 2025-02-03 07:51:08 +0000 UTC Normal Pod check-span-f7nxz.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:51:20 | collector-otlp-allinone-http | 2025-02-03 07:51:08 +0000 UTC Normal Pod report-span-vp5vc.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 910ms (910ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:51:20 | collector-otlp-allinone-http | 2025-02-03 07:51:08 +0000 UTC Normal Pod report-span-vp5vc.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:51:20 | collector-otlp-allinone-http | 2025-02-03 07:51:08 +0000 UTC Normal Pod report-span-vp5vc.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:51:20 | collector-otlp-allinone-http | 2025-02-03 07:51:19 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:51:20 | collector-otlp-allinone-http | Deleting namespace: kuttl-test-fit-minnow === CONT kuttl/harness/collector-otlp-allinone-grpc logger.go:42: 07:51:33 | collector-otlp-allinone-grpc | Creating namespace: kuttl-test-ready-spider logger.go:42: 07:51:33 | collector-otlp-allinone-grpc/0-install | starting test step 0-install logger.go:42: 07:51:33 | collector-otlp-allinone-grpc/0-install | Jaeger:kuttl-test-ready-spider/my-jaeger created logger.go:42: 07:51:40 | collector-otlp-allinone-grpc/0-install | test step completed 0-install logger.go:42: 07:51:40 | collector-otlp-allinone-grpc/1-smoke-test | starting test step 1-smoke-test logger.go:42: 07:51:40 | collector-otlp-allinone-grpc/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:51:41 | collector-otlp-allinone-grpc/1-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:51:48 | collector-otlp-allinone-grpc/1-smoke-test | running command: [sh -c REPORTING_PROTOCOL=grpc ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/otlp-smoke-test.yaml.template -o otlp-smoke-test-job.yaml] logger.go:42: 07:51:49 | collector-otlp-allinone-grpc/1-smoke-test | running command: [sh -c kubectl create -f otlp-smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:51:49 | collector-otlp-allinone-grpc/1-smoke-test | job.batch/report-span created logger.go:42: 07:51:49 | collector-otlp-allinone-grpc/1-smoke-test | job.batch/check-span created logger.go:42: 07:52:10 | collector-otlp-allinone-grpc/1-smoke-test | test step completed 1-smoke-test logger.go:42: 07:52:10 | collector-otlp-allinone-grpc | collector-otlp-allinone-grpc events from ns kuttl-test-ready-spider: logger.go:42: 07:52:10 | collector-otlp-allinone-grpc | 2025-02-03 07:51:37 +0000 UTC Normal Pod my-jaeger-69d85cf744-4crnk Binding Scheduled Successfully assigned kuttl-test-ready-spider/my-jaeger-69d85cf744-4crnk to ip-10-0-66-98.us-east-2.compute.internal default-scheduler logger.go:42: 07:52:10 | collector-otlp-allinone-grpc | 2025-02-03 07:51:37 +0000 UTC Warning Pod my-jaeger-69d85cf744-4crnk FailedMount MountVolume.SetUp failed for volume "my-jaeger-ui-oauth-proxy-tls" : secret "my-jaeger-ui-oauth-proxy-tls" not found kubelet logger.go:42: 07:52:10 | collector-otlp-allinone-grpc | 2025-02-03 07:51:37 +0000 UTC Warning Pod my-jaeger-69d85cf744-4crnk FailedMount MountVolume.SetUp failed for volume "my-jaeger-collector-tls-config-volume" : secret "my-jaeger-collector-headless-tls" not found kubelet logger.go:42: 07:52:10 | collector-otlp-allinone-grpc | 2025-02-03 07:51:37 +0000 UTC Normal ReplicaSet.apps my-jaeger-69d85cf744 SuccessfulCreate Created pod: my-jaeger-69d85cf744-4crnk replicaset-controller logger.go:42: 07:52:10 | collector-otlp-allinone-grpc | 2025-02-03 07:51:37 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-69d85cf744 to 1 deployment-controller logger.go:42: 07:52:10 | collector-otlp-allinone-grpc | 2025-02-03 07:51:38 +0000 UTC Normal Pod my-jaeger-69d85cf744-4crnk AddedInterface Add eth0 [10.129.2.62/23] from ovn-kubernetes multus logger.go:42: 07:52:10 | collector-otlp-allinone-grpc | 2025-02-03 07:51:38 +0000 UTC Normal Pod my-jaeger-69d85cf744-4crnk.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:fa9aaf9ae026228265ce7f10fb451d5c52a4e0269ce1fa9024cbefddf765ae8d" already present on machine kubelet logger.go:42: 07:52:10 | collector-otlp-allinone-grpc | 2025-02-03 07:51:38 +0000 UTC Normal Pod my-jaeger-69d85cf744-4crnk.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:52:10 | collector-otlp-allinone-grpc | 2025-02-03 07:51:38 +0000 UTC Normal Pod my-jaeger-69d85cf744-4crnk.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:52:10 | collector-otlp-allinone-grpc | 2025-02-03 07:51:38 +0000 UTC Normal Pod my-jaeger-69d85cf744-4crnk.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" already present on machine kubelet logger.go:42: 07:52:10 | collector-otlp-allinone-grpc | 2025-02-03 07:51:38 +0000 UTC Normal Pod my-jaeger-69d85cf744-4crnk.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:52:10 | collector-otlp-allinone-grpc | 2025-02-03 07:51:38 +0000 UTC Normal Pod my-jaeger-69d85cf744-4crnk.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:52:10 | collector-otlp-allinone-grpc | 2025-02-03 07:51:45 +0000 UTC Normal Pod my-jaeger-69d85cf744-4crnk.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:52:10 | collector-otlp-allinone-grpc | 2025-02-03 07:51:45 +0000 UTC Normal Pod my-jaeger-69d85cf744-4crnk.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:52:10 | collector-otlp-allinone-grpc | 2025-02-03 07:51:45 +0000 UTC Normal ReplicaSet.apps my-jaeger-69d85cf744 SuccessfulDelete Deleted pod: my-jaeger-69d85cf744-4crnk replicaset-controller logger.go:42: 07:52:10 | collector-otlp-allinone-grpc | 2025-02-03 07:51:45 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled down replica set my-jaeger-69d85cf744 to 0 from 1 deployment-controller logger.go:42: 07:52:10 | collector-otlp-allinone-grpc | 2025-02-03 07:51:46 +0000 UTC Normal Pod my-jaeger-6fcfbd8468-4wbg7 Binding Scheduled Successfully assigned kuttl-test-ready-spider/my-jaeger-6fcfbd8468-4wbg7 to ip-10-0-66-98.us-east-2.compute.internal default-scheduler logger.go:42: 07:52:10 | collector-otlp-allinone-grpc | 2025-02-03 07:51:46 +0000 UTC Normal ReplicaSet.apps my-jaeger-6fcfbd8468 SuccessfulCreate Created pod: my-jaeger-6fcfbd8468-4wbg7 replicaset-controller logger.go:42: 07:52:10 | collector-otlp-allinone-grpc | 2025-02-03 07:51:46 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-6fcfbd8468 to 1 deployment-controller logger.go:42: 07:52:10 | collector-otlp-allinone-grpc | 2025-02-03 07:51:47 +0000 UTC Normal Pod my-jaeger-6fcfbd8468-4wbg7 AddedInterface Add eth0 [10.129.2.63/23] from ovn-kubernetes multus logger.go:42: 07:52:10 | collector-otlp-allinone-grpc | 2025-02-03 07:51:47 +0000 UTC Normal Pod my-jaeger-6fcfbd8468-4wbg7.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:fa9aaf9ae026228265ce7f10fb451d5c52a4e0269ce1fa9024cbefddf765ae8d" already present on machine kubelet logger.go:42: 07:52:10 | collector-otlp-allinone-grpc | 2025-02-03 07:51:47 +0000 UTC Normal Pod my-jaeger-6fcfbd8468-4wbg7.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:52:10 | collector-otlp-allinone-grpc | 2025-02-03 07:51:47 +0000 UTC Normal Pod my-jaeger-6fcfbd8468-4wbg7.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:52:10 | collector-otlp-allinone-grpc | 2025-02-03 07:51:47 +0000 UTC Normal Pod my-jaeger-6fcfbd8468-4wbg7.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" already present on machine kubelet logger.go:42: 07:52:10 | collector-otlp-allinone-grpc | 2025-02-03 07:51:47 +0000 UTC Normal Pod my-jaeger-6fcfbd8468-4wbg7.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:52:10 | collector-otlp-allinone-grpc | 2025-02-03 07:51:47 +0000 UTC Normal Pod my-jaeger-6fcfbd8468-4wbg7.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:52:10 | collector-otlp-allinone-grpc | 2025-02-03 07:51:49 +0000 UTC Normal Pod check-span-67llm Binding Scheduled Successfully assigned kuttl-test-ready-spider/check-span-67llm to ip-10-0-26-111.us-east-2.compute.internal default-scheduler logger.go:42: 07:52:10 | collector-otlp-allinone-grpc | 2025-02-03 07:51:49 +0000 UTC Normal Pod check-span-67llm AddedInterface Add eth0 [10.128.2.68/23] from ovn-kubernetes multus logger.go:42: 07:52:10 | collector-otlp-allinone-grpc | 2025-02-03 07:51:49 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-67llm job-controller logger.go:42: 07:52:10 | collector-otlp-allinone-grpc | 2025-02-03 07:51:49 +0000 UTC Normal Pod report-span-mg2fs Binding Scheduled Successfully assigned kuttl-test-ready-spider/report-span-mg2fs to ip-10-0-6-172.us-east-2.compute.internal default-scheduler logger.go:42: 07:52:10 | collector-otlp-allinone-grpc | 2025-02-03 07:51:49 +0000 UTC Normal Pod report-span-mg2fs AddedInterface Add eth0 [10.131.0.51/23] from ovn-kubernetes multus logger.go:42: 07:52:10 | collector-otlp-allinone-grpc | 2025-02-03 07:51:49 +0000 UTC Normal Pod report-span-mg2fs.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:52:10 | collector-otlp-allinone-grpc | 2025-02-03 07:51:49 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-mg2fs job-controller logger.go:42: 07:52:10 | collector-otlp-allinone-grpc | 2025-02-03 07:51:50 +0000 UTC Normal Pod check-span-67llm.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:52:10 | collector-otlp-allinone-grpc | 2025-02-03 07:51:50 +0000 UTC Normal Pod check-span-67llm.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 536ms (536ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:52:10 | collector-otlp-allinone-grpc | 2025-02-03 07:51:50 +0000 UTC Normal Pod check-span-67llm.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:52:10 | collector-otlp-allinone-grpc | 2025-02-03 07:51:50 +0000 UTC Normal Pod check-span-67llm.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:52:10 | collector-otlp-allinone-grpc | 2025-02-03 07:51:50 +0000 UTC Normal Pod report-span-mg2fs.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 564ms (564ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:52:10 | collector-otlp-allinone-grpc | 2025-02-03 07:51:50 +0000 UTC Normal Pod report-span-mg2fs.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:52:10 | collector-otlp-allinone-grpc | 2025-02-03 07:51:50 +0000 UTC Normal Pod report-span-mg2fs.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:52:10 | collector-otlp-allinone-grpc | 2025-02-03 07:52:09 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:52:10 | collector-otlp-allinone-grpc | Deleting namespace: kuttl-test-ready-spider === CONT kuttl/harness/collector-autoscale logger.go:42: 07:52:23 | collector-autoscale | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:52:23 | collector-autoscale | Creating namespace: kuttl-test-climbing-minnow logger.go:42: 07:52:23 | collector-autoscale/1-install | starting test step 1-install logger.go:42: 07:52:23 | collector-autoscale/1-install | Jaeger:kuttl-test-climbing-minnow/simple-prod created logger.go:42: 07:53:18 | collector-autoscale/1-install | test step completed 1-install logger.go:42: 07:53:18 | collector-autoscale/2- | starting test step 2- logger.go:42: 07:53:18 | collector-autoscale/2- | test step completed 2- logger.go:42: 07:53:18 | collector-autoscale | collector-autoscale events from ns kuttl-test-climbing-minnow: logger.go:42: 07:53:18 | collector-autoscale | 2025-02-03 07:52:48 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestclimbingminnowsimpleprod-1-6c7c56dd69 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestclimbingminnowsimpleprod-1-6c7cwl4xf replicaset-controller logger.go:42: 07:53:18 | collector-autoscale | 2025-02-03 07:52:48 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestclimbingminnowsimpleprod-1-6c7cwl4xf Binding Scheduled Successfully assigned kuttl-test-climbing-minnow/elasticsearch-cdm-kuttltestclimbingminnowsimpleprod-1-6c7cwl4xf to ip-10-0-66-98.us-east-2.compute.internal default-scheduler logger.go:42: 07:53:18 | collector-autoscale | 2025-02-03 07:52:48 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestclimbingminnowsimpleprod-1-6c7cwl4xf AddedInterface Add eth0 [10.129.2.64/23] from ovn-kubernetes multus logger.go:42: 07:53:18 | collector-autoscale | 2025-02-03 07:52:48 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestclimbingminnowsimpleprod-1-6c7cwl4xf.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:410674bf2f77ff1d431f494cadfab3e81249b35695eb65aedfdb8498079b4d35" already present on machine kubelet logger.go:42: 07:53:18 | collector-autoscale | 2025-02-03 07:52:48 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestclimbingminnowsimpleprod-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestclimbingminnowsimpleprod-1-6c7c56dd69 to 1 deployment-controller logger.go:42: 07:53:18 | collector-autoscale | 2025-02-03 07:52:49 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestclimbingminnowsimpleprod-1-6c7cwl4xf.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:53:18 | collector-autoscale | 2025-02-03 07:52:49 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestclimbingminnowsimpleprod-1-6c7cwl4xf.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:53:18 | collector-autoscale | 2025-02-03 07:52:49 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestclimbingminnowsimpleprod-1-6c7cwl4xf.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:fe2439e5f9e14d828caace425214edfa3edc179d72d225c43f69e9e4c9a5837b" already present on machine kubelet logger.go:42: 07:53:18 | collector-autoscale | 2025-02-03 07:52:49 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestclimbingminnowsimpleprod-1-6c7cwl4xf.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:53:18 | collector-autoscale | 2025-02-03 07:52:49 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestclimbingminnowsimpleprod-1-6c7cwl4xf.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:53:18 | collector-autoscale | 2025-02-03 07:53:03 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestclimbingminnowsimpleprod-1-6c7cwl4xf.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:53:18 | collector-autoscale | 2025-02-03 07:53:15 +0000 UTC Normal Pod simple-prod-collector-6c8cd458b5-lkq7w Binding Scheduled Successfully assigned kuttl-test-climbing-minnow/simple-prod-collector-6c8cd458b5-lkq7w to ip-10-0-6-172.us-east-2.compute.internal default-scheduler logger.go:42: 07:53:18 | collector-autoscale | 2025-02-03 07:53:15 +0000 UTC Normal Pod simple-prod-collector-6c8cd458b5-lkq7w AddedInterface Add eth0 [10.131.0.52/23] from ovn-kubernetes multus logger.go:42: 07:53:18 | collector-autoscale | 2025-02-03 07:53:15 +0000 UTC Normal Pod simple-prod-collector-6c8cd458b5-lkq7w.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:40a1c7fa23aa5ffa64e8e22aa38022f5d4d7ff644c46a3da7169b713d486c3c1" already present on machine kubelet logger.go:42: 07:53:18 | collector-autoscale | 2025-02-03 07:53:15 +0000 UTC Normal Pod simple-prod-collector-6c8cd458b5-lkq7w.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:53:18 | collector-autoscale | 2025-02-03 07:53:15 +0000 UTC Normal Pod simple-prod-collector-6c8cd458b5-lkq7w.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:53:18 | collector-autoscale | 2025-02-03 07:53:15 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-6c8cd458b5 SuccessfulCreate Created pod: simple-prod-collector-6c8cd458b5-lkq7w replicaset-controller logger.go:42: 07:53:18 | collector-autoscale | 2025-02-03 07:53:15 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-6c8cd458b5 to 1 deployment-controller logger.go:42: 07:53:18 | collector-autoscale | 2025-02-03 07:53:15 +0000 UTC Normal Pod simple-prod-query-6484c49449-h4kzs Binding Scheduled Successfully assigned kuttl-test-climbing-minnow/simple-prod-query-6484c49449-h4kzs to ip-10-0-26-111.us-east-2.compute.internal default-scheduler logger.go:42: 07:53:18 | collector-autoscale | 2025-02-03 07:53:15 +0000 UTC Normal Pod simple-prod-query-6484c49449-h4kzs AddedInterface Add eth0 [10.128.2.69/23] from ovn-kubernetes multus logger.go:42: 07:53:18 | collector-autoscale | 2025-02-03 07:53:15 +0000 UTC Normal Pod simple-prod-query-6484c49449-h4kzs.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:84020ce29bcb5a4bd018e6596188ae919c5cd600e08f78a546c0e76ea477685e" already present on machine kubelet logger.go:42: 07:53:18 | collector-autoscale | 2025-02-03 07:53:15 +0000 UTC Normal Pod simple-prod-query-6484c49449-h4kzs.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:53:18 | collector-autoscale | 2025-02-03 07:53:15 +0000 UTC Normal ReplicaSet.apps simple-prod-query-6484c49449 SuccessfulCreate Created pod: simple-prod-query-6484c49449-h4kzs replicaset-controller logger.go:42: 07:53:18 | collector-autoscale | 2025-02-03 07:53:15 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-6484c49449 to 1 deployment-controller logger.go:42: 07:53:18 | collector-autoscale | 2025-02-03 07:53:16 +0000 UTC Normal Pod simple-prod-query-6484c49449-h4kzs.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:53:18 | collector-autoscale | 2025-02-03 07:53:16 +0000 UTC Normal Pod simple-prod-query-6484c49449-h4kzs.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" already present on machine kubelet logger.go:42: 07:53:18 | collector-autoscale | 2025-02-03 07:53:16 +0000 UTC Normal Pod simple-prod-query-6484c49449-h4kzs.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:53:18 | collector-autoscale | 2025-02-03 07:53:16 +0000 UTC Normal Pod simple-prod-query-6484c49449-h4kzs.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:53:18 | collector-autoscale | 2025-02-03 07:53:16 +0000 UTC Normal Pod simple-prod-query-6484c49449-h4kzs.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:65342c7d622f336741a8ae0b9dead79c7ecabc155d02cdd42b7b49ca36680e74" already present on machine kubelet logger.go:42: 07:53:18 | collector-autoscale | 2025-02-03 07:53:16 +0000 UTC Normal Pod simple-prod-query-6484c49449-h4kzs.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:53:18 | collector-autoscale | 2025-02-03 07:53:16 +0000 UTC Normal Pod simple-prod-query-6484c49449-h4kzs.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:53:18 | collector-autoscale | Deleting namespace: kuttl-test-climbing-minnow === CONT kuttl/harness/cassandra-spark logger.go:42: 07:53:24 | cassandra-spark | Ignoring 01-assert.yaml.template as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:53:24 | cassandra-spark | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:53:24 | cassandra-spark | Creating namespace: kuttl-test-stunning-caribou logger.go:42: 07:53:24 | cassandra-spark | cassandra-spark events from ns kuttl-test-stunning-caribou: logger.go:42: 07:53:24 | cassandra-spark | Deleting namespace: kuttl-test-stunning-caribou === CONT kuttl/harness/set-custom-img logger.go:42: 07:53:30 | set-custom-img | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:53:30 | set-custom-img | Ignoring check-collector-img.sh as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:53:30 | set-custom-img | Creating namespace: kuttl-test-elegant-chigger logger.go:42: 07:53:30 | set-custom-img/1-install | starting test step 1-install logger.go:42: 07:53:30 | set-custom-img/1-install | Jaeger:kuttl-test-elegant-chigger/my-jaeger created logger.go:42: 07:54:28 | set-custom-img/1-install | test step completed 1-install logger.go:42: 07:54:28 | set-custom-img/2-install | starting test step 2-install logger.go:42: 07:54:28 | set-custom-img/2-install | Jaeger:kuttl-test-elegant-chigger/my-jaeger updated logger.go:42: 07:54:28 | set-custom-img/2-install | test step completed 2-install logger.go:42: 07:54:28 | set-custom-img/3-check-image | starting test step 3-check-image logger.go:42: 07:54:28 | set-custom-img/3-check-image | running command: [sh -c ./check-collector-img.sh] logger.go:42: 07:54:28 | set-custom-img/3-check-image | Collector image missmatch. Expected: test. Has: registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:40a1c7fa23aa5ffa64e8e22aa38022f5d4d7ff644c46a3da7169b713d486c3c1 logger.go:42: 07:54:34 | set-custom-img/3-check-image | Collector image asserted properly! logger.go:42: 07:54:34 | set-custom-img/3-check-image | test step completed 3-check-image logger.go:42: 07:54:34 | set-custom-img | set-custom-img events from ns kuttl-test-elegant-chigger: logger.go:42: 07:54:34 | set-custom-img | 2025-02-03 07:53:55 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestelegantchiggermyjaeger-1-c744dd74c SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestelegantchiggermyjaeger-1-c744ddxm7kx replicaset-controller logger.go:42: 07:54:34 | set-custom-img | 2025-02-03 07:53:55 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestelegantchiggermyjaeger-1-c744ddxm7kx Binding Scheduled Successfully assigned kuttl-test-elegant-chigger/elasticsearch-cdm-kuttltestelegantchiggermyjaeger-1-c744ddxm7kx to ip-10-0-66-98.us-east-2.compute.internal default-scheduler logger.go:42: 07:54:34 | set-custom-img | 2025-02-03 07:53:55 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestelegantchiggermyjaeger-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestelegantchiggermyjaeger-1-c744dd74c to 1 deployment-controller logger.go:42: 07:54:34 | set-custom-img | 2025-02-03 07:53:56 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestelegantchiggermyjaeger-1-c744ddxm7kx FailedMount MountVolume.SetUp failed for volume "certificates" : failed to sync secret cache: timed out waiting for the condition kubelet logger.go:42: 07:54:34 | set-custom-img | 2025-02-03 07:53:56 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestelegantchiggermyjaeger-1-c744ddxm7kx FailedMount MountVolume.SetUp failed for volume "elasticsearch-metrics" : failed to sync secret cache: timed out waiting for the condition kubelet logger.go:42: 07:54:34 | set-custom-img | 2025-02-03 07:53:57 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestelegantchiggermyjaeger-1-c744ddxm7kx AddedInterface Add eth0 [10.129.2.65/23] from ovn-kubernetes multus logger.go:42: 07:54:34 | set-custom-img | 2025-02-03 07:53:57 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestelegantchiggermyjaeger-1-c744ddxm7kx.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:410674bf2f77ff1d431f494cadfab3e81249b35695eb65aedfdb8498079b4d35" already present on machine kubelet logger.go:42: 07:54:34 | set-custom-img | 2025-02-03 07:53:57 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestelegantchiggermyjaeger-1-c744ddxm7kx.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:54:34 | set-custom-img | 2025-02-03 07:53:57 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestelegantchiggermyjaeger-1-c744ddxm7kx.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:54:34 | set-custom-img | 2025-02-03 07:53:58 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestelegantchiggermyjaeger-1-c744ddxm7kx.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:fe2439e5f9e14d828caace425214edfa3edc179d72d225c43f69e9e4c9a5837b" already present on machine kubelet logger.go:42: 07:54:34 | set-custom-img | 2025-02-03 07:53:58 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestelegantchiggermyjaeger-1-c744ddxm7kx.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:54:34 | set-custom-img | 2025-02-03 07:53:58 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestelegantchiggermyjaeger-1-c744ddxm7kx.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:54:34 | set-custom-img | 2025-02-03 07:54:07 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestelegantchiggermyjaeger-1-c744ddxm7kx.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 07:54:34 | set-custom-img | 2025-02-03 07:54:12 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestelegantchiggermyjaeger-1-c744ddxm7kx.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:54:34 | set-custom-img | 2025-02-03 07:54:24 +0000 UTC Normal Pod my-jaeger-collector-6c94884b8f-nn7tv Binding Scheduled Successfully assigned kuttl-test-elegant-chigger/my-jaeger-collector-6c94884b8f-nn7tv to ip-10-0-6-172.us-east-2.compute.internal default-scheduler logger.go:42: 07:54:34 | set-custom-img | 2025-02-03 07:54:24 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-6c94884b8f SuccessfulCreate Created pod: my-jaeger-collector-6c94884b8f-nn7tv replicaset-controller logger.go:42: 07:54:34 | set-custom-img | 2025-02-03 07:54:24 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-6c94884b8f to 1 deployment-controller logger.go:42: 07:54:34 | set-custom-img | 2025-02-03 07:54:24 +0000 UTC Normal Pod my-jaeger-query-cf4db8db8-rfqml Binding Scheduled Successfully assigned kuttl-test-elegant-chigger/my-jaeger-query-cf4db8db8-rfqml to ip-10-0-26-111.us-east-2.compute.internal default-scheduler logger.go:42: 07:54:34 | set-custom-img | 2025-02-03 07:54:24 +0000 UTC Warning Pod my-jaeger-query-cf4db8db8-rfqml FailedMount MountVolume.SetUp failed for volume "my-jaeger-ui-oauth-proxy-tls" : secret "my-jaeger-ui-oauth-proxy-tls" not found kubelet logger.go:42: 07:54:34 | set-custom-img | 2025-02-03 07:54:24 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-cf4db8db8 SuccessfulCreate Created pod: my-jaeger-query-cf4db8db8-rfqml replicaset-controller logger.go:42: 07:54:34 | set-custom-img | 2025-02-03 07:54:24 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-cf4db8db8 to 1 deployment-controller logger.go:42: 07:54:34 | set-custom-img | 2025-02-03 07:54:25 +0000 UTC Normal Pod my-jaeger-collector-6c94884b8f-nn7tv AddedInterface Add eth0 [10.131.0.53/23] from ovn-kubernetes multus logger.go:42: 07:54:34 | set-custom-img | 2025-02-03 07:54:25 +0000 UTC Normal Pod my-jaeger-collector-6c94884b8f-nn7tv.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:40a1c7fa23aa5ffa64e8e22aa38022f5d4d7ff644c46a3da7169b713d486c3c1" already present on machine kubelet logger.go:42: 07:54:34 | set-custom-img | 2025-02-03 07:54:25 +0000 UTC Normal Pod my-jaeger-collector-6c94884b8f-nn7tv.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:54:34 | set-custom-img | 2025-02-03 07:54:25 +0000 UTC Normal Pod my-jaeger-collector-6c94884b8f-nn7tv.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:54:34 | set-custom-img | 2025-02-03 07:54:25 +0000 UTC Normal Pod my-jaeger-query-cf4db8db8-rfqml AddedInterface Add eth0 [10.128.2.70/23] from ovn-kubernetes multus logger.go:42: 07:54:34 | set-custom-img | 2025-02-03 07:54:25 +0000 UTC Normal Pod my-jaeger-query-cf4db8db8-rfqml.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:84020ce29bcb5a4bd018e6596188ae919c5cd600e08f78a546c0e76ea477685e" already present on machine kubelet logger.go:42: 07:54:34 | set-custom-img | 2025-02-03 07:54:25 +0000 UTC Normal Pod my-jaeger-query-cf4db8db8-rfqml.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:54:34 | set-custom-img | 2025-02-03 07:54:25 +0000 UTC Normal Pod my-jaeger-query-cf4db8db8-rfqml.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:54:34 | set-custom-img | 2025-02-03 07:54:25 +0000 UTC Normal Pod my-jaeger-query-cf4db8db8-rfqml.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" already present on machine kubelet logger.go:42: 07:54:34 | set-custom-img | 2025-02-03 07:54:26 +0000 UTC Normal Pod my-jaeger-query-cf4db8db8-rfqml.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:54:34 | set-custom-img | 2025-02-03 07:54:26 +0000 UTC Normal Pod my-jaeger-query-cf4db8db8-rfqml.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:54:34 | set-custom-img | 2025-02-03 07:54:26 +0000 UTC Normal Pod my-jaeger-query-cf4db8db8-rfqml.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:65342c7d622f336741a8ae0b9dead79c7ecabc155d02cdd42b7b49ca36680e74" already present on machine kubelet logger.go:42: 07:54:34 | set-custom-img | 2025-02-03 07:54:26 +0000 UTC Normal Pod my-jaeger-query-cf4db8db8-rfqml.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:54:34 | set-custom-img | 2025-02-03 07:54:26 +0000 UTC Normal Pod my-jaeger-query-cf4db8db8-rfqml.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:54:34 | set-custom-img | 2025-02-03 07:54:32 +0000 UTC Normal Pod my-jaeger-collector-6c94884b8f-nn7tv.spec.containers{jaeger-collector} Killing Stopping container jaeger-collector kubelet logger.go:42: 07:54:34 | set-custom-img | 2025-02-03 07:54:32 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-6c94884b8f SuccessfulDelete Deleted pod: my-jaeger-collector-6c94884b8f-nn7tv replicaset-controller logger.go:42: 07:54:34 | set-custom-img | 2025-02-03 07:54:32 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled down replica set my-jaeger-collector-6c94884b8f to 0 from 1 deployment-controller logger.go:42: 07:54:34 | set-custom-img | 2025-02-03 07:54:33 +0000 UTC Normal Pod my-jaeger-collector-67c88dd854-6nvpz Binding Scheduled Successfully assigned kuttl-test-elegant-chigger/my-jaeger-collector-67c88dd854-6nvpz to ip-10-0-6-172.us-east-2.compute.internal default-scheduler logger.go:42: 07:54:34 | set-custom-img | 2025-02-03 07:54:33 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-67c88dd854 SuccessfulCreate Created pod: my-jaeger-collector-67c88dd854-6nvpz replicaset-controller logger.go:42: 07:54:34 | set-custom-img | 2025-02-03 07:54:33 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-67c88dd854 to 1 deployment-controller logger.go:42: 07:54:34 | set-custom-img | Deleting namespace: kuttl-test-elegant-chigger === CONT kuttl/harness/collector-otlp-production-http logger.go:42: 07:54:40 | collector-otlp-production-http | Creating namespace: kuttl-test-splendid-caribou logger.go:42: 07:54:40 | collector-otlp-production-http/1-install | starting test step 1-install logger.go:42: 07:54:40 | collector-otlp-production-http/1-install | Jaeger:kuttl-test-splendid-caribou/my-jaeger created logger.go:42: 07:55:39 | collector-otlp-production-http/1-install | test step completed 1-install logger.go:42: 07:55:39 | collector-otlp-production-http/2-smoke-test | starting test step 2-smoke-test logger.go:42: 07:55:39 | collector-otlp-production-http/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:55:41 | collector-otlp-production-http/2-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:55:49 | collector-otlp-production-http/2-smoke-test | running command: [sh -c REPORTING_PROTOCOL=http ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/otlp-smoke-test.yaml.template -o otlp-smoke-test-job.yaml] logger.go:42: 07:55:49 | collector-otlp-production-http/2-smoke-test | running command: [sh -c kubectl create -f otlp-smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:55:49 | collector-otlp-production-http/2-smoke-test | job.batch/report-span created logger.go:42: 07:55:49 | collector-otlp-production-http/2-smoke-test | job.batch/check-span created logger.go:42: 07:56:03 | collector-otlp-production-http/2-smoke-test | test step completed 2-smoke-test logger.go:42: 07:56:03 | collector-otlp-production-http | collector-otlp-production-http events from ns kuttl-test-splendid-caribou: logger.go:42: 07:56:03 | collector-otlp-production-http | 2025-02-03 07:55:09 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestsplendidcariboumyjaeger-1-7fbb56d7b7 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestsplendidcariboumyjaeger-1-7fbb5jgjbq replicaset-controller logger.go:42: 07:56:03 | collector-otlp-production-http | 2025-02-03 07:55:09 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsplendidcariboumyjaeger-1-7fbb5jgjbq Binding Scheduled Successfully assigned kuttl-test-splendid-caribou/elasticsearch-cdm-kuttltestsplendidcariboumyjaeger-1-7fbb5jgjbq to ip-10-0-66-98.us-east-2.compute.internal default-scheduler logger.go:42: 07:56:03 | collector-otlp-production-http | 2025-02-03 07:55:09 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestsplendidcariboumyjaeger-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestsplendidcariboumyjaeger-1-7fbb56d7b7 to 1 deployment-controller logger.go:42: 07:56:03 | collector-otlp-production-http | 2025-02-03 07:55:10 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsplendidcariboumyjaeger-1-7fbb5jgjbq AddedInterface Add eth0 [10.129.2.66/23] from ovn-kubernetes multus logger.go:42: 07:56:03 | collector-otlp-production-http | 2025-02-03 07:55:10 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsplendidcariboumyjaeger-1-7fbb5jgjbq.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:410674bf2f77ff1d431f494cadfab3e81249b35695eb65aedfdb8498079b4d35" already present on machine kubelet logger.go:42: 07:56:03 | collector-otlp-production-http | 2025-02-03 07:55:10 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsplendidcariboumyjaeger-1-7fbb5jgjbq.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:56:03 | collector-otlp-production-http | 2025-02-03 07:55:10 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsplendidcariboumyjaeger-1-7fbb5jgjbq.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:56:03 | collector-otlp-production-http | 2025-02-03 07:55:10 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsplendidcariboumyjaeger-1-7fbb5jgjbq.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:fe2439e5f9e14d828caace425214edfa3edc179d72d225c43f69e9e4c9a5837b" already present on machine kubelet logger.go:42: 07:56:03 | collector-otlp-production-http | 2025-02-03 07:55:10 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsplendidcariboumyjaeger-1-7fbb5jgjbq.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:56:03 | collector-otlp-production-http | 2025-02-03 07:55:10 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsplendidcariboumyjaeger-1-7fbb5jgjbq.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:56:03 | collector-otlp-production-http | 2025-02-03 07:55:25 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestsplendidcariboumyjaeger-1-7fbb5jgjbq.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:56:03 | collector-otlp-production-http | 2025-02-03 07:55:36 +0000 UTC Normal Pod my-jaeger-collector-579f5755d-jvkc9 Binding Scheduled Successfully assigned kuttl-test-splendid-caribou/my-jaeger-collector-579f5755d-jvkc9 to ip-10-0-6-172.us-east-2.compute.internal default-scheduler logger.go:42: 07:56:03 | collector-otlp-production-http | 2025-02-03 07:55:36 +0000 UTC Normal Pod my-jaeger-collector-579f5755d-jvkc9 AddedInterface Add eth0 [10.131.0.55/23] from ovn-kubernetes multus logger.go:42: 07:56:03 | collector-otlp-production-http | 2025-02-03 07:55:36 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-579f5755d SuccessfulCreate Created pod: my-jaeger-collector-579f5755d-jvkc9 replicaset-controller logger.go:42: 07:56:03 | collector-otlp-production-http | 2025-02-03 07:55:36 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-579f5755d to 1 deployment-controller logger.go:42: 07:56:03 | collector-otlp-production-http | 2025-02-03 07:55:36 +0000 UTC Normal Pod my-jaeger-query-b6c86b977-pgj88 Binding Scheduled Successfully assigned kuttl-test-splendid-caribou/my-jaeger-query-b6c86b977-pgj88 to ip-10-0-26-111.us-east-2.compute.internal default-scheduler logger.go:42: 07:56:03 | collector-otlp-production-http | 2025-02-03 07:55:36 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-b6c86b977 SuccessfulCreate Created pod: my-jaeger-query-b6c86b977-pgj88 replicaset-controller logger.go:42: 07:56:03 | collector-otlp-production-http | 2025-02-03 07:55:36 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-b6c86b977 to 1 deployment-controller logger.go:42: 07:56:03 | collector-otlp-production-http | 2025-02-03 07:55:37 +0000 UTC Normal Pod my-jaeger-collector-579f5755d-jvkc9.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:40a1c7fa23aa5ffa64e8e22aa38022f5d4d7ff644c46a3da7169b713d486c3c1" already present on machine kubelet logger.go:42: 07:56:03 | collector-otlp-production-http | 2025-02-03 07:55:37 +0000 UTC Normal Pod my-jaeger-collector-579f5755d-jvkc9.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:56:03 | collector-otlp-production-http | 2025-02-03 07:55:37 +0000 UTC Normal Pod my-jaeger-collector-579f5755d-jvkc9.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:56:03 | collector-otlp-production-http | 2025-02-03 07:55:37 +0000 UTC Normal Pod my-jaeger-query-b6c86b977-pgj88 AddedInterface Add eth0 [10.128.2.71/23] from ovn-kubernetes multus logger.go:42: 07:56:03 | collector-otlp-production-http | 2025-02-03 07:55:37 +0000 UTC Normal Pod my-jaeger-query-b6c86b977-pgj88.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:84020ce29bcb5a4bd018e6596188ae919c5cd600e08f78a546c0e76ea477685e" already present on machine kubelet logger.go:42: 07:56:03 | collector-otlp-production-http | 2025-02-03 07:55:37 +0000 UTC Normal Pod my-jaeger-query-b6c86b977-pgj88.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:56:03 | collector-otlp-production-http | 2025-02-03 07:55:37 +0000 UTC Normal Pod my-jaeger-query-b6c86b977-pgj88.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:56:03 | collector-otlp-production-http | 2025-02-03 07:55:37 +0000 UTC Normal Pod my-jaeger-query-b6c86b977-pgj88.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" already present on machine kubelet logger.go:42: 07:56:03 | collector-otlp-production-http | 2025-02-03 07:55:37 +0000 UTC Normal Pod my-jaeger-query-b6c86b977-pgj88.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:56:03 | collector-otlp-production-http | 2025-02-03 07:55:37 +0000 UTC Normal Pod my-jaeger-query-b6c86b977-pgj88.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:56:03 | collector-otlp-production-http | 2025-02-03 07:55:37 +0000 UTC Normal Pod my-jaeger-query-b6c86b977-pgj88.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:65342c7d622f336741a8ae0b9dead79c7ecabc155d02cdd42b7b49ca36680e74" already present on machine kubelet logger.go:42: 07:56:03 | collector-otlp-production-http | 2025-02-03 07:55:37 +0000 UTC Normal Pod my-jaeger-query-b6c86b977-pgj88.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:56:03 | collector-otlp-production-http | 2025-02-03 07:55:37 +0000 UTC Normal Pod my-jaeger-query-b6c86b977-pgj88.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:56:03 | collector-otlp-production-http | 2025-02-03 07:55:46 +0000 UTC Normal Pod my-jaeger-query-b6c86b977-pgj88.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:56:03 | collector-otlp-production-http | 2025-02-03 07:55:46 +0000 UTC Normal Pod my-jaeger-query-b6c86b977-pgj88.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:56:03 | collector-otlp-production-http | 2025-02-03 07:55:46 +0000 UTC Normal Pod my-jaeger-query-b6c86b977-pgj88.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:56:03 | collector-otlp-production-http | 2025-02-03 07:55:46 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-b6c86b977 SuccessfulDelete Deleted pod: my-jaeger-query-b6c86b977-pgj88 replicaset-controller logger.go:42: 07:56:03 | collector-otlp-production-http | 2025-02-03 07:55:46 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled down replica set my-jaeger-query-b6c86b977 to 0 from 1 deployment-controller logger.go:42: 07:56:03 | collector-otlp-production-http | 2025-02-03 07:55:47 +0000 UTC Normal Pod my-jaeger-query-7f59bc4cb4-4p5n4 Binding Scheduled Successfully assigned kuttl-test-splendid-caribou/my-jaeger-query-7f59bc4cb4-4p5n4 to ip-10-0-26-111.us-east-2.compute.internal default-scheduler logger.go:42: 07:56:03 | collector-otlp-production-http | 2025-02-03 07:55:47 +0000 UTC Normal Pod my-jaeger-query-7f59bc4cb4-4p5n4 AddedInterface Add eth0 [10.128.2.72/23] from ovn-kubernetes multus logger.go:42: 07:56:03 | collector-otlp-production-http | 2025-02-03 07:55:47 +0000 UTC Normal Pod my-jaeger-query-7f59bc4cb4-4p5n4.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:84020ce29bcb5a4bd018e6596188ae919c5cd600e08f78a546c0e76ea477685e" already present on machine kubelet logger.go:42: 07:56:03 | collector-otlp-production-http | 2025-02-03 07:55:47 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-7f59bc4cb4 SuccessfulCreate Created pod: my-jaeger-query-7f59bc4cb4-4p5n4 replicaset-controller logger.go:42: 07:56:03 | collector-otlp-production-http | 2025-02-03 07:55:47 +0000 UTC Warning Pod my-jaeger-query-b6c86b977-pgj88.spec.containers{jaeger-query} Unhealthy Readiness probe failed: Get "http://10.128.2.71:16687/": dial tcp 10.128.2.71:16687: i/o timeout (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:56:03 | collector-otlp-production-http | 2025-02-03 07:55:47 +0000 UTC Warning Pod my-jaeger-query-b6c86b977-pgj88.spec.containers{jaeger-agent} Unhealthy Readiness probe failed: Get "http://10.128.2.71:14271/": dial tcp 10.128.2.71:14271: i/o timeout (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:56:03 | collector-otlp-production-http | 2025-02-03 07:55:47 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-7f59bc4cb4 to 1 deployment-controller logger.go:42: 07:56:03 | collector-otlp-production-http | 2025-02-03 07:55:48 +0000 UTC Normal Pod my-jaeger-query-7f59bc4cb4-4p5n4.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:56:03 | collector-otlp-production-http | 2025-02-03 07:55:48 +0000 UTC Normal Pod my-jaeger-query-7f59bc4cb4-4p5n4.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:56:03 | collector-otlp-production-http | 2025-02-03 07:55:48 +0000 UTC Normal Pod my-jaeger-query-7f59bc4cb4-4p5n4.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" already present on machine kubelet logger.go:42: 07:56:03 | collector-otlp-production-http | 2025-02-03 07:55:48 +0000 UTC Normal Pod my-jaeger-query-7f59bc4cb4-4p5n4.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:56:03 | collector-otlp-production-http | 2025-02-03 07:55:48 +0000 UTC Normal Pod my-jaeger-query-7f59bc4cb4-4p5n4.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:56:03 | collector-otlp-production-http | 2025-02-03 07:55:48 +0000 UTC Normal Pod my-jaeger-query-7f59bc4cb4-4p5n4.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:65342c7d622f336741a8ae0b9dead79c7ecabc155d02cdd42b7b49ca36680e74" already present on machine kubelet logger.go:42: 07:56:03 | collector-otlp-production-http | 2025-02-03 07:55:48 +0000 UTC Normal Pod my-jaeger-query-7f59bc4cb4-4p5n4.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:56:03 | collector-otlp-production-http | 2025-02-03 07:55:48 +0000 UTC Normal Pod my-jaeger-query-7f59bc4cb4-4p5n4.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:56:03 | collector-otlp-production-http | 2025-02-03 07:55:49 +0000 UTC Normal Pod check-span-qkx7r Binding Scheduled Successfully assigned kuttl-test-splendid-caribou/check-span-qkx7r to ip-10-0-6-172.us-east-2.compute.internal default-scheduler logger.go:42: 07:56:03 | collector-otlp-production-http | 2025-02-03 07:55:49 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-qkx7r job-controller logger.go:42: 07:56:03 | collector-otlp-production-http | 2025-02-03 07:55:49 +0000 UTC Normal Pod report-span-nzs8d Binding Scheduled Successfully assigned kuttl-test-splendid-caribou/report-span-nzs8d to ip-10-0-6-172.us-east-2.compute.internal default-scheduler logger.go:42: 07:56:03 | collector-otlp-production-http | 2025-02-03 07:55:49 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-nzs8d job-controller logger.go:42: 07:56:03 | collector-otlp-production-http | 2025-02-03 07:55:50 +0000 UTC Normal Pod check-span-qkx7r AddedInterface Add eth0 [10.131.0.57/23] from ovn-kubernetes multus logger.go:42: 07:56:03 | collector-otlp-production-http | 2025-02-03 07:55:50 +0000 UTC Normal Pod check-span-qkx7r.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:56:03 | collector-otlp-production-http | 2025-02-03 07:55:50 +0000 UTC Normal Pod report-span-nzs8d AddedInterface Add eth0 [10.131.0.56/23] from ovn-kubernetes multus logger.go:42: 07:56:03 | collector-otlp-production-http | 2025-02-03 07:55:50 +0000 UTC Normal Pod report-span-nzs8d.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:56:03 | collector-otlp-production-http | 2025-02-03 07:55:51 +0000 UTC Normal Pod check-span-qkx7r.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 1.167s (1.167s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:56:03 | collector-otlp-production-http | 2025-02-03 07:55:51 +0000 UTC Normal Pod report-span-nzs8d.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 922ms (922ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:56:03 | collector-otlp-production-http | 2025-02-03 07:55:52 +0000 UTC Normal Pod check-span-qkx7r.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:56:03 | collector-otlp-production-http | 2025-02-03 07:55:52 +0000 UTC Normal Pod check-span-qkx7r.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:56:03 | collector-otlp-production-http | 2025-02-03 07:55:52 +0000 UTC Normal Pod report-span-nzs8d.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:56:03 | collector-otlp-production-http | 2025-02-03 07:55:52 +0000 UTC Normal Pod report-span-nzs8d.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:56:03 | collector-otlp-production-http | 2025-02-03 07:55:54 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 07:56:03 | collector-otlp-production-http | 2025-02-03 07:55:54 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get memory utilization: missing request for memory in container jaeger-collector of Pod my-jaeger-collector-579f5755d-jvkc9 horizontal-pod-autoscaler logger.go:42: 07:56:03 | collector-otlp-production-http | 2025-02-03 07:55:54 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 07:56:03 | collector-otlp-production-http | 2025-02-03 07:56:02 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:56:03 | collector-otlp-production-http | Deleting namespace: kuttl-test-splendid-caribou === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (419.93s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/collector-otlp-production-grpc (93.72s) --- PASS: kuttl/harness/artifacts (6.48s) --- PASS: kuttl/harness/collector-otlp-allinone-http (41.49s) --- PASS: kuttl/harness/collector-otlp-allinone-grpc (50.53s) --- PASS: kuttl/harness/collector-autoscale (61.10s) --- PASS: kuttl/harness/cassandra-spark (6.17s) --- PASS: kuttl/harness/set-custom-img (69.96s) --- PASS: kuttl/harness/collector-otlp-production-http (90.40s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name miscellaneous --report --output /logs/artifacts/miscellaneous.xml ./artifacts/kuttl-report.xml time="2025-02-03T07:56:11Z" level=debug msg="Setting a new name for the test suites" time="2025-02-03T07:56:11Z" level=debug msg="Removing 'artifacts' TestCase" time="2025-02-03T07:56:11Z" level=debug msg="normalizing test case names" time="2025-02-03T07:56:11Z" level=debug msg="miscellaneous/collector-otlp-production-grpc -> miscellaneous_collector_otlp_production_grpc" time="2025-02-03T07:56:11Z" level=debug msg="miscellaneous/artifacts -> miscellaneous_artifacts" time="2025-02-03T07:56:11Z" level=debug msg="miscellaneous/collector-otlp-allinone-http -> miscellaneous_collector_otlp_allinone_http" time="2025-02-03T07:56:11Z" level=debug msg="miscellaneous/collector-otlp-allinone-grpc -> miscellaneous_collector_otlp_allinone_grpc" time="2025-02-03T07:56:11Z" level=debug msg="miscellaneous/collector-autoscale -> miscellaneous_collector_autoscale" time="2025-02-03T07:56:11Z" level=debug msg="miscellaneous/cassandra-spark -> miscellaneous_cassandra_spark" time="2025-02-03T07:56:11Z" level=debug msg="miscellaneous/set-custom-img -> miscellaneous_set_custom_img" time="2025-02-03T07:56:11Z" level=debug msg="miscellaneous/collector-otlp-production-http -> miscellaneous_collector_otlp_production_http" +----------------------------------------------+--------+ | NAME | RESULT | +----------------------------------------------+--------+ | miscellaneous_collector_otlp_production_grpc | passed | | miscellaneous_artifacts | passed | | miscellaneous_collector_otlp_allinone_http | passed | | miscellaneous_collector_otlp_allinone_grpc | passed | | miscellaneous_collector_autoscale | passed | | miscellaneous_cassandra_spark | passed | | miscellaneous_set_custom_img | passed | | miscellaneous_collector_otlp_production_http | passed | +----------------------------------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh sidecar false true + '[' 3 -ne 3 ']' + test_suite_name=sidecar + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/sidecar.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-sidecar make[2]: Entering directory '/tmp/jaeger-tests' ./tests/e2e/sidecar/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.18.0-0.nightly-2025-02-01-185117 True False 47m Cluster version is 4.18.0-0.nightly-2025-02-01-185117' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.18.0-0.nightly-2025-02-01-185117 True False 47m Cluster version is 4.18.0-0.nightly-2025-02-01-185117' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/sidecar/render.sh ++ export SUITE_DIR=./tests/e2e/sidecar ++ SUITE_DIR=./tests/e2e/sidecar ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/sidecar ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + jaeger_service_name=order + start_test sidecar-deployment + '[' 1 -ne 1 ']' + test_name=sidecar-deployment + echo =========================================================================== =========================================================================== + info 'Rendering files for test sidecar-deployment' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test sidecar-deployment\e[0m' Rendering files for test sidecar-deployment + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/sidecar/_build + '[' _build '!=' _build ']' + mkdir -p sidecar-deployment + cd sidecar-deployment + render_install_vertx 01 + '[' 1 -ne 1 ']' + test_step=01 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-assert.yaml.template -o ./01-assert.yaml + render_find_service agent-as-sidecar allInOne order 00 03 + '[' 5 -ne 5 ']' + jaeger=agent-as-sidecar + deployment_strategy=allInOne + service_name=order + job_number=00 + test_step=03 + export JAEGER_NAME=agent-as-sidecar + JAEGER_NAME=agent-as-sidecar + export JOB_NUMBER=00 + JOB_NUMBER=00 + export SERVICE_NAME=order + SERVICE_NAME=order + export JAEGER_QUERY_ENDPOINT + '[' true = true ']' + '[' allInOne '!=' allInOne ']' + template=/tmp/jaeger-tests/tests/templates/find-service.yaml.template + JAEGER_QUERY_ENDPOINT=http://agent-as-sidecar-query:16686 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/find-service.yaml.template -o ./03-find-service.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-find-service.yaml.template -o ./03-assert.yaml + unset JAEGER_NAME + unset SERVICE_NAME + unset JOB_NUMBER + unset JAEGER_COLLECTOR_ENDPOINT + render_find_service agent-as-sidecar2 allInOne order 01 06 + '[' 5 -ne 5 ']' + jaeger=agent-as-sidecar2 + deployment_strategy=allInOne + service_name=order + job_number=01 + test_step=06 + export JAEGER_NAME=agent-as-sidecar2 + JAEGER_NAME=agent-as-sidecar2 + export JOB_NUMBER=01 + JOB_NUMBER=01 + export SERVICE_NAME=order + SERVICE_NAME=order + export JAEGER_QUERY_ENDPOINT + '[' true = true ']' + '[' allInOne '!=' allInOne ']' + template=/tmp/jaeger-tests/tests/templates/find-service.yaml.template + JAEGER_QUERY_ENDPOINT=http://agent-as-sidecar2-query:16686 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/find-service.yaml.template -o ./06-find-service.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-find-service.yaml.template -o ./06-assert.yaml + unset JAEGER_NAME + unset SERVICE_NAME + unset JOB_NUMBER + unset JAEGER_COLLECTOR_ENDPOINT + start_test sidecar-namespace + '[' 1 -ne 1 ']' + test_name=sidecar-namespace + echo =========================================================================== =========================================================================== + info 'Rendering files for test sidecar-namespace' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test sidecar-namespace\e[0m' Rendering files for test sidecar-namespace + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/sidecar/_build/sidecar-deployment + '[' sidecar-deployment '!=' _build ']' + cd .. + mkdir -p sidecar-namespace + cd sidecar-namespace + render_install_vertx 01 + '[' 1 -ne 1 ']' + test_step=01 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-assert.yaml.template -o ./01-assert.yaml + render_find_service agent-as-sidecar allInOne order 00 03 + '[' 5 -ne 5 ']' + jaeger=agent-as-sidecar + deployment_strategy=allInOne + service_name=order + job_number=00 + test_step=03 + export JAEGER_NAME=agent-as-sidecar + JAEGER_NAME=agent-as-sidecar + export JOB_NUMBER=00 + JOB_NUMBER=00 + export SERVICE_NAME=order + SERVICE_NAME=order + export JAEGER_QUERY_ENDPOINT + '[' true = true ']' + '[' allInOne '!=' allInOne ']' + template=/tmp/jaeger-tests/tests/templates/find-service.yaml.template + JAEGER_QUERY_ENDPOINT=http://agent-as-sidecar-query:16686 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/find-service.yaml.template -o ./03-find-service.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-find-service.yaml.template -o ./03-assert.yaml + unset JAEGER_NAME + unset SERVICE_NAME + unset JOB_NUMBER + unset JAEGER_COLLECTOR_ENDPOINT + render_find_service agent-as-sidecar2 allInOne order 01 06 + '[' 5 -ne 5 ']' + jaeger=agent-as-sidecar2 + deployment_strategy=allInOne + service_name=order + job_number=01 + test_step=06 + export JAEGER_NAME=agent-as-sidecar2 + JAEGER_NAME=agent-as-sidecar2 + export JOB_NUMBER=01 + JOB_NUMBER=01 + export SERVICE_NAME=order + SERVICE_NAME=order + export JAEGER_QUERY_ENDPOINT + '[' true = true ']' + '[' allInOne '!=' allInOne ']' + template=/tmp/jaeger-tests/tests/templates/find-service.yaml.template + JAEGER_QUERY_ENDPOINT=http://agent-as-sidecar2-query:16686 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/find-service.yaml.template -o ./06-find-service.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-find-service.yaml.template -o ./06-assert.yaml + unset JAEGER_NAME + unset SERVICE_NAME + unset JOB_NUMBER + unset JAEGER_COLLECTOR_ENDPOINT + start_test sidecar-skip-webhook + '[' 1 -ne 1 ']' + test_name=sidecar-skip-webhook + echo =========================================================================== =========================================================================== + info 'Rendering files for test sidecar-skip-webhook' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test sidecar-skip-webhook\e[0m' Rendering files for test sidecar-skip-webhook + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/sidecar/_build/sidecar-namespace + '[' sidecar-namespace '!=' _build ']' + cd .. + mkdir -p sidecar-skip-webhook + cd sidecar-skip-webhook + render_install_vertx 01 + '[' 1 -ne 1 ']' + test_step=01 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-assert.yaml.template -o ./01-assert.yaml make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running sidecar E2E tests' Running sidecar E2E tests + cd tests/e2e/sidecar/_build + set +e + KUBECONFIG=/tmp/kubeconfig-2531963025 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-d102g9k0-f1a62.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 4 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/sidecar-deployment === PAUSE kuttl/harness/sidecar-deployment === RUN kuttl/harness/sidecar-namespace === PAUSE kuttl/harness/sidecar-namespace === RUN kuttl/harness/sidecar-skip-webhook === PAUSE kuttl/harness/sidecar-skip-webhook === CONT kuttl/harness/artifacts logger.go:42: 07:56:19 | artifacts | Creating namespace: kuttl-test-more-bluebird logger.go:42: 07:56:19 | artifacts | artifacts events from ns kuttl-test-more-bluebird: logger.go:42: 07:56:19 | artifacts | Deleting namespace: kuttl-test-more-bluebird === CONT kuttl/harness/sidecar-namespace logger.go:42: 07:56:26 | sidecar-namespace | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:56:26 | sidecar-namespace | Creating namespace: kuttl-test-on-porpoise logger.go:42: 07:56:26 | sidecar-namespace/0-install | starting test step 0-install logger.go:42: 07:56:26 | sidecar-namespace/0-install | Jaeger:kuttl-test-on-porpoise/agent-as-sidecar created logger.go:42: 07:56:32 | sidecar-namespace/0-install | test step completed 0-install logger.go:42: 07:56:32 | sidecar-namespace/1-install | starting test step 1-install logger.go:42: 07:56:32 | sidecar-namespace/1-install | Deployment:kuttl-test-on-porpoise/vertx-create-span-sidecar created logger.go:42: 07:56:34 | sidecar-namespace/1-install | test step completed 1-install logger.go:42: 07:56:34 | sidecar-namespace/2-enable-injection | starting test step 2-enable-injection logger.go:42: 07:56:34 | sidecar-namespace/2-enable-injection | running command: [sh -c kubectl annotate --overwrite namespaces $NAMESPACE "sidecar.jaegertracing.io/inject"="true"] logger.go:42: 07:56:34 | sidecar-namespace/2-enable-injection | namespace/kuttl-test-on-porpoise annotated logger.go:42: 07:56:36 | sidecar-namespace/2-enable-injection | test step completed 2-enable-injection logger.go:42: 07:56:36 | sidecar-namespace/3-find-service | starting test step 3-find-service logger.go:42: 07:56:36 | sidecar-namespace/3-find-service | Job:kuttl-test-on-porpoise/00-find-service created logger.go:42: 07:56:48 | sidecar-namespace/3-find-service | test step completed 3-find-service logger.go:42: 07:56:48 | sidecar-namespace/4-other-instance | starting test step 4-other-instance logger.go:42: 07:56:48 | sidecar-namespace/4-other-instance | Jaeger:kuttl-test-on-porpoise/agent-as-sidecar2 created logger.go:42: 07:56:59 | sidecar-namespace/4-other-instance | test step completed 4-other-instance logger.go:42: 07:56:59 | sidecar-namespace/5-delete-first-instance | starting test step 5-delete-first-instance logger.go:42: 07:56:59 | sidecar-namespace/5-delete-first-instance | test step completed 5-delete-first-instance logger.go:42: 07:56:59 | sidecar-namespace/6-find-service | starting test step 6-find-service logger.go:42: 07:56:59 | sidecar-namespace/6-find-service | Job:kuttl-test-on-porpoise/01-find-service created logger.go:42: 07:57:20 | sidecar-namespace/6-find-service | test step completed 6-find-service logger.go:42: 07:57:20 | sidecar-namespace/7-disable-injection | starting test step 7-disable-injection logger.go:42: 07:57:20 | sidecar-namespace/7-disable-injection | running command: [sh -c kubectl annotate --overwrite namespaces $NAMESPACE "sidecar.jaegertracing.io/inject"="false"] logger.go:42: 07:57:21 | sidecar-namespace/7-disable-injection | namespace/kuttl-test-on-porpoise annotated logger.go:42: 07:57:23 | sidecar-namespace/7-disable-injection | test step completed 7-disable-injection logger.go:42: 07:57:23 | sidecar-namespace | sidecar-namespace events from ns kuttl-test-on-porpoise: logger.go:42: 07:57:23 | sidecar-namespace | 2025-02-03 07:56:30 +0000 UTC Normal Pod agent-as-sidecar-68778d4fb7-s8bsj Binding Scheduled Successfully assigned kuttl-test-on-porpoise/agent-as-sidecar-68778d4fb7-s8bsj to ip-10-0-66-98.us-east-2.compute.internal default-scheduler logger.go:42: 07:57:23 | sidecar-namespace | 2025-02-03 07:56:30 +0000 UTC Normal Pod agent-as-sidecar-68778d4fb7-s8bsj AddedInterface Add eth0 [10.129.2.67/23] from ovn-kubernetes multus logger.go:42: 07:57:23 | sidecar-namespace | 2025-02-03 07:56:30 +0000 UTC Normal Pod agent-as-sidecar-68778d4fb7-s8bsj.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:fa9aaf9ae026228265ce7f10fb451d5c52a4e0269ce1fa9024cbefddf765ae8d" already present on machine kubelet logger.go:42: 07:57:23 | sidecar-namespace | 2025-02-03 07:56:30 +0000 UTC Normal Pod agent-as-sidecar-68778d4fb7-s8bsj.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:57:23 | sidecar-namespace | 2025-02-03 07:56:30 +0000 UTC Normal ReplicaSet.apps agent-as-sidecar-68778d4fb7 SuccessfulCreate Created pod: agent-as-sidecar-68778d4fb7-s8bsj replicaset-controller logger.go:42: 07:57:23 | sidecar-namespace | 2025-02-03 07:56:30 +0000 UTC Normal Deployment.apps agent-as-sidecar ScalingReplicaSet Scaled up replica set agent-as-sidecar-68778d4fb7 to 1 deployment-controller logger.go:42: 07:57:23 | sidecar-namespace | 2025-02-03 07:56:31 +0000 UTC Normal Pod agent-as-sidecar-68778d4fb7-s8bsj.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:57:23 | sidecar-namespace | 2025-02-03 07:56:32 +0000 UTC Normal Pod vertx-create-span-sidecar-797645c8fc-kg9h6 Binding Scheduled Successfully assigned kuttl-test-on-porpoise/vertx-create-span-sidecar-797645c8fc-kg9h6 to ip-10-0-6-172.us-east-2.compute.internal default-scheduler logger.go:42: 07:57:23 | sidecar-namespace | 2025-02-03 07:56:32 +0000 UTC Normal Pod vertx-create-span-sidecar-797645c8fc-kg9h6 AddedInterface Add eth0 [10.131.0.58/23] from ovn-kubernetes multus logger.go:42: 07:57:23 | sidecar-namespace | 2025-02-03 07:56:32 +0000 UTC Normal Pod vertx-create-span-sidecar-797645c8fc-kg9h6.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:57:23 | sidecar-namespace | 2025-02-03 07:56:32 +0000 UTC Normal Pod vertx-create-span-sidecar-797645c8fc-kg9h6.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:57:23 | sidecar-namespace | 2025-02-03 07:56:32 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-797645c8fc SuccessfulCreate Created pod: vertx-create-span-sidecar-797645c8fc-kg9h6 replicaset-controller logger.go:42: 07:57:23 | sidecar-namespace | 2025-02-03 07:56:32 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-797645c8fc to 1 deployment-controller logger.go:42: 07:57:23 | sidecar-namespace | 2025-02-03 07:56:33 +0000 UTC Normal Pod vertx-create-span-sidecar-797645c8fc-kg9h6.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:57:23 | sidecar-namespace | 2025-02-03 07:56:34 +0000 UTC Normal Pod vertx-create-span-sidecar-7f7797c89b-9v88g Binding Scheduled Successfully assigned kuttl-test-on-porpoise/vertx-create-span-sidecar-7f7797c89b-9v88g to ip-10-0-6-172.us-east-2.compute.internal default-scheduler logger.go:42: 07:57:23 | sidecar-namespace | 2025-02-03 07:56:34 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-7f7797c89b SuccessfulCreate Created pod: vertx-create-span-sidecar-7f7797c89b-9v88g replicaset-controller logger.go:42: 07:57:23 | sidecar-namespace | 2025-02-03 07:56:34 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-7f7797c89b to 1 deployment-controller logger.go:42: 07:57:23 | sidecar-namespace | 2025-02-03 07:56:35 +0000 UTC Normal Pod vertx-create-span-sidecar-7f7797c89b-9v88g AddedInterface Add eth0 [10.131.0.59/23] from ovn-kubernetes multus logger.go:42: 07:57:23 | sidecar-namespace | 2025-02-03 07:56:35 +0000 UTC Normal Pod vertx-create-span-sidecar-7f7797c89b-9v88g.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:57:23 | sidecar-namespace | 2025-02-03 07:56:35 +0000 UTC Normal Pod vertx-create-span-sidecar-7f7797c89b-9v88g.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:57:23 | sidecar-namespace | 2025-02-03 07:56:35 +0000 UTC Normal Pod vertx-create-span-sidecar-7f7797c89b-9v88g.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:57:23 | sidecar-namespace | 2025-02-03 07:56:35 +0000 UTC Normal Pod vertx-create-span-sidecar-7f7797c89b-9v88g.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:65342c7d622f336741a8ae0b9dead79c7ecabc155d02cdd42b7b49ca36680e74" already present on machine kubelet logger.go:42: 07:57:23 | sidecar-namespace | 2025-02-03 07:56:35 +0000 UTC Normal Pod vertx-create-span-sidecar-7f7797c89b-9v88g.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:57:23 | sidecar-namespace | 2025-02-03 07:56:35 +0000 UTC Normal Pod vertx-create-span-sidecar-7f7797c89b-9v88g.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:57:23 | sidecar-namespace | 2025-02-03 07:56:36 +0000 UTC Normal Pod 00-find-service-sxdhv Binding Scheduled Successfully assigned kuttl-test-on-porpoise/00-find-service-sxdhv to ip-10-0-26-111.us-east-2.compute.internal default-scheduler logger.go:42: 07:57:23 | sidecar-namespace | 2025-02-03 07:56:36 +0000 UTC Normal Job.batch 00-find-service SuccessfulCreate Created pod: 00-find-service-sxdhv job-controller logger.go:42: 07:57:23 | sidecar-namespace | 2025-02-03 07:56:37 +0000 UTC Normal Pod 00-find-service-sxdhv AddedInterface Add eth0 [10.128.2.73/23] from ovn-kubernetes multus logger.go:42: 07:57:23 | sidecar-namespace | 2025-02-03 07:56:37 +0000 UTC Normal Pod 00-find-service-sxdhv.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:57:23 | sidecar-namespace | 2025-02-03 07:56:37 +0000 UTC Normal Pod 00-find-service-sxdhv.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 472ms (472ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:57:23 | sidecar-namespace | 2025-02-03 07:56:37 +0000 UTC Normal Pod 00-find-service-sxdhv.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:57:23 | sidecar-namespace | 2025-02-03 07:56:37 +0000 UTC Normal Pod 00-find-service-sxdhv.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:57:23 | sidecar-namespace | 2025-02-03 07:56:40 +0000 UTC Warning Pod vertx-create-span-sidecar-797645c8fc-kg9h6.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.131.0.58:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:57:23 | sidecar-namespace | 2025-02-03 07:56:40 +0000 UTC Warning Pod vertx-create-span-sidecar-797645c8fc-kg9h6.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.131.0.58:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:57:23 | sidecar-namespace | 2025-02-03 07:56:42 +0000 UTC Normal Pod vertx-create-span-sidecar-797645c8fc-kg9h6.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 07:57:23 | sidecar-namespace | 2025-02-03 07:56:43 +0000 UTC Warning Pod vertx-create-span-sidecar-797645c8fc-kg9h6.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.131.0.58:8080/": read tcp 10.131.0.2:32960->10.131.0.58:8080: read: connection reset by peer kubelet logger.go:42: 07:57:23 | sidecar-namespace | 2025-02-03 07:56:43 +0000 UTC Warning Pod vertx-create-span-sidecar-797645c8fc-kg9h6.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.131.0.58:8080/": dial tcp 10.131.0.58:8080: connect: connection refused kubelet logger.go:42: 07:57:23 | sidecar-namespace | 2025-02-03 07:56:43 +0000 UTC Warning Pod vertx-create-span-sidecar-7f7797c89b-9v88g.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.131.0.59:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:57:23 | sidecar-namespace | 2025-02-03 07:56:43 +0000 UTC Warning Pod vertx-create-span-sidecar-7f7797c89b-9v88g.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.131.0.59:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:57:23 | sidecar-namespace | 2025-02-03 07:56:45 +0000 UTC Normal Pod vertx-create-span-sidecar-7f7797c89b-9v88g.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 07:57:23 | sidecar-namespace | 2025-02-03 07:56:46 +0000 UTC Warning Pod vertx-create-span-sidecar-7f7797c89b-9v88g.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.131.0.59:8080/": read tcp 10.131.0.2:35638->10.131.0.59:8080: read: connection reset by peer kubelet logger.go:42: 07:57:23 | sidecar-namespace | 2025-02-03 07:56:46 +0000 UTC Warning Pod vertx-create-span-sidecar-7f7797c89b-9v88g.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.131.0.59:8080/": dial tcp 10.131.0.59:8080: connect: connection refused kubelet logger.go:42: 07:57:23 | sidecar-namespace | 2025-02-03 07:56:48 +0000 UTC Normal Job.batch 00-find-service Completed Job completed job-controller logger.go:42: 07:57:23 | sidecar-namespace | 2025-02-03 07:56:54 +0000 UTC Warning Pod vertx-create-span-sidecar-797645c8fc-kg9h6.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.131.0.58:8080/": read tcp 10.131.0.2:36970->10.131.0.58:8080: read: connection reset by peer kubelet logger.go:42: 07:57:23 | sidecar-namespace | 2025-02-03 07:56:55 +0000 UTC Normal Pod agent-as-sidecar2-bb8887c45-zr4gq Binding Scheduled Successfully assigned kuttl-test-on-porpoise/agent-as-sidecar2-bb8887c45-zr4gq to ip-10-0-26-111.us-east-2.compute.internal default-scheduler logger.go:42: 07:57:23 | sidecar-namespace | 2025-02-03 07:56:55 +0000 UTC Normal Pod agent-as-sidecar2-bb8887c45-zr4gq AddedInterface Add eth0 [10.128.2.74/23] from ovn-kubernetes multus logger.go:42: 07:57:23 | sidecar-namespace | 2025-02-03 07:56:55 +0000 UTC Normal Pod agent-as-sidecar2-bb8887c45-zr4gq.spec.containers{jaeger} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:fa9aaf9ae026228265ce7f10fb451d5c52a4e0269ce1fa9024cbefddf765ae8d" kubelet logger.go:42: 07:57:23 | sidecar-namespace | 2025-02-03 07:56:55 +0000 UTC Normal ReplicaSet.apps agent-as-sidecar2-bb8887c45 SuccessfulCreate Created pod: agent-as-sidecar2-bb8887c45-zr4gq replicaset-controller logger.go:42: 07:57:23 | sidecar-namespace | 2025-02-03 07:56:55 +0000 UTC Normal Deployment.apps agent-as-sidecar2 ScalingReplicaSet Scaled up replica set agent-as-sidecar2-bb8887c45 to 1 deployment-controller logger.go:42: 07:57:23 | sidecar-namespace | 2025-02-03 07:56:57 +0000 UTC Normal Pod agent-as-sidecar2-bb8887c45-zr4gq.spec.containers{jaeger} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:fa9aaf9ae026228265ce7f10fb451d5c52a4e0269ce1fa9024cbefddf765ae8d" in 1.971s (1.971s including waiting). Image size: 144304495 bytes. kubelet logger.go:42: 07:57:23 | sidecar-namespace | 2025-02-03 07:56:57 +0000 UTC Normal Pod agent-as-sidecar2-bb8887c45-zr4gq.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:57:23 | sidecar-namespace | 2025-02-03 07:56:58 +0000 UTC Normal Pod agent-as-sidecar2-bb8887c45-zr4gq.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:57:23 | sidecar-namespace | 2025-02-03 07:56:59 +0000 UTC Normal Pod 01-find-service-8vz4q Binding Scheduled Successfully assigned kuttl-test-on-porpoise/01-find-service-8vz4q to ip-10-0-66-98.us-east-2.compute.internal default-scheduler logger.go:42: 07:57:23 | sidecar-namespace | 2025-02-03 07:56:59 +0000 UTC Normal Job.batch 01-find-service SuccessfulCreate Created pod: 01-find-service-8vz4q job-controller logger.go:42: 07:57:23 | sidecar-namespace | 2025-02-03 07:56:59 +0000 UTC Normal Pod agent-as-sidecar-68778d4fb7-s8bsj.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:57:23 | sidecar-namespace | 2025-02-03 07:57:00 +0000 UTC Normal Pod 01-find-service-8vz4q AddedInterface Add eth0 [10.129.2.68/23] from ovn-kubernetes multus logger.go:42: 07:57:23 | sidecar-namespace | 2025-02-03 07:57:00 +0000 UTC Normal Pod 01-find-service-8vz4q.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:57:23 | sidecar-namespace | 2025-02-03 07:57:01 +0000 UTC Normal Pod 01-find-service-8vz4q.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 690ms (690ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:57:23 | sidecar-namespace | 2025-02-03 07:57:01 +0000 UTC Normal Pod 01-find-service-8vz4q.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:57:23 | sidecar-namespace | 2025-02-03 07:57:01 +0000 UTC Normal Pod 01-find-service-8vz4q.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:57:23 | sidecar-namespace | 2025-02-03 07:57:02 +0000 UTC Normal Pod vertx-create-span-sidecar-6fc85b58-2n4ts Binding Scheduled Successfully assigned kuttl-test-on-porpoise/vertx-create-span-sidecar-6fc85b58-2n4ts to ip-10-0-66-98.us-east-2.compute.internal default-scheduler logger.go:42: 07:57:23 | sidecar-namespace | 2025-02-03 07:57:02 +0000 UTC Normal Pod vertx-create-span-sidecar-6fc85b58-2n4ts AddedInterface Add eth0 [10.129.2.69/23] from ovn-kubernetes multus logger.go:42: 07:57:23 | sidecar-namespace | 2025-02-03 07:57:02 +0000 UTC Normal Pod vertx-create-span-sidecar-6fc85b58-2n4ts.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:57:23 | sidecar-namespace | 2025-02-03 07:57:02 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-6fc85b58 SuccessfulCreate Created pod: vertx-create-span-sidecar-6fc85b58-2n4ts replicaset-controller logger.go:42: 07:57:23 | sidecar-namespace | 2025-02-03 07:57:02 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-797645c8fc SuccessfulDelete Deleted pod: vertx-create-span-sidecar-797645c8fc-kg9h6 replicaset-controller logger.go:42: 07:57:23 | sidecar-namespace | 2025-02-03 07:57:02 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled down replica set vertx-create-span-sidecar-797645c8fc to 0 from 1 deployment-controller logger.go:42: 07:57:23 | sidecar-namespace | 2025-02-03 07:57:02 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-6fc85b58 to 1 from 0 deployment-controller logger.go:42: 07:57:23 | sidecar-namespace | 2025-02-03 07:57:03 +0000 UTC Normal Pod vertx-create-span-sidecar-6fc85b58-2n4ts.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:57:23 | sidecar-namespace | 2025-02-03 07:57:03 +0000 UTC Normal Pod vertx-create-span-sidecar-6fc85b58-2n4ts.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:57:23 | sidecar-namespace | 2025-02-03 07:57:03 +0000 UTC Normal Pod vertx-create-span-sidecar-6fc85b58-2n4ts.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:65342c7d622f336741a8ae0b9dead79c7ecabc155d02cdd42b7b49ca36680e74" already present on machine kubelet logger.go:42: 07:57:23 | sidecar-namespace | 2025-02-03 07:57:03 +0000 UTC Normal Pod vertx-create-span-sidecar-6fc85b58-2n4ts.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:57:23 | sidecar-namespace | 2025-02-03 07:57:03 +0000 UTC Normal Pod vertx-create-span-sidecar-6fc85b58-2n4ts.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:57:23 | sidecar-namespace | 2025-02-03 07:57:11 +0000 UTC Warning Pod vertx-create-span-sidecar-6fc85b58-2n4ts.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.69:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:57:23 | sidecar-namespace | 2025-02-03 07:57:11 +0000 UTC Warning Pod vertx-create-span-sidecar-6fc85b58-2n4ts.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.129.2.69:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:57:23 | sidecar-namespace | 2025-02-03 07:57:13 +0000 UTC Normal Pod vertx-create-span-sidecar-6fc85b58-2n4ts.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 07:57:23 | sidecar-namespace | 2025-02-03 07:57:14 +0000 UTC Warning Pod vertx-create-span-sidecar-6fc85b58-2n4ts.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.69:8080/": read tcp 10.129.2.2:32834->10.129.2.69:8080: read: connection reset by peer kubelet logger.go:42: 07:57:23 | sidecar-namespace | 2025-02-03 07:57:14 +0000 UTC Warning Pod vertx-create-span-sidecar-6fc85b58-2n4ts.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.69:8080/": dial tcp 10.129.2.69:8080: connect: connection refused kubelet logger.go:42: 07:57:23 | sidecar-namespace | 2025-02-03 07:57:20 +0000 UTC Normal Job.batch 01-find-service Completed Job completed job-controller logger.go:42: 07:57:23 | sidecar-namespace | 2025-02-03 07:57:21 +0000 UTC Normal Pod vertx-create-span-sidecar-56d44b9595-sp6qv Binding Scheduled Successfully assigned kuttl-test-on-porpoise/vertx-create-span-sidecar-56d44b9595-sp6qv to ip-10-0-66-98.us-east-2.compute.internal default-scheduler logger.go:42: 07:57:23 | sidecar-namespace | 2025-02-03 07:57:21 +0000 UTC Normal Pod vertx-create-span-sidecar-56d44b9595-sp6qv AddedInterface Add eth0 [10.129.2.70/23] from ovn-kubernetes multus logger.go:42: 07:57:23 | sidecar-namespace | 2025-02-03 07:57:21 +0000 UTC Normal Pod vertx-create-span-sidecar-56d44b9595-sp6qv.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:57:23 | sidecar-namespace | 2025-02-03 07:57:21 +0000 UTC Normal Pod vertx-create-span-sidecar-56d44b9595-sp6qv.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:57:23 | sidecar-namespace | 2025-02-03 07:57:21 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-56d44b9595 SuccessfulCreate Created pod: vertx-create-span-sidecar-56d44b9595-sp6qv replicaset-controller logger.go:42: 07:57:23 | sidecar-namespace | 2025-02-03 07:57:21 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-7f7797c89b SuccessfulDelete Deleted pod: vertx-create-span-sidecar-7f7797c89b-9v88g replicaset-controller logger.go:42: 07:57:23 | sidecar-namespace | 2025-02-03 07:57:21 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled down replica set vertx-create-span-sidecar-7f7797c89b to 0 from 1 deployment-controller logger.go:42: 07:57:23 | sidecar-namespace | 2025-02-03 07:57:21 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-56d44b9595 to 1 from 0 deployment-controller logger.go:42: 07:57:23 | sidecar-namespace | 2025-02-03 07:57:22 +0000 UTC Normal Pod vertx-create-span-sidecar-56d44b9595-sp6qv.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:57:23 | sidecar-namespace | Deleting namespace: kuttl-test-on-porpoise === CONT kuttl/harness/sidecar-skip-webhook logger.go:42: 07:57:30 | sidecar-skip-webhook | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:57:30 | sidecar-skip-webhook | Creating namespace: kuttl-test-lasting-amoeba logger.go:42: 07:57:30 | sidecar-skip-webhook/0-install | starting test step 0-install logger.go:42: 07:57:30 | sidecar-skip-webhook/0-install | Jaeger:kuttl-test-lasting-amoeba/agent-as-sidecar created logger.go:42: 07:57:37 | sidecar-skip-webhook/0-install | test step completed 0-install logger.go:42: 07:57:37 | sidecar-skip-webhook/1-install | starting test step 1-install logger.go:42: 07:57:37 | sidecar-skip-webhook/1-install | Deployment:kuttl-test-lasting-amoeba/vertx-create-span-sidecar created logger.go:42: 07:57:39 | sidecar-skip-webhook/1-install | test step completed 1-install logger.go:42: 07:57:39 | sidecar-skip-webhook/2-add-anotation-and-label | starting test step 2-add-anotation-and-label logger.go:42: 07:57:39 | sidecar-skip-webhook/2-add-anotation-and-label | running command: [kubectl label deployment vertx-create-span-sidecar app.kubernetes.io/name=jaeger-operator --namespace kuttl-test-lasting-amoeba] logger.go:42: 07:57:39 | sidecar-skip-webhook/2-add-anotation-and-label | deployment.apps/vertx-create-span-sidecar labeled logger.go:42: 07:57:39 | sidecar-skip-webhook/2-add-anotation-and-label | running command: [kubectl annotate --overwrite deployment vertx-create-span-sidecar sidecar.jaegertracing.io/inject=true --namespace kuttl-test-lasting-amoeba] logger.go:42: 07:57:39 | sidecar-skip-webhook/2-add-anotation-and-label | deployment.apps/vertx-create-span-sidecar annotated logger.go:42: 07:57:39 | sidecar-skip-webhook/2-add-anotation-and-label | test step completed 2-add-anotation-and-label logger.go:42: 07:57:39 | sidecar-skip-webhook/3-remove-label | starting test step 3-remove-label logger.go:42: 07:57:39 | sidecar-skip-webhook/3-remove-label | running command: [kubectl label deployment vertx-create-span-sidecar app.kubernetes.io/name- --namespace kuttl-test-lasting-amoeba] logger.go:42: 07:57:40 | sidecar-skip-webhook/3-remove-label | deployment.apps/vertx-create-span-sidecar unlabeled logger.go:42: 07:57:42 | sidecar-skip-webhook/3-remove-label | test step completed 3-remove-label logger.go:42: 07:57:42 | sidecar-skip-webhook | sidecar-skip-webhook events from ns kuttl-test-lasting-amoeba: logger.go:42: 07:57:42 | sidecar-skip-webhook | 2025-02-03 07:57:34 +0000 UTC Normal Pod agent-as-sidecar-649bbbccf6-hvsn6 Binding Scheduled Successfully assigned kuttl-test-lasting-amoeba/agent-as-sidecar-649bbbccf6-hvsn6 to ip-10-0-66-98.us-east-2.compute.internal default-scheduler logger.go:42: 07:57:42 | sidecar-skip-webhook | 2025-02-03 07:57:34 +0000 UTC Normal ReplicaSet.apps agent-as-sidecar-649bbbccf6 SuccessfulCreate Created pod: agent-as-sidecar-649bbbccf6-hvsn6 replicaset-controller logger.go:42: 07:57:42 | sidecar-skip-webhook | 2025-02-03 07:57:34 +0000 UTC Normal Deployment.apps agent-as-sidecar ScalingReplicaSet Scaled up replica set agent-as-sidecar-649bbbccf6 to 1 deployment-controller logger.go:42: 07:57:42 | sidecar-skip-webhook | 2025-02-03 07:57:35 +0000 UTC Normal Pod agent-as-sidecar-649bbbccf6-hvsn6 AddedInterface Add eth0 [10.129.2.71/23] from ovn-kubernetes multus logger.go:42: 07:57:42 | sidecar-skip-webhook | 2025-02-03 07:57:35 +0000 UTC Normal Pod agent-as-sidecar-649bbbccf6-hvsn6.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:fa9aaf9ae026228265ce7f10fb451d5c52a4e0269ce1fa9024cbefddf765ae8d" already present on machine kubelet logger.go:42: 07:57:42 | sidecar-skip-webhook | 2025-02-03 07:57:35 +0000 UTC Normal Pod agent-as-sidecar-649bbbccf6-hvsn6.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:57:42 | sidecar-skip-webhook | 2025-02-03 07:57:35 +0000 UTC Normal Pod agent-as-sidecar-649bbbccf6-hvsn6.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:57:42 | sidecar-skip-webhook | 2025-02-03 07:57:37 +0000 UTC Normal Pod vertx-create-span-sidecar-797645c8fc-6mhhl Binding Scheduled Successfully assigned kuttl-test-lasting-amoeba/vertx-create-span-sidecar-797645c8fc-6mhhl to ip-10-0-6-172.us-east-2.compute.internal default-scheduler logger.go:42: 07:57:42 | sidecar-skip-webhook | 2025-02-03 07:57:37 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-797645c8fc SuccessfulCreate Created pod: vertx-create-span-sidecar-797645c8fc-6mhhl replicaset-controller logger.go:42: 07:57:42 | sidecar-skip-webhook | 2025-02-03 07:57:37 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-797645c8fc to 1 deployment-controller logger.go:42: 07:57:42 | sidecar-skip-webhook | 2025-02-03 07:57:38 +0000 UTC Normal Pod vertx-create-span-sidecar-797645c8fc-6mhhl AddedInterface Add eth0 [10.131.0.60/23] from ovn-kubernetes multus logger.go:42: 07:57:42 | sidecar-skip-webhook | 2025-02-03 07:57:38 +0000 UTC Normal Pod vertx-create-span-sidecar-797645c8fc-6mhhl.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:57:42 | sidecar-skip-webhook | 2025-02-03 07:57:38 +0000 UTC Normal Pod vertx-create-span-sidecar-797645c8fc-6mhhl.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:57:42 | sidecar-skip-webhook | 2025-02-03 07:57:38 +0000 UTC Normal Pod vertx-create-span-sidecar-797645c8fc-6mhhl.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:57:42 | sidecar-skip-webhook | 2025-02-03 07:57:39 +0000 UTC Normal Pod vertx-create-span-sidecar-5bd68b5fc-n2qkp Binding Scheduled Successfully assigned kuttl-test-lasting-amoeba/vertx-create-span-sidecar-5bd68b5fc-n2qkp to ip-10-0-66-98.us-east-2.compute.internal default-scheduler logger.go:42: 07:57:42 | sidecar-skip-webhook | 2025-02-03 07:57:39 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-5bd68b5fc SuccessfulCreate Created pod: vertx-create-span-sidecar-5bd68b5fc-n2qkp replicaset-controller logger.go:42: 07:57:42 | sidecar-skip-webhook | 2025-02-03 07:57:39 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-5bd68b5fc to 1 deployment-controller logger.go:42: 07:57:42 | sidecar-skip-webhook | 2025-02-03 07:57:40 +0000 UTC Normal Pod vertx-create-span-sidecar-5bd68b5fc-n2qkp AddedInterface Add eth0 [10.129.2.72/23] from ovn-kubernetes multus logger.go:42: 07:57:42 | sidecar-skip-webhook | 2025-02-03 07:57:40 +0000 UTC Normal Pod vertx-create-span-sidecar-5bd68b5fc-n2qkp.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:57:42 | sidecar-skip-webhook | 2025-02-03 07:57:40 +0000 UTC Normal Pod vertx-create-span-sidecar-5bd68b5fc-n2qkp.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:57:42 | sidecar-skip-webhook | 2025-02-03 07:57:40 +0000 UTC Normal Pod vertx-create-span-sidecar-5bd68b5fc-n2qkp.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:57:42 | sidecar-skip-webhook | 2025-02-03 07:57:40 +0000 UTC Normal Pod vertx-create-span-sidecar-5bd68b5fc-n2qkp.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:65342c7d622f336741a8ae0b9dead79c7ecabc155d02cdd42b7b49ca36680e74" already present on machine kubelet logger.go:42: 07:57:42 | sidecar-skip-webhook | 2025-02-03 07:57:40 +0000 UTC Normal Pod vertx-create-span-sidecar-5bd68b5fc-n2qkp.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:57:42 | sidecar-skip-webhook | 2025-02-03 07:57:40 +0000 UTC Normal Pod vertx-create-span-sidecar-5bd68b5fc-n2qkp.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:57:42 | sidecar-skip-webhook | Deleting namespace: kuttl-test-lasting-amoeba === CONT kuttl/harness/sidecar-deployment logger.go:42: 07:57:49 | sidecar-deployment | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:57:49 | sidecar-deployment | Creating namespace: kuttl-test-enabling-liger logger.go:42: 07:57:49 | sidecar-deployment/0-install | starting test step 0-install logger.go:42: 07:57:49 | sidecar-deployment/0-install | Jaeger:kuttl-test-enabling-liger/agent-as-sidecar created logger.go:42: 07:57:55 | sidecar-deployment/0-install | test step completed 0-install logger.go:42: 07:57:55 | sidecar-deployment/1-install | starting test step 1-install logger.go:42: 07:57:55 | sidecar-deployment/1-install | Deployment:kuttl-test-enabling-liger/vertx-create-span-sidecar created logger.go:42: 07:57:57 | sidecar-deployment/1-install | test step completed 1-install logger.go:42: 07:57:57 | sidecar-deployment/2-enable-injection | starting test step 2-enable-injection logger.go:42: 07:57:57 | sidecar-deployment/2-enable-injection | running command: [kubectl annotate --overwrite deployment vertx-create-span-sidecar sidecar.jaegertracing.io/inject=true --namespace kuttl-test-enabling-liger] logger.go:42: 07:57:57 | sidecar-deployment/2-enable-injection | deployment.apps/vertx-create-span-sidecar annotated logger.go:42: 07:58:03 | sidecar-deployment/2-enable-injection | test step completed 2-enable-injection logger.go:42: 07:58:03 | sidecar-deployment/3-find-service | starting test step 3-find-service logger.go:42: 07:58:03 | sidecar-deployment/3-find-service | Job:kuttl-test-enabling-liger/00-find-service created logger.go:42: 07:58:16 | sidecar-deployment/3-find-service | test step completed 3-find-service logger.go:42: 07:58:16 | sidecar-deployment/4-other-instance | starting test step 4-other-instance logger.go:42: 07:58:16 | sidecar-deployment/4-other-instance | Jaeger:kuttl-test-enabling-liger/agent-as-sidecar2 created logger.go:42: 07:58:24 | sidecar-deployment/4-other-instance | test step completed 4-other-instance logger.go:42: 07:58:24 | sidecar-deployment/5-delete-first-instance | starting test step 5-delete-first-instance logger.go:42: 07:58:25 | sidecar-deployment/5-delete-first-instance | test step completed 5-delete-first-instance logger.go:42: 07:58:25 | sidecar-deployment/6-find-service | starting test step 6-find-service logger.go:42: 07:58:25 | sidecar-deployment/6-find-service | Job:kuttl-test-enabling-liger/01-find-service created logger.go:42: 07:58:45 | sidecar-deployment/6-find-service | test step completed 6-find-service logger.go:42: 07:58:45 | sidecar-deployment/7-disable-injection | starting test step 7-disable-injection logger.go:42: 07:58:45 | sidecar-deployment/7-disable-injection | running command: [kubectl annotate --overwrite deployment vertx-create-span-sidecar sidecar.jaegertracing.io/inject=false --namespace kuttl-test-enabling-liger] logger.go:42: 07:58:45 | sidecar-deployment/7-disable-injection | deployment.apps/vertx-create-span-sidecar annotated logger.go:42: 07:58:48 | sidecar-deployment/7-disable-injection | test step completed 7-disable-injection logger.go:42: 07:58:48 | sidecar-deployment | sidecar-deployment events from ns kuttl-test-enabling-liger: logger.go:42: 07:58:48 | sidecar-deployment | 2025-02-03 07:57:53 +0000 UTC Normal Pod agent-as-sidecar-689b547f8b-9kdls Binding Scheduled Successfully assigned kuttl-test-enabling-liger/agent-as-sidecar-689b547f8b-9kdls to ip-10-0-66-98.us-east-2.compute.internal default-scheduler logger.go:42: 07:58:48 | sidecar-deployment | 2025-02-03 07:57:53 +0000 UTC Normal Pod agent-as-sidecar-689b547f8b-9kdls AddedInterface Add eth0 [10.129.2.73/23] from ovn-kubernetes multus logger.go:42: 07:58:48 | sidecar-deployment | 2025-02-03 07:57:53 +0000 UTC Normal Pod agent-as-sidecar-689b547f8b-9kdls.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:fa9aaf9ae026228265ce7f10fb451d5c52a4e0269ce1fa9024cbefddf765ae8d" already present on machine kubelet logger.go:42: 07:58:48 | sidecar-deployment | 2025-02-03 07:57:53 +0000 UTC Normal Pod agent-as-sidecar-689b547f8b-9kdls.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:58:48 | sidecar-deployment | 2025-02-03 07:57:53 +0000 UTC Normal ReplicaSet.apps agent-as-sidecar-689b547f8b SuccessfulCreate Created pod: agent-as-sidecar-689b547f8b-9kdls replicaset-controller logger.go:42: 07:58:48 | sidecar-deployment | 2025-02-03 07:57:53 +0000 UTC Normal Deployment.apps agent-as-sidecar ScalingReplicaSet Scaled up replica set agent-as-sidecar-689b547f8b to 1 deployment-controller logger.go:42: 07:58:48 | sidecar-deployment | 2025-02-03 07:57:54 +0000 UTC Normal Pod agent-as-sidecar-689b547f8b-9kdls.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:58:48 | sidecar-deployment | 2025-02-03 07:57:55 +0000 UTC Normal Pod vertx-create-span-sidecar-797645c8fc-qf4mk Binding Scheduled Successfully assigned kuttl-test-enabling-liger/vertx-create-span-sidecar-797645c8fc-qf4mk to ip-10-0-6-172.us-east-2.compute.internal default-scheduler logger.go:42: 07:58:48 | sidecar-deployment | 2025-02-03 07:57:55 +0000 UTC Normal Pod vertx-create-span-sidecar-797645c8fc-qf4mk AddedInterface Add eth0 [10.131.0.61/23] from ovn-kubernetes multus logger.go:42: 07:58:48 | sidecar-deployment | 2025-02-03 07:57:55 +0000 UTC Normal Pod vertx-create-span-sidecar-797645c8fc-qf4mk.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:58:48 | sidecar-deployment | 2025-02-03 07:57:55 +0000 UTC Normal Pod vertx-create-span-sidecar-797645c8fc-qf4mk.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:58:48 | sidecar-deployment | 2025-02-03 07:57:55 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-797645c8fc SuccessfulCreate Created pod: vertx-create-span-sidecar-797645c8fc-qf4mk replicaset-controller logger.go:42: 07:58:48 | sidecar-deployment | 2025-02-03 07:57:55 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-797645c8fc to 1 deployment-controller logger.go:42: 07:58:48 | sidecar-deployment | 2025-02-03 07:57:56 +0000 UTC Normal Pod vertx-create-span-sidecar-797645c8fc-qf4mk.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:58:48 | sidecar-deployment | 2025-02-03 07:57:57 +0000 UTC Normal Pod vertx-create-span-sidecar-7bf6ffc9-7lmlj Binding Scheduled Successfully assigned kuttl-test-enabling-liger/vertx-create-span-sidecar-7bf6ffc9-7lmlj to ip-10-0-26-111.us-east-2.compute.internal default-scheduler logger.go:42: 07:58:48 | sidecar-deployment | 2025-02-03 07:57:57 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-7bf6ffc9 SuccessfulCreate Created pod: vertx-create-span-sidecar-7bf6ffc9-7lmlj replicaset-controller logger.go:42: 07:58:48 | sidecar-deployment | 2025-02-03 07:57:57 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-7bf6ffc9 to 1 deployment-controller logger.go:42: 07:58:48 | sidecar-deployment | 2025-02-03 07:57:58 +0000 UTC Normal Pod vertx-create-span-sidecar-7bf6ffc9-7lmlj AddedInterface Add eth0 [10.128.2.75/23] from ovn-kubernetes multus logger.go:42: 07:58:48 | sidecar-deployment | 2025-02-03 07:57:58 +0000 UTC Normal Pod vertx-create-span-sidecar-7bf6ffc9-7lmlj.spec.containers{vertx-create-span-sidecar} Pulling Pulling image "jaegertracing/vertx-create-span:operator-e2e-tests" kubelet logger.go:42: 07:58:48 | sidecar-deployment | 2025-02-03 07:58:01 +0000 UTC Normal Pod vertx-create-span-sidecar-7bf6ffc9-7lmlj.spec.containers{vertx-create-span-sidecar} Pulled Successfully pulled image "jaegertracing/vertx-create-span:operator-e2e-tests" in 3.382s (3.382s including waiting). Image size: 282912835 bytes. kubelet logger.go:42: 07:58:48 | sidecar-deployment | 2025-02-03 07:58:01 +0000 UTC Normal Pod vertx-create-span-sidecar-7bf6ffc9-7lmlj.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:58:48 | sidecar-deployment | 2025-02-03 07:58:01 +0000 UTC Normal Pod vertx-create-span-sidecar-7bf6ffc9-7lmlj.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:58:48 | sidecar-deployment | 2025-02-03 07:58:01 +0000 UTC Normal Pod vertx-create-span-sidecar-7bf6ffc9-7lmlj.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:65342c7d622f336741a8ae0b9dead79c7ecabc155d02cdd42b7b49ca36680e74" already present on machine kubelet logger.go:42: 07:58:48 | sidecar-deployment | 2025-02-03 07:58:01 +0000 UTC Normal Pod vertx-create-span-sidecar-7bf6ffc9-7lmlj.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:58:48 | sidecar-deployment | 2025-02-03 07:58:01 +0000 UTC Normal Pod vertx-create-span-sidecar-7bf6ffc9-7lmlj.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:58:48 | sidecar-deployment | 2025-02-03 07:58:03 +0000 UTC Normal Pod 00-find-service-7dgk7 Binding Scheduled Successfully assigned kuttl-test-enabling-liger/00-find-service-7dgk7 to ip-10-0-6-172.us-east-2.compute.internal default-scheduler logger.go:42: 07:58:48 | sidecar-deployment | 2025-02-03 07:58:03 +0000 UTC Normal Job.batch 00-find-service SuccessfulCreate Created pod: 00-find-service-7dgk7 job-controller logger.go:42: 07:58:48 | sidecar-deployment | 2025-02-03 07:58:03 +0000 UTC Warning Pod vertx-create-span-sidecar-797645c8fc-qf4mk.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.131.0.61:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:58:48 | sidecar-deployment | 2025-02-03 07:58:03 +0000 UTC Warning Pod vertx-create-span-sidecar-797645c8fc-qf4mk.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.131.0.61:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:58:48 | sidecar-deployment | 2025-02-03 07:58:04 +0000 UTC Normal Pod 00-find-service-7dgk7 AddedInterface Add eth0 [10.131.0.62/23] from ovn-kubernetes multus logger.go:42: 07:58:48 | sidecar-deployment | 2025-02-03 07:58:04 +0000 UTC Normal Pod 00-find-service-7dgk7.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:58:48 | sidecar-deployment | 2025-02-03 07:58:05 +0000 UTC Normal Pod 00-find-service-7dgk7.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 552ms (552ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:58:48 | sidecar-deployment | 2025-02-03 07:58:05 +0000 UTC Normal Pod 00-find-service-7dgk7.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:58:48 | sidecar-deployment | 2025-02-03 07:58:05 +0000 UTC Normal Pod 00-find-service-7dgk7.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:58:48 | sidecar-deployment | 2025-02-03 07:58:05 +0000 UTC Normal Pod vertx-create-span-sidecar-797645c8fc-qf4mk.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 07:58:48 | sidecar-deployment | 2025-02-03 07:58:06 +0000 UTC Warning Pod vertx-create-span-sidecar-797645c8fc-qf4mk.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.131.0.61:8080/": read tcp 10.131.0.2:41810->10.131.0.61:8080: read: connection reset by peer kubelet logger.go:42: 07:58:48 | sidecar-deployment | 2025-02-03 07:58:06 +0000 UTC Warning Pod vertx-create-span-sidecar-797645c8fc-qf4mk.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.131.0.61:8080/": dial tcp 10.131.0.61:8080: connect: connection refused kubelet logger.go:42: 07:58:48 | sidecar-deployment | 2025-02-03 07:58:09 +0000 UTC Warning Pod vertx-create-span-sidecar-7bf6ffc9-7lmlj.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.128.2.75:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:58:48 | sidecar-deployment | 2025-02-03 07:58:09 +0000 UTC Warning Pod vertx-create-span-sidecar-7bf6ffc9-7lmlj.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.2.75:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:58:48 | sidecar-deployment | 2025-02-03 07:58:11 +0000 UTC Normal Pod vertx-create-span-sidecar-7bf6ffc9-7lmlj.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 07:58:48 | sidecar-deployment | 2025-02-03 07:58:12 +0000 UTC Warning Pod vertx-create-span-sidecar-7bf6ffc9-7lmlj.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.2.75:8080/": read tcp 10.128.2.2:48382->10.128.2.75:8080: read: connection reset by peer kubelet logger.go:42: 07:58:48 | sidecar-deployment | 2025-02-03 07:58:12 +0000 UTC Warning Pod vertx-create-span-sidecar-7bf6ffc9-7lmlj.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.128.2.75:8080/": dial tcp 10.128.2.75:8080: connect: connection refused kubelet logger.go:42: 07:58:48 | sidecar-deployment | 2025-02-03 07:58:12 +0000 UTC Normal Pod vertx-create-span-sidecar-7bf6ffc9-7lmlj.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:58:48 | sidecar-deployment | 2025-02-03 07:58:16 +0000 UTC Normal Job.batch 00-find-service Completed Job completed job-controller logger.go:42: 07:58:48 | sidecar-deployment | 2025-02-03 07:58:17 +0000 UTC Warning Pod vertx-create-span-sidecar-797645c8fc-qf4mk.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.131.0.61:8080/": read tcp 10.131.0.2:53260->10.131.0.61:8080: read: connection reset by peer kubelet logger.go:42: 07:58:48 | sidecar-deployment | 2025-02-03 07:58:20 +0000 UTC Normal Pod agent-as-sidecar2-75b75bbd66-s8q2f Binding Scheduled Successfully assigned kuttl-test-enabling-liger/agent-as-sidecar2-75b75bbd66-s8q2f to ip-10-0-66-98.us-east-2.compute.internal default-scheduler logger.go:42: 07:58:48 | sidecar-deployment | 2025-02-03 07:58:20 +0000 UTC Warning Pod agent-as-sidecar2-75b75bbd66-s8q2f FailedMount MountVolume.SetUp failed for volume "agent-as-sidecar2-collector-tls-config-volume" : secret "agent-as-sidecar2-collector-headless-tls" not found kubelet logger.go:42: 07:58:48 | sidecar-deployment | 2025-02-03 07:58:20 +0000 UTC Normal ReplicaSet.apps agent-as-sidecar2-75b75bbd66 SuccessfulCreate Created pod: agent-as-sidecar2-75b75bbd66-s8q2f replicaset-controller logger.go:42: 07:58:48 | sidecar-deployment | 2025-02-03 07:58:20 +0000 UTC Normal Deployment.apps agent-as-sidecar2 ScalingReplicaSet Scaled up replica set agent-as-sidecar2-75b75bbd66 to 1 deployment-controller logger.go:42: 07:58:48 | sidecar-deployment | 2025-02-03 07:58:21 +0000 UTC Normal Pod agent-as-sidecar2-75b75bbd66-s8q2f AddedInterface Add eth0 [10.129.2.74/23] from ovn-kubernetes multus logger.go:42: 07:58:48 | sidecar-deployment | 2025-02-03 07:58:21 +0000 UTC Normal Pod agent-as-sidecar2-75b75bbd66-s8q2f.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:fa9aaf9ae026228265ce7f10fb451d5c52a4e0269ce1fa9024cbefddf765ae8d" already present on machine kubelet logger.go:42: 07:58:48 | sidecar-deployment | 2025-02-03 07:58:22 +0000 UTC Normal Pod agent-as-sidecar2-75b75bbd66-s8q2f.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:58:48 | sidecar-deployment | 2025-02-03 07:58:22 +0000 UTC Normal Pod agent-as-sidecar2-75b75bbd66-s8q2f.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:58:48 | sidecar-deployment | 2025-02-03 07:58:24 +0000 UTC Normal Pod agent-as-sidecar-689b547f8b-9kdls.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:58:48 | sidecar-deployment | 2025-02-03 07:58:25 +0000 UTC Normal Pod 01-find-service-fbdn2 Binding Scheduled Successfully assigned kuttl-test-enabling-liger/01-find-service-fbdn2 to ip-10-0-66-98.us-east-2.compute.internal default-scheduler logger.go:42: 07:58:48 | sidecar-deployment | 2025-02-03 07:58:25 +0000 UTC Normal Job.batch 01-find-service SuccessfulCreate Created pod: 01-find-service-fbdn2 job-controller logger.go:42: 07:58:48 | sidecar-deployment | 2025-02-03 07:58:26 +0000 UTC Normal Pod 01-find-service-fbdn2 AddedInterface Add eth0 [10.129.2.75/23] from ovn-kubernetes multus logger.go:42: 07:58:48 | sidecar-deployment | 2025-02-03 07:58:26 +0000 UTC Normal Pod 01-find-service-fbdn2.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:58:48 | sidecar-deployment | 2025-02-03 07:58:27 +0000 UTC Normal Pod 01-find-service-fbdn2.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 753ms (753ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:58:48 | sidecar-deployment | 2025-02-03 07:58:27 +0000 UTC Normal Pod 01-find-service-fbdn2.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:58:48 | sidecar-deployment | 2025-02-03 07:58:27 +0000 UTC Normal Pod 01-find-service-fbdn2.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:58:48 | sidecar-deployment | 2025-02-03 07:58:27 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-797645c8fc SuccessfulDelete Deleted pod: vertx-create-span-sidecar-797645c8fc-qf4mk replicaset-controller logger.go:42: 07:58:48 | sidecar-deployment | 2025-02-03 07:58:27 +0000 UTC Normal Pod vertx-create-span-sidecar-b4f5d59cc-vlw9l Binding Scheduled Successfully assigned kuttl-test-enabling-liger/vertx-create-span-sidecar-b4f5d59cc-vlw9l to ip-10-0-6-172.us-east-2.compute.internal default-scheduler logger.go:42: 07:58:48 | sidecar-deployment | 2025-02-03 07:58:27 +0000 UTC Normal Pod vertx-create-span-sidecar-b4f5d59cc-vlw9l AddedInterface Add eth0 [10.131.0.63/23] from ovn-kubernetes multus logger.go:42: 07:58:48 | sidecar-deployment | 2025-02-03 07:58:27 +0000 UTC Normal Pod vertx-create-span-sidecar-b4f5d59cc-vlw9l.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:58:48 | sidecar-deployment | 2025-02-03 07:58:27 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-b4f5d59cc SuccessfulCreate Created pod: vertx-create-span-sidecar-b4f5d59cc-vlw9l replicaset-controller logger.go:42: 07:58:48 | sidecar-deployment | 2025-02-03 07:58:27 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled down replica set vertx-create-span-sidecar-797645c8fc to 0 from 1 deployment-controller logger.go:42: 07:58:48 | sidecar-deployment | 2025-02-03 07:58:27 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-b4f5d59cc to 1 from 0 deployment-controller logger.go:42: 07:58:48 | sidecar-deployment | 2025-02-03 07:58:28 +0000 UTC Normal Pod vertx-create-span-sidecar-b4f5d59cc-vlw9l.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:58:48 | sidecar-deployment | 2025-02-03 07:58:28 +0000 UTC Normal Pod vertx-create-span-sidecar-b4f5d59cc-vlw9l.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:58:48 | sidecar-deployment | 2025-02-03 07:58:28 +0000 UTC Normal Pod vertx-create-span-sidecar-b4f5d59cc-vlw9l.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:65342c7d622f336741a8ae0b9dead79c7ecabc155d02cdd42b7b49ca36680e74" already present on machine kubelet logger.go:42: 07:58:48 | sidecar-deployment | 2025-02-03 07:58:28 +0000 UTC Normal Pod vertx-create-span-sidecar-b4f5d59cc-vlw9l.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:58:48 | sidecar-deployment | 2025-02-03 07:58:28 +0000 UTC Normal Pod vertx-create-span-sidecar-b4f5d59cc-vlw9l.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:58:48 | sidecar-deployment | 2025-02-03 07:58:36 +0000 UTC Warning Pod vertx-create-span-sidecar-b4f5d59cc-vlw9l.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.131.0.63:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:58:48 | sidecar-deployment | 2025-02-03 07:58:36 +0000 UTC Warning Pod vertx-create-span-sidecar-b4f5d59cc-vlw9l.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.131.0.63:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:58:48 | sidecar-deployment | 2025-02-03 07:58:38 +0000 UTC Normal Pod vertx-create-span-sidecar-b4f5d59cc-vlw9l.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 07:58:48 | sidecar-deployment | 2025-02-03 07:58:39 +0000 UTC Warning Pod vertx-create-span-sidecar-b4f5d59cc-vlw9l.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.131.0.63:8080/": read tcp 10.131.0.2:59126->10.131.0.63:8080: read: connection reset by peer kubelet logger.go:42: 07:58:48 | sidecar-deployment | 2025-02-03 07:58:39 +0000 UTC Warning Pod vertx-create-span-sidecar-b4f5d59cc-vlw9l.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.131.0.63:8080/": dial tcp 10.131.0.63:8080: connect: connection refused kubelet logger.go:42: 07:58:48 | sidecar-deployment | 2025-02-03 07:58:45 +0000 UTC Normal Job.batch 01-find-service Completed Job completed job-controller logger.go:42: 07:58:48 | sidecar-deployment | 2025-02-03 07:58:45 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-7bf6ffc9 SuccessfulDelete Deleted pod: vertx-create-span-sidecar-7bf6ffc9-7lmlj replicaset-controller logger.go:42: 07:58:48 | sidecar-deployment | 2025-02-03 07:58:45 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled down replica set vertx-create-span-sidecar-7bf6ffc9 to 0 from 1 deployment-controller logger.go:42: 07:58:48 | sidecar-deployment | 2025-02-03 07:58:46 +0000 UTC Normal Pod vertx-create-span-sidecar-b7b5c5dcd-6ghlf Binding Scheduled Successfully assigned kuttl-test-enabling-liger/vertx-create-span-sidecar-b7b5c5dcd-6ghlf to ip-10-0-66-98.us-east-2.compute.internal default-scheduler logger.go:42: 07:58:48 | sidecar-deployment | 2025-02-03 07:58:46 +0000 UTC Normal Pod vertx-create-span-sidecar-b7b5c5dcd-6ghlf AddedInterface Add eth0 [10.129.2.76/23] from ovn-kubernetes multus logger.go:42: 07:58:48 | sidecar-deployment | 2025-02-03 07:58:46 +0000 UTC Normal Pod vertx-create-span-sidecar-b7b5c5dcd-6ghlf.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:58:48 | sidecar-deployment | 2025-02-03 07:58:46 +0000 UTC Normal Pod vertx-create-span-sidecar-b7b5c5dcd-6ghlf.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:58:48 | sidecar-deployment | 2025-02-03 07:58:46 +0000 UTC Normal Pod vertx-create-span-sidecar-b7b5c5dcd-6ghlf.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:58:48 | sidecar-deployment | 2025-02-03 07:58:46 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-b7b5c5dcd SuccessfulCreate Created pod: vertx-create-span-sidecar-b7b5c5dcd-6ghlf replicaset-controller logger.go:42: 07:58:48 | sidecar-deployment | 2025-02-03 07:58:46 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-b7b5c5dcd to 1 from 0 deployment-controller logger.go:42: 07:58:48 | sidecar-deployment | Deleting namespace: kuttl-test-enabling-liger === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (155.45s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (6.15s) --- PASS: kuttl/harness/sidecar-namespace (64.41s) --- PASS: kuttl/harness/sidecar-skip-webhook (18.60s) --- PASS: kuttl/harness/sidecar-deployment (66.23s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name sidecar --report --output /logs/artifacts/sidecar.xml ./artifacts/kuttl-report.xml time="2025-02-03T07:58:55Z" level=debug msg="Setting a new name for the test suites" time="2025-02-03T07:58:55Z" level=debug msg="Removing 'artifacts' TestCase" time="2025-02-03T07:58:55Z" level=debug msg="normalizing test case names" time="2025-02-03T07:58:55Z" level=debug msg="sidecar/artifacts -> sidecar_artifacts" time="2025-02-03T07:58:55Z" level=debug msg="sidecar/sidecar-namespace -> sidecar_sidecar_namespace" time="2025-02-03T07:58:55Z" level=debug msg="sidecar/sidecar-skip-webhook -> sidecar_sidecar_skip_webhook" time="2025-02-03T07:58:55Z" level=debug msg="sidecar/sidecar-deployment -> sidecar_sidecar_deployment" +------------------------------+--------+ | NAME | RESULT | +------------------------------+--------+ | sidecar_artifacts | passed | | sidecar_sidecar_namespace | passed | | sidecar_sidecar_skip_webhook | passed | | sidecar_sidecar_deployment | passed | +------------------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh streaming false true + '[' 3 -ne 3 ']' + test_suite_name=streaming + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/streaming.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-streaming make[2]: Entering directory '/tmp/jaeger-tests' >>>> Elasticsearch image not loaded because SKIP_ES_EXTERNAL is true KAFKA_VERSION=3.6.0 \ SKIP_KAFKA=false \ SKIP_ES_EXTERNAL=true \ ./tests/e2e/streaming/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.18.0-0.nightly-2025-02-01-185117 True False 50m Cluster version is 4.18.0-0.nightly-2025-02-01-185117' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.18.0-0.nightly-2025-02-01-185117 True False 50m Cluster version is 4.18.0-0.nightly-2025-02-01-185117' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/streaming/render.sh ++ export SUITE_DIR=./tests/e2e/streaming ++ SUITE_DIR=./tests/e2e/streaming ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/streaming ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + is_secured=false + '[' true = true ']' + is_secured=true + '[' false = true ']' + start_test streaming-simple + '[' 1 -ne 1 ']' + test_name=streaming-simple + echo =========================================================================== =========================================================================== + info 'Rendering files for test streaming-simple' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test streaming-simple\e[0m' Rendering files for test streaming-simple + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/streaming/_build + '[' _build '!=' _build ']' + mkdir -p streaming-simple + cd streaming-simple + render_install_kafka my-cluster 00 + '[' 2 -ne 2 ']' + cluster_name=my-cluster + test_step=00 + CLUSTER_NAME=my-cluster + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/kafka-install.yaml.template -o ./00-install.yaml + render_assert_kafka false my-cluster 00 + '[' 3 -ne 3 ']' + autoprovisioned=false + cluster_name=my-cluster + test_step=00 + '[' false = true ']' + '[' false = true ']' + '[' false = false ']' + replicas=1 + CLUSTER_NAME=my-cluster + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-zookeeper-cluster.yaml.template -o ./00-assert.yaml ++ expr 00 + 1 + CLUSTER_NAME=my-cluster + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-kafka-cluster.yaml.template -o ./01-assert.yaml ++ expr 00 + 2 + CLUSTER_NAME=my-cluster + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-entity-operator.yaml.template -o ./02-assert.yaml + render_install_elasticsearch upstream 03 + '[' 2 -ne 2 ']' + deploy_mode=upstream + test_step=03 + '[' upstream = upstream ']' + '[' true = true ']' + template=/tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template + /tmp/jaeger-tests/bin/yq eval -s '"elasticsearch_" + $index' /tmp/jaeger-tests/tests/elasticsearch.yml + /tmp/jaeger-tests/bin/yq eval -i '.spec.template.spec.serviceAccountName="deploy-elasticsearch"' ./elasticsearch_0.yml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template -o ./03-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/elasticsearch-assert.yaml.template -o ./03-assert.yaml + JAEGER_NAME=simple-streaming + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/streaming-jaeger-assert.yaml.template -o ./04-assert.yaml + render_smoke_test simple-streaming true 05 + '[' 3 -ne 3 ']' + jaeger=simple-streaming + is_secured=true + test_step=05 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simple-streaming-query:443 + JAEGER_QUERY_ENDPOINT=https://simple-streaming-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simple-streaming-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simple-streaming-collector-headless:14268 + export JAEGER_NAME=simple-streaming + JAEGER_NAME=simple-streaming + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./05-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./05-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + '[' false = true ']' + start_test streaming-with-tls + '[' 1 -ne 1 ']' + test_name=streaming-with-tls + echo =========================================================================== =========================================================================== + info 'Rendering files for test streaming-with-tls' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test streaming-with-tls\e[0m' Rendering files for test streaming-with-tls + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/streaming/_build/streaming-simple + '[' streaming-simple '!=' _build ']' + cd .. + mkdir -p streaming-with-tls + cd streaming-with-tls + render_install_kafka my-cluster 00 + '[' 2 -ne 2 ']' + cluster_name=my-cluster + test_step=00 + CLUSTER_NAME=my-cluster + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/kafka-install.yaml.template -o ./00-install.yaml + render_assert_kafka false my-cluster 00 + '[' 3 -ne 3 ']' + autoprovisioned=false + cluster_name=my-cluster + test_step=00 + '[' false = true ']' + '[' false = true ']' + '[' false = false ']' + replicas=1 + CLUSTER_NAME=my-cluster + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-zookeeper-cluster.yaml.template -o ./00-assert.yaml ++ expr 00 + 1 + CLUSTER_NAME=my-cluster + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-kafka-cluster.yaml.template -o ./01-assert.yaml ++ expr 00 + 2 + CLUSTER_NAME=my-cluster + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-entity-operator.yaml.template -o ./02-assert.yaml + render_install_elasticsearch upstream 03 + '[' 2 -ne 2 ']' + deploy_mode=upstream + test_step=03 + '[' upstream = upstream ']' + '[' true = true ']' + template=/tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template + /tmp/jaeger-tests/bin/yq eval -s '"elasticsearch_" + $index' /tmp/jaeger-tests/tests/elasticsearch.yml + /tmp/jaeger-tests/bin/yq eval -i '.spec.template.spec.serviceAccountName="deploy-elasticsearch"' ./elasticsearch_0.yml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template -o ./03-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/elasticsearch-assert.yaml.template -o ./03-assert.yaml + render_smoke_test tls-streaming true 05 + '[' 3 -ne 3 ']' + jaeger=tls-streaming + is_secured=true + test_step=05 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://tls-streaming-query:443 + JAEGER_QUERY_ENDPOINT=https://tls-streaming-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://tls-streaming-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://tls-streaming-collector-headless:14268 + export JAEGER_NAME=tls-streaming + JAEGER_NAME=tls-streaming + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./05-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./05-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + '[' false = true ']' + start_test streaming-with-autoprovisioning-autoscale + '[' 1 -ne 1 ']' + test_name=streaming-with-autoprovisioning-autoscale + echo =========================================================================== =========================================================================== + info 'Rendering files for test streaming-with-autoprovisioning-autoscale' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test streaming-with-autoprovisioning-autoscale\e[0m' Rendering files for test streaming-with-autoprovisioning-autoscale + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/streaming/_build/streaming-with-tls + '[' streaming-with-tls '!=' _build ']' + cd .. + mkdir -p streaming-with-autoprovisioning-autoscale + cd streaming-with-autoprovisioning-autoscale + '[' true = true ']' + rm ./00-install.yaml ./00-assert.yaml + render_install_elasticsearch upstream 01 + '[' 2 -ne 2 ']' + deploy_mode=upstream + test_step=01 + '[' upstream = upstream ']' + '[' true = true ']' + template=/tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template + /tmp/jaeger-tests/bin/yq eval -s '"elasticsearch_" + $index' /tmp/jaeger-tests/tests/elasticsearch.yml + /tmp/jaeger-tests/bin/yq eval -i '.spec.template.spec.serviceAccountName="deploy-elasticsearch"' ./elasticsearch_0.yml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/elasticsearch-assert.yaml.template -o ./01-assert.yaml + jaeger_name=auto-provisioned + /tmp/jaeger-tests/bin/yq e -i '.spec.ingester.resources.requests.memory="20Mi"' ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.ingester.resources.requests.memory="500m"' ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.ingester.autoscale=true ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.ingester.minReplicas=1 ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.ingester.maxReplicas=2 ./02-install.yaml + render_assert_kafka true auto-provisioned 03 + '[' 3 -ne 3 ']' + autoprovisioned=true + cluster_name=auto-provisioned + test_step=03 + '[' true = true ']' + is_kafka_minimal_enabled + namespaces=(observability openshift-operators openshift-distributed-tracing) + for i in "${namespaces[@]}" ++ kubectl get pods -n observability -l name=jaeger-operator -o yaml ++ /tmp/jaeger-tests/bin/yq e '.items[0].spec.containers[0].env[] | select(.name=="KAFKA-PROVISIONING-MINIMAL").value' + enabled= + '[' '' == true ']' + for i in "${namespaces[@]}" ++ kubectl get pods -n openshift-operators -l name=jaeger-operator -o yaml ++ /tmp/jaeger-tests/bin/yq e '.items[0].spec.containers[0].env[] | select(.name=="KAFKA-PROVISIONING-MINIMAL").value' + enabled= + '[' '' == true ']' + for i in "${namespaces[@]}" ++ kubectl get pods -n openshift-distributed-tracing -l name=jaeger-operator -o yaml ++ /tmp/jaeger-tests/bin/yq e '.items[0].spec.containers[0].env[] | select(.name=="KAFKA-PROVISIONING-MINIMAL").value' + enabled=true + '[' true == true ']' + return 0 + replicas=1 + CLUSTER_NAME=auto-provisioned + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-zookeeper-cluster.yaml.template -o ./03-assert.yaml ++ expr 03 + 1 + CLUSTER_NAME=auto-provisioned + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-kafka-cluster.yaml.template -o ./04-assert.yaml ++ expr 03 + 2 + CLUSTER_NAME=auto-provisioned + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-entity-operator.yaml.template -o ./05-assert.yaml + version_lt 1.30 1.23 ++ echo 1.30 1.23 ++ tr ' ' '\n' ++ sort -rV ++ head -n 1 + test 1.30 '!=' 1.30 + rm ./08-assert.yaml + skip_test streaming-with-tls 'This test is flaky in Prow CI' + '[' 2 -ne 2 ']' + test_name=streaming-with-tls + message='This test is flaky in Prow CI' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/streaming/_build/streaming-with-autoprovisioning-autoscale + '[' streaming-with-autoprovisioning-autoscale '!=' _build ']' + cd .. + rm -rf streaming-with-tls + warning 'streaming-with-tls: This test is flaky in Prow CI' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: streaming-with-tls: This test is flaky in Prow CI\e[0m' WAR: streaming-with-tls: This test is flaky in Prow CI + skip_test streaming-simple 'This test is flaky in Prow CI' + '[' 2 -ne 2 ']' + test_name=streaming-simple + message='This test is flaky in Prow CI' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/streaming/_build + '[' _build '!=' _build ']' + rm -rf streaming-simple + warning 'streaming-simple: This test is flaky in Prow CI' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: streaming-simple: This test is flaky in Prow CI\e[0m' WAR: streaming-simple: This test is flaky in Prow CI make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running streaming E2E tests' Running streaming E2E tests + cd tests/e2e/streaming/_build + set +e + KUBECONFIG=/tmp/kubeconfig-2531963025 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-d102g9k0-f1a62.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 2 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/streaming-with-autoprovisioning-autoscale === PAUSE kuttl/harness/streaming-with-autoprovisioning-autoscale === CONT kuttl/harness/artifacts logger.go:42: 07:59:07 | artifacts | Creating namespace: kuttl-test-upright-quagga logger.go:42: 07:59:07 | artifacts | artifacts events from ns kuttl-test-upright-quagga: logger.go:42: 07:59:07 | artifacts | Deleting namespace: kuttl-test-upright-quagga === CONT kuttl/harness/streaming-with-autoprovisioning-autoscale logger.go:42: 07:59:13 | streaming-with-autoprovisioning-autoscale | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:59:13 | streaming-with-autoprovisioning-autoscale | Ignoring elasticsearch_0.yml as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:59:13 | streaming-with-autoprovisioning-autoscale | Ignoring elasticsearch_1.yml as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:59:13 | streaming-with-autoprovisioning-autoscale | Creating namespace: kuttl-test-cunning-lamb logger.go:42: 07:59:13 | streaming-with-autoprovisioning-autoscale/1-install | starting test step 1-install logger.go:42: 07:59:13 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c oc create sa deploy-elasticsearch -n $NAMESPACE 2>&1 | grep -v "already exists" || true] logger.go:42: 07:59:13 | streaming-with-autoprovisioning-autoscale/1-install | serviceaccount/deploy-elasticsearch created logger.go:42: 07:59:13 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c oc adm policy add-scc-to-user privileged -z deploy-elasticsearch -n $NAMESPACE 2>&1 | grep -v "already exists" || true] logger.go:42: 07:59:14 | streaming-with-autoprovisioning-autoscale/1-install | clusterrole.rbac.authorization.k8s.io/system:openshift:scc:privileged added: "deploy-elasticsearch" logger.go:42: 07:59:14 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c sleep 6] logger.go:42: 07:59:20 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c kubectl apply -f elasticsearch_0.yml -n $NAMESPACE] logger.go:42: 07:59:20 | streaming-with-autoprovisioning-autoscale/1-install | statefulset.apps/elasticsearch created logger.go:42: 07:59:20 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c sleep 3] logger.go:42: 07:59:23 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c kubectl apply -f elasticsearch_1.yml -n $NAMESPACE] logger.go:42: 07:59:23 | streaming-with-autoprovisioning-autoscale/1-install | service/elasticsearch created logger.go:42: 07:59:41 | streaming-with-autoprovisioning-autoscale/1-install | test step completed 1-install logger.go:42: 07:59:41 | streaming-with-autoprovisioning-autoscale/2-install | starting test step 2-install logger.go:42: 07:59:41 | streaming-with-autoprovisioning-autoscale/2-install | Jaeger:kuttl-test-cunning-lamb/auto-provisioned created logger.go:42: 07:59:41 | streaming-with-autoprovisioning-autoscale/2-install | test step completed 2-install logger.go:42: 07:59:41 | streaming-with-autoprovisioning-autoscale/3- | starting test step 3- logger.go:42: 08:00:41 | streaming-with-autoprovisioning-autoscale/3- | test step completed 3- logger.go:42: 08:00:41 | streaming-with-autoprovisioning-autoscale/4- | starting test step 4- logger.go:42: 08:01:12 | streaming-with-autoprovisioning-autoscale/4- | test step completed 4- logger.go:42: 08:01:12 | streaming-with-autoprovisioning-autoscale/5- | starting test step 5- logger.go:42: 08:01:35 | streaming-with-autoprovisioning-autoscale/5- | test step completed 5- logger.go:42: 08:01:35 | streaming-with-autoprovisioning-autoscale/6- | starting test step 6- logger.go:42: 08:01:49 | streaming-with-autoprovisioning-autoscale/6- | test step completed 6- logger.go:42: 08:01:49 | streaming-with-autoprovisioning-autoscale/7- | starting test step 7- logger.go:42: 08:01:49 | streaming-with-autoprovisioning-autoscale/7- | test step completed 7- logger.go:42: 08:01:49 | streaming-with-autoprovisioning-autoscale | streaming-with-autoprovisioning-autoscale events from ns kuttl-test-cunning-lamb: logger.go:42: 08:01:49 | streaming-with-autoprovisioning-autoscale | 2025-02-03 07:59:20 +0000 UTC Normal Pod elasticsearch-0 Binding Scheduled Successfully assigned kuttl-test-cunning-lamb/elasticsearch-0 to ip-10-0-66-98.us-east-2.compute.internal default-scheduler logger.go:42: 08:01:49 | streaming-with-autoprovisioning-autoscale | 2025-02-03 07:59:20 +0000 UTC Normal StatefulSet.apps elasticsearch SuccessfulCreate create Pod elasticsearch-0 in StatefulSet elasticsearch successful statefulset-controller logger.go:42: 08:01:49 | streaming-with-autoprovisioning-autoscale | 2025-02-03 07:59:21 +0000 UTC Normal Pod elasticsearch-0 AddedInterface Add eth0 [10.129.2.77/23] from ovn-kubernetes multus logger.go:42: 08:01:49 | streaming-with-autoprovisioning-autoscale | 2025-02-03 07:59:21 +0000 UTC Normal Pod elasticsearch-0.spec.containers{elasticsearch} Pulling Pulling image "docker.elastic.co/elasticsearch/elasticsearch-oss:6.8.6" kubelet logger.go:42: 08:01:49 | streaming-with-autoprovisioning-autoscale | 2025-02-03 07:59:28 +0000 UTC Normal Pod elasticsearch-0.spec.containers{elasticsearch} Pulled Successfully pulled image "docker.elastic.co/elasticsearch/elasticsearch-oss:6.8.6" in 7.86s (7.86s including waiting). Image size: 758467647 bytes. kubelet logger.go:42: 08:01:49 | streaming-with-autoprovisioning-autoscale | 2025-02-03 07:59:29 +0000 UTC Normal Pod elasticsearch-0.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 08:01:49 | streaming-with-autoprovisioning-autoscale | 2025-02-03 07:59:29 +0000 UTC Normal Pod elasticsearch-0.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 08:01:49 | streaming-with-autoprovisioning-autoscale | 2025-02-03 07:59:35 +0000 UTC Warning Pod elasticsearch-0.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Get "http://10.129.2.77:9200/": dial tcp 10.129.2.77:9200: connect: connection refused kubelet logger.go:42: 08:01:49 | streaming-with-autoprovisioning-autoscale | 2025-02-03 07:59:50 +0000 UTC Normal PodDisruptionBudget.policy auto-provisioned-zookeeper NoPods No matching pods found controllermanager logger.go:42: 08:01:49 | streaming-with-autoprovisioning-autoscale | 2025-02-03 07:59:50 +0000 UTC Normal PersistentVolumeClaim data-auto-provisioned-zookeeper-0 WaitForFirstConsumer waiting for first consumer to be created before binding persistentvolume-controller logger.go:42: 08:01:49 | streaming-with-autoprovisioning-autoscale | 2025-02-03 07:59:50 +0000 UTC Normal PersistentVolumeClaim data-auto-provisioned-zookeeper-0 Provisioning External provisioner is provisioning volume for claim "kuttl-test-cunning-lamb/data-auto-provisioned-zookeeper-0" ebs.csi.aws.com_aws-ebs-csi-driver-controller-7786bff46b-tnfgt_16799f7a-1018-4495-ab6f-76e3d33221ca logger.go:42: 08:01:49 | streaming-with-autoprovisioning-autoscale | 2025-02-03 07:59:50 +0000 UTC Normal PersistentVolumeClaim data-auto-provisioned-zookeeper-0 ExternalProvisioning Waiting for a volume to be created either by the external provisioner 'ebs.csi.aws.com' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered. persistentvolume-controller logger.go:42: 08:01:49 | streaming-with-autoprovisioning-autoscale | 2025-02-03 07:59:53 +0000 UTC Normal Pod auto-provisioned-zookeeper-0 Binding Scheduled Successfully assigned kuttl-test-cunning-lamb/auto-provisioned-zookeeper-0 to ip-10-0-6-172.us-east-2.compute.internal default-scheduler logger.go:42: 08:01:49 | streaming-with-autoprovisioning-autoscale | 2025-02-03 07:59:53 +0000 UTC Normal PersistentVolumeClaim data-auto-provisioned-zookeeper-0 ProvisioningSucceeded Successfully provisioned volume pvc-b518acb5-a31f-4626-89bf-ea02b3df0199 ebs.csi.aws.com_aws-ebs-csi-driver-controller-7786bff46b-tnfgt_16799f7a-1018-4495-ab6f-76e3d33221ca logger.go:42: 08:01:49 | streaming-with-autoprovisioning-autoscale | 2025-02-03 07:59:57 +0000 UTC Normal Pod auto-provisioned-zookeeper-0 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-b518acb5-a31f-4626-89bf-ea02b3df0199" attachdetach-controller logger.go:42: 08:01:49 | streaming-with-autoprovisioning-autoscale | 2025-02-03 08:00:01 +0000 UTC Normal Pod auto-provisioned-zookeeper-0 AddedInterface Add eth0 [10.131.0.64/23] from ovn-kubernetes multus logger.go:42: 08:01:49 | streaming-with-autoprovisioning-autoscale | 2025-02-03 08:00:01 +0000 UTC Normal Pod auto-provisioned-zookeeper-0.spec.containers{zookeeper} Pulling Pulling image "registry.redhat.io/amq-streams/kafka-38-rhel9@sha256:c9a60d03827466ee37dd4aff6803eda2c1d65cb2508cb57d13e1d47f982e5a20" kubelet logger.go:42: 08:01:49 | streaming-with-autoprovisioning-autoscale | 2025-02-03 08:00:15 +0000 UTC Normal Pod auto-provisioned-zookeeper-0.spec.containers{zookeeper} Pulled Successfully pulled image "registry.redhat.io/amq-streams/kafka-38-rhel9@sha256:c9a60d03827466ee37dd4aff6803eda2c1d65cb2508cb57d13e1d47f982e5a20" in 14.414s (14.414s including waiting). Image size: 616791996 bytes. kubelet logger.go:42: 08:01:49 | streaming-with-autoprovisioning-autoscale | 2025-02-03 08:00:16 +0000 UTC Normal Pod auto-provisioned-zookeeper-0.spec.containers{zookeeper} Created Created container zookeeper kubelet logger.go:42: 08:01:49 | streaming-with-autoprovisioning-autoscale | 2025-02-03 08:00:16 +0000 UTC Normal Pod auto-provisioned-zookeeper-0.spec.containers{zookeeper} Started Started container zookeeper kubelet logger.go:42: 08:01:49 | streaming-with-autoprovisioning-autoscale | 2025-02-03 08:00:42 +0000 UTC Normal PersistentVolumeClaim data-0-auto-provisioned-kafka-0 WaitForFirstConsumer waiting for first consumer to be created before binding persistentvolume-controller logger.go:42: 08:01:49 | streaming-with-autoprovisioning-autoscale | 2025-02-03 08:00:43 +0000 UTC Normal PodDisruptionBudget.policy auto-provisioned-kafka NoPods No matching pods found controllermanager logger.go:42: 08:01:49 | streaming-with-autoprovisioning-autoscale | 2025-02-03 08:00:43 +0000 UTC Normal PersistentVolumeClaim data-0-auto-provisioned-kafka-0 ExternalProvisioning Waiting for a volume to be created either by the external provisioner 'ebs.csi.aws.com' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered. persistentvolume-controller logger.go:42: 08:01:49 | streaming-with-autoprovisioning-autoscale | 2025-02-03 08:00:43 +0000 UTC Normal PersistentVolumeClaim data-0-auto-provisioned-kafka-0 Provisioning External provisioner is provisioning volume for claim "kuttl-test-cunning-lamb/data-0-auto-provisioned-kafka-0" ebs.csi.aws.com_aws-ebs-csi-driver-controller-7786bff46b-tnfgt_16799f7a-1018-4495-ab6f-76e3d33221ca logger.go:42: 08:01:49 | streaming-with-autoprovisioning-autoscale | 2025-02-03 08:00:45 +0000 UTC Normal PersistentVolumeClaim data-0-auto-provisioned-kafka-0 ProvisioningSucceeded Successfully provisioned volume pvc-1ae8e667-cec2-49ec-8b26-ef7949ab003d ebs.csi.aws.com_aws-ebs-csi-driver-controller-7786bff46b-tnfgt_16799f7a-1018-4495-ab6f-76e3d33221ca logger.go:42: 08:01:49 | streaming-with-autoprovisioning-autoscale | 2025-02-03 08:00:46 +0000 UTC Normal Pod auto-provisioned-kafka-0 Binding Scheduled Successfully assigned kuttl-test-cunning-lamb/auto-provisioned-kafka-0 to ip-10-0-6-172.us-east-2.compute.internal default-scheduler logger.go:42: 08:01:49 | streaming-with-autoprovisioning-autoscale | 2025-02-03 08:00:50 +0000 UTC Normal Pod auto-provisioned-kafka-0 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-1ae8e667-cec2-49ec-8b26-ef7949ab003d" attachdetach-controller logger.go:42: 08:01:49 | streaming-with-autoprovisioning-autoscale | 2025-02-03 08:00:52 +0000 UTC Normal Pod auto-provisioned-kafka-0 AddedInterface Add eth0 [10.131.0.66/23] from ovn-kubernetes multus logger.go:42: 08:01:49 | streaming-with-autoprovisioning-autoscale | 2025-02-03 08:00:52 +0000 UTC Normal Pod auto-provisioned-kafka-0.spec.containers{kafka} Pulled Container image "registry.redhat.io/amq-streams/kafka-38-rhel9@sha256:c9a60d03827466ee37dd4aff6803eda2c1d65cb2508cb57d13e1d47f982e5a20" already present on machine kubelet logger.go:42: 08:01:49 | streaming-with-autoprovisioning-autoscale | 2025-02-03 08:00:52 +0000 UTC Normal Pod auto-provisioned-kafka-0.spec.containers{kafka} Created Created container kafka kubelet logger.go:42: 08:01:49 | streaming-with-autoprovisioning-autoscale | 2025-02-03 08:00:52 +0000 UTC Normal Pod auto-provisioned-kafka-0.spec.containers{kafka} Started Started container kafka kubelet logger.go:42: 08:01:49 | streaming-with-autoprovisioning-autoscale | 2025-02-03 08:01:14 +0000 UTC Normal Pod auto-provisioned-entity-operator-85cfffbc4b-dgm6x Binding Scheduled Successfully assigned kuttl-test-cunning-lamb/auto-provisioned-entity-operator-85cfffbc4b-dgm6x to ip-10-0-26-111.us-east-2.compute.internal default-scheduler logger.go:42: 08:01:49 | streaming-with-autoprovisioning-autoscale | 2025-02-03 08:01:14 +0000 UTC Normal ReplicaSet.apps auto-provisioned-entity-operator-85cfffbc4b SuccessfulCreate Created pod: auto-provisioned-entity-operator-85cfffbc4b-dgm6x replicaset-controller logger.go:42: 08:01:49 | streaming-with-autoprovisioning-autoscale | 2025-02-03 08:01:14 +0000 UTC Normal Deployment.apps auto-provisioned-entity-operator ScalingReplicaSet Scaled up replica set auto-provisioned-entity-operator-85cfffbc4b to 1 deployment-controller logger.go:42: 08:01:49 | streaming-with-autoprovisioning-autoscale | 2025-02-03 08:01:15 +0000 UTC Normal Pod auto-provisioned-entity-operator-85cfffbc4b-dgm6x AddedInterface Add eth0 [10.128.2.76/23] from ovn-kubernetes multus logger.go:42: 08:01:49 | streaming-with-autoprovisioning-autoscale | 2025-02-03 08:01:15 +0000 UTC Normal Pod auto-provisioned-entity-operator-85cfffbc4b-dgm6x.spec.containers{topic-operator} Pulled Container image "registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:07a9540a8b906c6d52e6a9684cfb838aadf9849cc24b1d80218ea3ad5545cb5a" already present on machine kubelet logger.go:42: 08:01:49 | streaming-with-autoprovisioning-autoscale | 2025-02-03 08:01:15 +0000 UTC Normal Pod auto-provisioned-entity-operator-85cfffbc4b-dgm6x.spec.containers{topic-operator} Created Created container topic-operator kubelet logger.go:42: 08:01:49 | streaming-with-autoprovisioning-autoscale | 2025-02-03 08:01:15 +0000 UTC Normal Pod auto-provisioned-entity-operator-85cfffbc4b-dgm6x.spec.containers{topic-operator} Started Started container topic-operator kubelet logger.go:42: 08:01:49 | streaming-with-autoprovisioning-autoscale | 2025-02-03 08:01:15 +0000 UTC Normal Pod auto-provisioned-entity-operator-85cfffbc4b-dgm6x.spec.containers{user-operator} Pulled Container image "registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:07a9540a8b906c6d52e6a9684cfb838aadf9849cc24b1d80218ea3ad5545cb5a" already present on machine kubelet logger.go:42: 08:01:49 | streaming-with-autoprovisioning-autoscale | 2025-02-03 08:01:15 +0000 UTC Normal Pod auto-provisioned-entity-operator-85cfffbc4b-dgm6x.spec.containers{user-operator} Created Created container user-operator kubelet logger.go:42: 08:01:49 | streaming-with-autoprovisioning-autoscale | 2025-02-03 08:01:15 +0000 UTC Normal Pod auto-provisioned-entity-operator-85cfffbc4b-dgm6x.spec.containers{user-operator} Started Started container user-operator kubelet logger.go:42: 08:01:49 | streaming-with-autoprovisioning-autoscale | 2025-02-03 08:01:37 +0000 UTC Normal Pod auto-provisioned-collector-58d55796fc-745gf Binding Scheduled Successfully assigned kuttl-test-cunning-lamb/auto-provisioned-collector-58d55796fc-745gf to ip-10-0-66-98.us-east-2.compute.internal default-scheduler logger.go:42: 08:01:49 | streaming-with-autoprovisioning-autoscale | 2025-02-03 08:01:37 +0000 UTC Normal ReplicaSet.apps auto-provisioned-collector-58d55796fc SuccessfulCreate Created pod: auto-provisioned-collector-58d55796fc-745gf replicaset-controller logger.go:42: 08:01:49 | streaming-with-autoprovisioning-autoscale | 2025-02-03 08:01:37 +0000 UTC Normal Deployment.apps auto-provisioned-collector ScalingReplicaSet Scaled up replica set auto-provisioned-collector-58d55796fc to 1 deployment-controller logger.go:42: 08:01:49 | streaming-with-autoprovisioning-autoscale | 2025-02-03 08:01:37 +0000 UTC Normal Deployment.apps auto-provisioned-ingester ScalingReplicaSet Scaled up replica set auto-provisioned-ingester-5cbf994565 to 1 deployment-controller logger.go:42: 08:01:49 | streaming-with-autoprovisioning-autoscale | 2025-02-03 08:01:37 +0000 UTC Normal Pod auto-provisioned-query-cc69446db-cjlrv Binding Scheduled Successfully assigned kuttl-test-cunning-lamb/auto-provisioned-query-cc69446db-cjlrv to ip-10-0-66-98.us-east-2.compute.internal default-scheduler logger.go:42: 08:01:49 | streaming-with-autoprovisioning-autoscale | 2025-02-03 08:01:37 +0000 UTC Normal ReplicaSet.apps auto-provisioned-query-cc69446db SuccessfulCreate Created pod: auto-provisioned-query-cc69446db-cjlrv replicaset-controller logger.go:42: 08:01:49 | streaming-with-autoprovisioning-autoscale | 2025-02-03 08:01:37 +0000 UTC Normal Deployment.apps auto-provisioned-query ScalingReplicaSet Scaled up replica set auto-provisioned-query-cc69446db to 1 deployment-controller logger.go:42: 08:01:49 | streaming-with-autoprovisioning-autoscale | 2025-02-03 08:01:38 +0000 UTC Normal Pod auto-provisioned-collector-58d55796fc-745gf AddedInterface Add eth0 [10.129.2.78/23] from ovn-kubernetes multus logger.go:42: 08:01:49 | streaming-with-autoprovisioning-autoscale | 2025-02-03 08:01:38 +0000 UTC Normal Pod auto-provisioned-collector-58d55796fc-745gf.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:40a1c7fa23aa5ffa64e8e22aa38022f5d4d7ff644c46a3da7169b713d486c3c1" already present on machine kubelet logger.go:42: 08:01:49 | streaming-with-autoprovisioning-autoscale | 2025-02-03 08:01:38 +0000 UTC Normal Pod auto-provisioned-ingester-5cbf994565-8fcr7 Binding Scheduled Successfully assigned kuttl-test-cunning-lamb/auto-provisioned-ingester-5cbf994565-8fcr7 to ip-10-0-6-172.us-east-2.compute.internal default-scheduler logger.go:42: 08:01:49 | streaming-with-autoprovisioning-autoscale | 2025-02-03 08:01:38 +0000 UTC Normal ReplicaSet.apps auto-provisioned-ingester-5cbf994565 SuccessfulCreate Created pod: auto-provisioned-ingester-5cbf994565-8fcr7 replicaset-controller logger.go:42: 08:01:49 | streaming-with-autoprovisioning-autoscale | 2025-02-03 08:01:39 +0000 UTC Normal Pod auto-provisioned-collector-58d55796fc-745gf.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 08:01:49 | streaming-with-autoprovisioning-autoscale | 2025-02-03 08:01:39 +0000 UTC Normal Pod auto-provisioned-collector-58d55796fc-745gf.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 08:01:49 | streaming-with-autoprovisioning-autoscale | 2025-02-03 08:01:39 +0000 UTC Normal Pod auto-provisioned-ingester-5cbf994565-8fcr7 AddedInterface Add eth0 [10.131.0.67/23] from ovn-kubernetes multus logger.go:42: 08:01:49 | streaming-with-autoprovisioning-autoscale | 2025-02-03 08:01:39 +0000 UTC Normal Pod auto-provisioned-ingester-5cbf994565-8fcr7.spec.containers{jaeger-ingester} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-ingester-rhel8@sha256:b06a13d35b7ff03ed3a5512b64e5a472a8365e99dc5c9ba69e844db2b5284ee8" kubelet logger.go:42: 08:01:49 | streaming-with-autoprovisioning-autoscale | 2025-02-03 08:01:39 +0000 UTC Normal Pod auto-provisioned-query-cc69446db-cjlrv AddedInterface Add eth0 [10.129.2.79/23] from ovn-kubernetes multus logger.go:42: 08:01:49 | streaming-with-autoprovisioning-autoscale | 2025-02-03 08:01:39 +0000 UTC Normal Pod auto-provisioned-query-cc69446db-cjlrv.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:84020ce29bcb5a4bd018e6596188ae919c5cd600e08f78a546c0e76ea477685e" already present on machine kubelet logger.go:42: 08:01:49 | streaming-with-autoprovisioning-autoscale | 2025-02-03 08:01:39 +0000 UTC Normal Pod auto-provisioned-query-cc69446db-cjlrv.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 08:01:49 | streaming-with-autoprovisioning-autoscale | 2025-02-03 08:01:39 +0000 UTC Normal Pod auto-provisioned-query-cc69446db-cjlrv.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 08:01:49 | streaming-with-autoprovisioning-autoscale | 2025-02-03 08:01:39 +0000 UTC Normal Pod auto-provisioned-query-cc69446db-cjlrv.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" already present on machine kubelet logger.go:42: 08:01:49 | streaming-with-autoprovisioning-autoscale | 2025-02-03 08:01:39 +0000 UTC Normal Pod auto-provisioned-query-cc69446db-cjlrv.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 08:01:49 | streaming-with-autoprovisioning-autoscale | 2025-02-03 08:01:39 +0000 UTC Normal Pod auto-provisioned-query-cc69446db-cjlrv.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 08:01:49 | streaming-with-autoprovisioning-autoscale | 2025-02-03 08:01:39 +0000 UTC Normal Pod auto-provisioned-query-cc69446db-cjlrv.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:65342c7d622f336741a8ae0b9dead79c7ecabc155d02cdd42b7b49ca36680e74" already present on machine kubelet logger.go:42: 08:01:49 | streaming-with-autoprovisioning-autoscale | 2025-02-03 08:01:39 +0000 UTC Normal Pod auto-provisioned-query-cc69446db-cjlrv.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 08:01:49 | streaming-with-autoprovisioning-autoscale | 2025-02-03 08:01:39 +0000 UTC Normal Pod auto-provisioned-query-cc69446db-cjlrv.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 08:01:49 | streaming-with-autoprovisioning-autoscale | 2025-02-03 08:01:42 +0000 UTC Normal Pod auto-provisioned-ingester-5cbf994565-8fcr7.spec.containers{jaeger-ingester} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-ingester-rhel8@sha256:b06a13d35b7ff03ed3a5512b64e5a472a8365e99dc5c9ba69e844db2b5284ee8" in 3.564s (3.564s including waiting). Image size: 137309437 bytes. kubelet logger.go:42: 08:01:49 | streaming-with-autoprovisioning-autoscale | 2025-02-03 08:01:42 +0000 UTC Normal Pod auto-provisioned-ingester-5cbf994565-8fcr7.spec.containers{jaeger-ingester} Created Created container jaeger-ingester kubelet logger.go:42: 08:01:49 | streaming-with-autoprovisioning-autoscale | 2025-02-03 08:01:42 +0000 UTC Normal Pod auto-provisioned-ingester-5cbf994565-8fcr7.spec.containers{jaeger-ingester} Started Started container jaeger-ingester kubelet logger.go:42: 08:01:49 | streaming-with-autoprovisioning-autoscale | 2025-02-03 08:01:43 +0000 UTC Warning Pod auto-provisioned-ingester-5cbf994565-8fcr7.spec.containers{jaeger-ingester} Unhealthy Readiness probe failed: HTTP probe failed with statuscode: 503 kubelet logger.go:42: 08:01:49 | streaming-with-autoprovisioning-autoscale | Deleting namespace: kuttl-test-cunning-lamb === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (176.12s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (6.18s) --- PASS: kuttl/harness/streaming-with-autoprovisioning-autoscale (169.88s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name streaming --report --output /logs/artifacts/streaming.xml ./artifacts/kuttl-report.xml time="2025-02-03T08:02:03Z" level=debug msg="Setting a new name for the test suites" time="2025-02-03T08:02:03Z" level=debug msg="Removing 'artifacts' TestCase" time="2025-02-03T08:02:03Z" level=debug msg="normalizing test case names" time="2025-02-03T08:02:03Z" level=debug msg="streaming/artifacts -> streaming_artifacts" time="2025-02-03T08:02:03Z" level=debug msg="streaming/streaming-with-autoprovisioning-autoscale -> streaming_streaming_with_autoprovisioning_autoscale" +-----------------------------------------------------+--------+ | NAME | RESULT | +-----------------------------------------------------+--------+ | streaming_artifacts | passed | | streaming_streaming_with_autoprovisioning_autoscale | passed | +-----------------------------------------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh ui false true + '[' 3 -ne 3 ']' + test_suite_name=ui + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/ui.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-ui make[2]: Entering directory '/tmp/jaeger-tests' >>>> Elasticsearch image not loaded because SKIP_ES_EXTERNAL is true SKIP_ES_EXTERNAL=true ./tests/e2e/ui/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.18.0-0.nightly-2025-02-01-185117 True False 53m Cluster version is 4.18.0-0.nightly-2025-02-01-185117' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.18.0-0.nightly-2025-02-01-185117 True False 53m Cluster version is 4.18.0-0.nightly-2025-02-01-185117' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/ui/render.sh ++ export SUITE_DIR=./tests/e2e/ui ++ SUITE_DIR=./tests/e2e/ui ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/ui ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + start_test allinone + '[' 1 -ne 1 ']' + test_name=allinone + echo =========================================================================== =========================================================================== + info 'Rendering files for test allinone' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test allinone\e[0m' Rendering files for test allinone + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/ui/_build + '[' _build '!=' _build ']' + mkdir -p allinone + cd allinone + export GET_URL_COMMAND + export URL + export JAEGER_NAME=all-in-one-ui + JAEGER_NAME=all-in-one-ui + '[' true = true ']' + GET_URL_COMMAND='kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE' + URL='https://$(kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE)/search' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/ensure-ingress-host.sh.template -o ./ensure-ingress-host.sh + chmod +x ./ensure-ingress-host.sh + EXPECTED_CODE=200 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./01-curl.yaml + ASSERT_PRESENT=true + TRACKING_ID=MyTrackingId + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/test-ui-config.yaml.template -o ./04-test-ui-config.yaml + start_test production + '[' 1 -ne 1 ']' + test_name=production + echo =========================================================================== =========================================================================== + info 'Rendering files for test production' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test production\e[0m' Rendering files for test production + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/ui/_build/allinone + '[' allinone '!=' _build ']' + cd .. + mkdir -p production + cd production + export JAEGER_NAME=production-ui + JAEGER_NAME=production-ui + [[ true = true ]] + [[ true = true ]] + render_install_jaeger production-ui production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=production-ui + JAEGER_NAME=production-ui + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/ensure-ingress-host.sh.template -o ./ensure-ingress-host.sh + chmod +x ./ensure-ingress-host.sh + '[' true = true ']' + INSECURE=true + EXPECTED_CODE=403 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./02-check-forbbiden-access.yaml + EXPECTED_CODE=200 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./03-curl.yaml + INSECURE=true + EXPECTED_CODE=200 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./05-check-disabled-security.yaml + ASSERT_PRESENT=false + TRACKING_ID=MyTrackingId + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/test-ui-config.yaml.template -o ./06-check-NO-gaID.yaml + ASSERT_PRESENT=true + TRACKING_ID=MyTrackingId + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/test-ui-config.yaml.template -o ./08-check-gaID.yaml make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running ui E2E tests' Running ui E2E tests + cd tests/e2e/ui/_build + set +e + KUBECONFIG=/tmp/kubeconfig-2531963025 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-d102g9k0-f1a62.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 3 tests === RUN kuttl/harness === RUN kuttl/harness/allinone === PAUSE kuttl/harness/allinone === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/production === PAUSE kuttl/harness/production === CONT kuttl/harness/allinone logger.go:42: 08:02:10 | allinone | Ignoring ensure-ingress-host.sh as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 08:02:10 | allinone | Creating namespace: kuttl-test-poetic-joey logger.go:42: 08:02:10 | allinone/0-install | starting test step 0-install logger.go:42: 08:02:10 | allinone/0-install | Jaeger:kuttl-test-poetic-joey/all-in-one-ui created logger.go:42: 08:02:14 | allinone/0-install | test step completed 0-install logger.go:42: 08:02:14 | allinone/1-curl | starting test step 1-curl logger.go:42: 08:02:14 | allinone/1-curl | running command: [./ensure-ingress-host.sh] logger.go:42: 08:02:14 | allinone/1-curl | Checking the Ingress host value was populated logger.go:42: 08:02:14 | allinone/1-curl | Try number 0 logger.go:42: 08:02:14 | allinone/1-curl | Hostname is all-in-one-ui-kuttl-test-poetic-joey.apps.ci-op-d102g9k0-f1a62.cspilp.interop.ccitredhat.com logger.go:42: 08:02:14 | allinone/1-curl | running command: [sh -c ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 200 true $NAMESPACE all-in-one-ui] logger.go:42: 08:02:14 | allinone/1-curl | Checking an expected HTTP response logger.go:42: 08:02:14 | allinone/1-curl | Running in OpenShift logger.go:42: 08:02:14 | allinone/1-curl | User not provided. Getting the token... logger.go:42: 08:02:15 | allinone/1-curl | Warning: resource jaegers/all-in-one-ui is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 08:02:22 | allinone/1-curl | Try number 1/30 the https://all-in-one-ui-kuttl-test-poetic-joey.apps.ci-op-d102g9k0-f1a62.cspilp.interop.ccitredhat.com/search logger.go:42: 08:02:22 | allinone/1-curl | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 08:02:22 | allinone/1-curl | Try number 2/30 the https://all-in-one-ui-kuttl-test-poetic-joey.apps.ci-op-d102g9k0-f1a62.cspilp.interop.ccitredhat.com/search logger.go:42: 08:02:42 | allinone/1-curl | HTTP response is 503. 200 expected. Waiting 10 s logger.go:42: 08:02:52 | allinone/1-curl | Try number 3/30 the https://all-in-one-ui-kuttl-test-poetic-joey.apps.ci-op-d102g9k0-f1a62.cspilp.interop.ccitredhat.com/search logger.go:42: 08:02:52 | allinone/1-curl | curl response asserted properly logger.go:42: 08:02:52 | allinone/1-curl | test step completed 1-curl logger.go:42: 08:02:52 | allinone/2-delete | starting test step 2-delete logger.go:42: 08:02:52 | allinone/2-delete | Jaeger:kuttl-test-poetic-joey/all-in-one-ui created logger.go:42: 08:02:52 | allinone/2-delete | test step completed 2-delete logger.go:42: 08:02:52 | allinone/3-install | starting test step 3-install logger.go:42: 08:02:53 | allinone/3-install | Jaeger:kuttl-test-poetic-joey/all-in-one-ui updated logger.go:42: 08:02:53 | allinone/3-install | test step completed 3-install logger.go:42: 08:02:53 | allinone/4-test-ui-config | starting test step 4-test-ui-config logger.go:42: 08:02:53 | allinone/4-test-ui-config | running command: [./ensure-ingress-host.sh] logger.go:42: 08:02:53 | allinone/4-test-ui-config | Checking the Ingress host value was populated logger.go:42: 08:02:53 | allinone/4-test-ui-config | Try number 0 logger.go:42: 08:02:53 | allinone/4-test-ui-config | error: error executing jsonpath "{.items[0].status.ingress[0].host}": Error executing template: array index out of bounds: index 0, length 0. Printing more information for debugging the template: logger.go:42: 08:02:53 | allinone/4-test-ui-config | template was: logger.go:42: 08:02:53 | allinone/4-test-ui-config | {.items[0].status.ingress[0].host} logger.go:42: 08:02:53 | allinone/4-test-ui-config | object given to jsonpath engine was: logger.go:42: 08:02:53 | allinone/4-test-ui-config | map[string]interface {}{"apiVersion":"v1", "items":[]interface {}{}, "kind":"List", "metadata":map[string]interface {}{"resourceVersion":""}} logger.go:42: 08:02:53 | allinone/4-test-ui-config | logger.go:42: 08:02:53 | allinone/4-test-ui-config | logger.go:42: 08:03:03 | allinone/4-test-ui-config | Try number 1 logger.go:42: 08:03:03 | allinone/4-test-ui-config | Hostname is all-in-one-ui-kuttl-test-poetic-joey.apps.ci-op-d102g9k0-f1a62.cspilp.interop.ccitredhat.com logger.go:42: 08:03:03 | allinone/4-test-ui-config | running command: [sh -c ASSERT_PRESENT=true EXPECTED_CONTENT=MyTrackingId QUERY_HOSTNAME=https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search go run ../../../../cmd-utils/uiconfig/main.go] logger.go:42: 08:03:03 | allinone/4-test-ui-config | time="2025-02-03T08:03:03Z" level=info msg="Querying https://all-in-one-ui-kuttl-test-poetic-joey.apps.ci-op-d102g9k0-f1a62.cspilp.interop.ccitredhat.com/search..." logger.go:42: 08:03:03 | allinone/4-test-ui-config | time="2025-02-03T08:03:03Z" level=info msg="No secret provided for the Authorization header" logger.go:42: 08:03:03 | allinone/4-test-ui-config | time="2025-02-03T08:03:03Z" level=info msg="Polling to https://all-in-one-ui-kuttl-test-poetic-joey.apps.ci-op-d102g9k0-f1a62.cspilp.interop.ccitredhat.com/search" logger.go:42: 08:03:03 | allinone/4-test-ui-config | time="2025-02-03T08:03:03Z" level=info msg="Doing request number 0" logger.go:42: 08:03:03 | allinone/4-test-ui-config | time="2025-02-03T08:03:03Z" level=info msg="Content found and asserted!" logger.go:42: 08:03:03 | allinone/4-test-ui-config | time="2025-02-03T08:03:03Z" level=info msg="Success!" logger.go:42: 08:03:03 | allinone/4-test-ui-config | test step completed 4-test-ui-config logger.go:42: 08:03:03 | allinone | allinone events from ns kuttl-test-poetic-joey: logger.go:42: 08:03:03 | allinone | 2025-02-03 08:02:14 +0000 UTC Normal Pod all-in-one-ui-6d55958444-fdcb2 Binding Scheduled Successfully assigned kuttl-test-poetic-joey/all-in-one-ui-6d55958444-fdcb2 to ip-10-0-66-98.us-east-2.compute.internal default-scheduler logger.go:42: 08:03:03 | allinone | 2025-02-03 08:02:14 +0000 UTC Normal ReplicaSet.apps all-in-one-ui-6d55958444 SuccessfulCreate Created pod: all-in-one-ui-6d55958444-fdcb2 replicaset-controller logger.go:42: 08:03:03 | allinone | 2025-02-03 08:02:14 +0000 UTC Normal Deployment.apps all-in-one-ui ScalingReplicaSet Scaled up replica set all-in-one-ui-6d55958444 to 1 deployment-controller logger.go:42: 08:03:03 | allinone | 2025-02-03 08:02:15 +0000 UTC Normal Pod all-in-one-ui-6d55958444-fdcb2 AddedInterface Add eth0 [10.129.2.80/23] from ovn-kubernetes multus logger.go:42: 08:03:03 | allinone | 2025-02-03 08:02:15 +0000 UTC Normal Pod all-in-one-ui-6d55958444-fdcb2.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:fa9aaf9ae026228265ce7f10fb451d5c52a4e0269ce1fa9024cbefddf765ae8d" already present on machine kubelet logger.go:42: 08:03:03 | allinone | 2025-02-03 08:02:15 +0000 UTC Normal Pod all-in-one-ui-6d55958444-fdcb2.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 08:03:03 | allinone | 2025-02-03 08:02:15 +0000 UTC Normal Pod all-in-one-ui-6d55958444-fdcb2.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 08:03:03 | allinone | 2025-02-03 08:02:15 +0000 UTC Normal Pod all-in-one-ui-6d55958444-fdcb2.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" already present on machine kubelet logger.go:42: 08:03:03 | allinone | 2025-02-03 08:02:15 +0000 UTC Normal Pod all-in-one-ui-6d55958444-fdcb2.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 08:03:03 | allinone | 2025-02-03 08:02:15 +0000 UTC Normal Pod all-in-one-ui-6d55958444-fdcb2.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 08:03:03 | allinone | 2025-02-03 08:02:19 +0000 UTC Normal Pod all-in-one-ui-6d55958444-fdcb2.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 08:03:03 | allinone | 2025-02-03 08:02:19 +0000 UTC Normal Pod all-in-one-ui-6d55958444-fdcb2.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 08:03:03 | allinone | 2025-02-03 08:02:19 +0000 UTC Normal ReplicaSet.apps all-in-one-ui-6d55958444 SuccessfulDelete Deleted pod: all-in-one-ui-6d55958444-fdcb2 replicaset-controller logger.go:42: 08:03:03 | allinone | 2025-02-03 08:02:19 +0000 UTC Normal Deployment.apps all-in-one-ui ScalingReplicaSet Scaled down replica set all-in-one-ui-6d55958444 to 0 from 1 deployment-controller logger.go:42: 08:03:03 | allinone | 2025-02-03 08:02:20 +0000 UTC Normal Pod all-in-one-ui-6cb675545-g9kdd Binding Scheduled Successfully assigned kuttl-test-poetic-joey/all-in-one-ui-6cb675545-g9kdd to ip-10-0-66-98.us-east-2.compute.internal default-scheduler logger.go:42: 08:03:03 | allinone | 2025-02-03 08:02:20 +0000 UTC Normal ReplicaSet.apps all-in-one-ui-6cb675545 SuccessfulCreate Created pod: all-in-one-ui-6cb675545-g9kdd replicaset-controller logger.go:42: 08:03:03 | allinone | 2025-02-03 08:02:20 +0000 UTC Normal Deployment.apps all-in-one-ui ScalingReplicaSet Scaled up replica set all-in-one-ui-6cb675545 to 1 deployment-controller logger.go:42: 08:03:03 | allinone | 2025-02-03 08:02:21 +0000 UTC Normal Pod all-in-one-ui-6cb675545-g9kdd AddedInterface Add eth0 [10.129.2.81/23] from ovn-kubernetes multus logger.go:42: 08:03:03 | allinone | 2025-02-03 08:02:21 +0000 UTC Normal Pod all-in-one-ui-6cb675545-g9kdd.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:fa9aaf9ae026228265ce7f10fb451d5c52a4e0269ce1fa9024cbefddf765ae8d" already present on machine kubelet logger.go:42: 08:03:03 | allinone | 2025-02-03 08:02:21 +0000 UTC Normal Pod all-in-one-ui-6cb675545-g9kdd.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 08:03:03 | allinone | 2025-02-03 08:02:21 +0000 UTC Normal Pod all-in-one-ui-6cb675545-g9kdd.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 08:03:03 | allinone | 2025-02-03 08:02:21 +0000 UTC Normal Pod all-in-one-ui-6cb675545-g9kdd.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" already present on machine kubelet logger.go:42: 08:03:03 | allinone | 2025-02-03 08:02:21 +0000 UTC Normal Pod all-in-one-ui-6cb675545-g9kdd.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 08:03:03 | allinone | 2025-02-03 08:02:22 +0000 UTC Normal Pod all-in-one-ui-6cb675545-g9kdd.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 08:03:03 | allinone | 2025-02-03 08:02:53 +0000 UTC Normal Pod all-in-one-ui-6cb675545-g9kdd.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 08:03:03 | allinone | 2025-02-03 08:02:53 +0000 UTC Normal Pod all-in-one-ui-6cb675545-g9kdd.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 08:03:03 | allinone | 2025-02-03 08:02:59 +0000 UTC Normal Pod all-in-one-ui-7864cdc58f-qmklx Binding Scheduled Successfully assigned kuttl-test-poetic-joey/all-in-one-ui-7864cdc58f-qmklx to ip-10-0-66-98.us-east-2.compute.internal default-scheduler logger.go:42: 08:03:03 | allinone | 2025-02-03 08:02:59 +0000 UTC Normal ReplicaSet.apps all-in-one-ui-7864cdc58f SuccessfulCreate Created pod: all-in-one-ui-7864cdc58f-qmklx replicaset-controller logger.go:42: 08:03:03 | allinone | 2025-02-03 08:02:59 +0000 UTC Normal Deployment.apps all-in-one-ui ScalingReplicaSet Scaled up replica set all-in-one-ui-7864cdc58f to 1 deployment-controller logger.go:42: 08:03:03 | allinone | 2025-02-03 08:03:00 +0000 UTC Normal Pod all-in-one-ui-7864cdc58f-qmklx AddedInterface Add eth0 [10.129.2.82/23] from ovn-kubernetes multus logger.go:42: 08:03:03 | allinone | 2025-02-03 08:03:00 +0000 UTC Normal Pod all-in-one-ui-7864cdc58f-qmklx.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:fa9aaf9ae026228265ce7f10fb451d5c52a4e0269ce1fa9024cbefddf765ae8d" already present on machine kubelet logger.go:42: 08:03:03 | allinone | 2025-02-03 08:03:00 +0000 UTC Normal Pod all-in-one-ui-7864cdc58f-qmklx.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 08:03:03 | allinone | 2025-02-03 08:03:00 +0000 UTC Normal Pod all-in-one-ui-7864cdc58f-qmklx.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 08:03:03 | allinone | Deleting namespace: kuttl-test-poetic-joey === CONT kuttl/harness/production logger.go:42: 08:03:10 | production | Ignoring add-tracking-id.yaml as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 08:03:10 | production | Ignoring ensure-ingress-host.sh as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 08:03:10 | production | Creating namespace: kuttl-test-fitting-kite logger.go:42: 08:03:10 | production/1-install | starting test step 1-install logger.go:42: 08:03:10 | production/1-install | Jaeger:kuttl-test-fitting-kite/production-ui created logger.go:42: 08:04:01 | production/1-install | test step completed 1-install logger.go:42: 08:04:01 | production/2-check-forbbiden-access | starting test step 2-check-forbbiden-access logger.go:42: 08:04:01 | production/2-check-forbbiden-access | running command: [./ensure-ingress-host.sh] logger.go:42: 08:04:01 | production/2-check-forbbiden-access | Checking the Ingress host value was populated logger.go:42: 08:04:01 | production/2-check-forbbiden-access | Try number 0 logger.go:42: 08:04:01 | production/2-check-forbbiden-access | Hostname is production-ui-kuttl-test-fitting-kite.apps.ci-op-d102g9k0-f1a62.cspilp.interop.ccitredhat.com logger.go:42: 08:04:01 | production/2-check-forbbiden-access | running command: [sh -c INSECURE=true ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 403 true $NAMESPACE production-ui] logger.go:42: 08:04:01 | production/2-check-forbbiden-access | Checking an expected HTTP response logger.go:42: 08:04:01 | production/2-check-forbbiden-access | Running in OpenShift logger.go:42: 08:04:01 | production/2-check-forbbiden-access | Not using any secret logger.go:42: 08:04:01 | production/2-check-forbbiden-access | Try number 1/30 the https://production-ui-kuttl-test-fitting-kite.apps.ci-op-d102g9k0-f1a62.cspilp.interop.ccitredhat.com/search logger.go:42: 08:04:01 | production/2-check-forbbiden-access | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 08:04:01 | production/2-check-forbbiden-access | Try number 2/30 the https://production-ui-kuttl-test-fitting-kite.apps.ci-op-d102g9k0-f1a62.cspilp.interop.ccitredhat.com/search logger.go:42: 08:04:01 | production/2-check-forbbiden-access | HTTP response is 503. 403 expected. Waiting 10 s logger.go:42: 08:04:11 | production/2-check-forbbiden-access | Try number 3/30 the https://production-ui-kuttl-test-fitting-kite.apps.ci-op-d102g9k0-f1a62.cspilp.interop.ccitredhat.com/search logger.go:42: 08:04:11 | production/2-check-forbbiden-access | curl response asserted properly logger.go:42: 08:04:11 | production/2-check-forbbiden-access | test step completed 2-check-forbbiden-access logger.go:42: 08:04:11 | production/3-curl | starting test step 3-curl logger.go:42: 08:04:11 | production/3-curl | running command: [./ensure-ingress-host.sh] logger.go:42: 08:04:11 | production/3-curl | Checking the Ingress host value was populated logger.go:42: 08:04:11 | production/3-curl | Try number 0 logger.go:42: 08:04:11 | production/3-curl | Hostname is production-ui-kuttl-test-fitting-kite.apps.ci-op-d102g9k0-f1a62.cspilp.interop.ccitredhat.com logger.go:42: 08:04:11 | production/3-curl | running command: [sh -c ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 200 true $NAMESPACE production-ui] logger.go:42: 08:04:11 | production/3-curl | Checking an expected HTTP response logger.go:42: 08:04:11 | production/3-curl | Running in OpenShift logger.go:42: 08:04:11 | production/3-curl | User not provided. Getting the token... logger.go:42: 08:04:12 | production/3-curl | Warning: resource jaegers/production-ui is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 08:04:19 | production/3-curl | Try number 1/30 the https://production-ui-kuttl-test-fitting-kite.apps.ci-op-d102g9k0-f1a62.cspilp.interop.ccitredhat.com/search logger.go:42: 08:04:19 | production/3-curl | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 08:04:19 | production/3-curl | Try number 2/30 the https://production-ui-kuttl-test-fitting-kite.apps.ci-op-d102g9k0-f1a62.cspilp.interop.ccitredhat.com/search logger.go:42: 08:04:19 | production/3-curl | HTTP response is 503. 200 expected. Waiting 10 s logger.go:42: 08:04:29 | production/3-curl | Try number 3/30 the https://production-ui-kuttl-test-fitting-kite.apps.ci-op-d102g9k0-f1a62.cspilp.interop.ccitredhat.com/search logger.go:42: 08:04:29 | production/3-curl | curl response asserted properly logger.go:42: 08:04:29 | production/3-curl | test step completed 3-curl logger.go:42: 08:04:29 | production/4-install | starting test step 4-install logger.go:42: 08:04:29 | production/4-install | Jaeger:kuttl-test-fitting-kite/production-ui updated logger.go:42: 08:04:29 | production/4-install | test step completed 4-install logger.go:42: 08:04:29 | production/5-check-disabled-security | starting test step 5-check-disabled-security logger.go:42: 08:04:29 | production/5-check-disabled-security | running command: [./ensure-ingress-host.sh] logger.go:42: 08:04:29 | production/5-check-disabled-security | Checking the Ingress host value was populated logger.go:42: 08:04:29 | production/5-check-disabled-security | Try number 0 logger.go:42: 08:04:29 | production/5-check-disabled-security | Hostname is production-ui-kuttl-test-fitting-kite.apps.ci-op-d102g9k0-f1a62.cspilp.interop.ccitredhat.com logger.go:42: 08:04:29 | production/5-check-disabled-security | running command: [sh -c INSECURE=true ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 200 true $NAMESPACE production-ui] logger.go:42: 08:04:29 | production/5-check-disabled-security | Checking an expected HTTP response logger.go:42: 08:04:29 | production/5-check-disabled-security | Running in OpenShift logger.go:42: 08:04:29 | production/5-check-disabled-security | Not using any secret logger.go:42: 08:04:29 | production/5-check-disabled-security | Try number 1/30 the https://production-ui-kuttl-test-fitting-kite.apps.ci-op-d102g9k0-f1a62.cspilp.interop.ccitredhat.com/search logger.go:42: 08:04:29 | production/5-check-disabled-security | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 08:04:29 | production/5-check-disabled-security | Try number 2/30 the https://production-ui-kuttl-test-fitting-kite.apps.ci-op-d102g9k0-f1a62.cspilp.interop.ccitredhat.com/search logger.go:42: 08:04:29 | production/5-check-disabled-security | HTTP response is 403. 200 expected. Waiting 10 s logger.go:42: 08:04:39 | production/5-check-disabled-security | Try number 3/30 the https://production-ui-kuttl-test-fitting-kite.apps.ci-op-d102g9k0-f1a62.cspilp.interop.ccitredhat.com/search logger.go:42: 08:04:39 | production/5-check-disabled-security | curl response asserted properly logger.go:42: 08:04:39 | production/5-check-disabled-security | test step completed 5-check-disabled-security logger.go:42: 08:04:39 | production/6-check-NO-gaID | starting test step 6-check-NO-gaID logger.go:42: 08:04:39 | production/6-check-NO-gaID | running command: [./ensure-ingress-host.sh] logger.go:42: 08:04:39 | production/6-check-NO-gaID | Checking the Ingress host value was populated logger.go:42: 08:04:39 | production/6-check-NO-gaID | Try number 0 logger.go:42: 08:04:39 | production/6-check-NO-gaID | Hostname is production-ui-kuttl-test-fitting-kite.apps.ci-op-d102g9k0-f1a62.cspilp.interop.ccitredhat.com logger.go:42: 08:04:39 | production/6-check-NO-gaID | running command: [sh -c ASSERT_PRESENT=false EXPECTED_CONTENT=MyTrackingId QUERY_HOSTNAME=https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search go run ../../../../cmd-utils/uiconfig/main.go] logger.go:42: 08:04:40 | production/6-check-NO-gaID | time="2025-02-03T08:04:40Z" level=info msg="Querying https://production-ui-kuttl-test-fitting-kite.apps.ci-op-d102g9k0-f1a62.cspilp.interop.ccitredhat.com/search..." logger.go:42: 08:04:40 | production/6-check-NO-gaID | time="2025-02-03T08:04:40Z" level=info msg="No secret provided for the Authorization header" logger.go:42: 08:04:40 | production/6-check-NO-gaID | time="2025-02-03T08:04:40Z" level=info msg="Polling to https://production-ui-kuttl-test-fitting-kite.apps.ci-op-d102g9k0-f1a62.cspilp.interop.ccitredhat.com/search" logger.go:42: 08:04:40 | production/6-check-NO-gaID | time="2025-02-03T08:04:40Z" level=info msg="Doing request number 0" logger.go:42: 08:04:40 | production/6-check-NO-gaID | time="2025-02-03T08:04:40Z" level=info msg="Content not found and asserted it was not found!" logger.go:42: 08:04:40 | production/6-check-NO-gaID | time="2025-02-03T08:04:40Z" level=info msg="Success!" logger.go:42: 08:04:40 | production/6-check-NO-gaID | test step completed 6-check-NO-gaID logger.go:42: 08:04:40 | production/7-add-tracking-id | starting test step 7-add-tracking-id logger.go:42: 08:04:40 | production/7-add-tracking-id | running command: [sh -c kubectl apply -f add-tracking-id.yaml -n $NAMESPACE] logger.go:42: 08:04:40 | production/7-add-tracking-id | jaeger.jaegertracing.io/production-ui configured logger.go:42: 08:04:40 | production/7-add-tracking-id | test step completed 7-add-tracking-id logger.go:42: 08:04:40 | production/8-check-gaID | starting test step 8-check-gaID logger.go:42: 08:04:40 | production/8-check-gaID | running command: [./ensure-ingress-host.sh] logger.go:42: 08:04:40 | production/8-check-gaID | Checking the Ingress host value was populated logger.go:42: 08:04:40 | production/8-check-gaID | Try number 0 logger.go:42: 08:04:40 | production/8-check-gaID | Hostname is production-ui-kuttl-test-fitting-kite.apps.ci-op-d102g9k0-f1a62.cspilp.interop.ccitredhat.com logger.go:42: 08:04:40 | production/8-check-gaID | running command: [sh -c ASSERT_PRESENT=true EXPECTED_CONTENT=MyTrackingId QUERY_HOSTNAME=https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search go run ../../../../cmd-utils/uiconfig/main.go] logger.go:42: 08:04:41 | production/8-check-gaID | time="2025-02-03T08:04:41Z" level=info msg="Querying https://production-ui-kuttl-test-fitting-kite.apps.ci-op-d102g9k0-f1a62.cspilp.interop.ccitredhat.com/search..." logger.go:42: 08:04:41 | production/8-check-gaID | time="2025-02-03T08:04:41Z" level=info msg="No secret provided for the Authorization header" logger.go:42: 08:04:41 | production/8-check-gaID | time="2025-02-03T08:04:41Z" level=info msg="Polling to https://production-ui-kuttl-test-fitting-kite.apps.ci-op-d102g9k0-f1a62.cspilp.interop.ccitredhat.com/search" logger.go:42: 08:04:41 | production/8-check-gaID | time="2025-02-03T08:04:41Z" level=info msg="Doing request number 0" logger.go:42: 08:04:41 | production/8-check-gaID | time="2025-02-03T08:04:41Z" level=warning msg="Found: false . Assert: true" logger.go:42: 08:04:41 | production/8-check-gaID | time="2025-02-03T08:04:41Z" level=warning msg="The condition of the test function was not accomplished" logger.go:42: 08:04:49 | production/8-check-gaID | time="2025-02-03T08:04:49Z" level=info msg="Doing request number 1" logger.go:42: 08:05:09 | production/8-check-gaID | time="2025-02-03T08:05:09Z" level=info msg="Content found and asserted!" logger.go:42: 08:05:09 | production/8-check-gaID | time="2025-02-03T08:05:09Z" level=info msg="Success!" logger.go:42: 08:05:09 | production/8-check-gaID | test step completed 8-check-gaID logger.go:42: 08:05:09 | production | production events from ns kuttl-test-fitting-kite: logger.go:42: 08:05:09 | production | 2025-02-03 08:03:30 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestfittingkiteproductionui-1-66c7d76bd7 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestfittingkiteproductionui-1-66c7dkp6sm replicaset-controller logger.go:42: 08:05:09 | production | 2025-02-03 08:03:30 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfittingkiteproductionui-1-66c7dkp6sm Binding Scheduled Successfully assigned kuttl-test-fitting-kite/elasticsearch-cdm-kuttltestfittingkiteproductionui-1-66c7dkp6sm to ip-10-0-6-172.us-east-2.compute.internal default-scheduler logger.go:42: 08:05:09 | production | 2025-02-03 08:03:30 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestfittingkiteproductionui-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestfittingkiteproductionui-1-66c7d76bd7 to 1 deployment-controller logger.go:42: 08:05:09 | production | 2025-02-03 08:03:31 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfittingkiteproductionui-1-66c7dkp6sm AddedInterface Add eth0 [10.131.0.68/23] from ovn-kubernetes multus logger.go:42: 08:05:09 | production | 2025-02-03 08:03:31 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfittingkiteproductionui-1-66c7dkp6sm.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:410674bf2f77ff1d431f494cadfab3e81249b35695eb65aedfdb8498079b4d35" already present on machine kubelet logger.go:42: 08:05:09 | production | 2025-02-03 08:03:31 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfittingkiteproductionui-1-66c7dkp6sm.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 08:05:09 | production | 2025-02-03 08:03:32 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfittingkiteproductionui-1-66c7dkp6sm.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 08:05:09 | production | 2025-02-03 08:03:32 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfittingkiteproductionui-1-66c7dkp6sm.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:fe2439e5f9e14d828caace425214edfa3edc179d72d225c43f69e9e4c9a5837b" already present on machine kubelet logger.go:42: 08:05:09 | production | 2025-02-03 08:03:32 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfittingkiteproductionui-1-66c7dkp6sm.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 08:05:09 | production | 2025-02-03 08:03:32 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfittingkiteproductionui-1-66c7dkp6sm.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 08:05:09 | production | 2025-02-03 08:03:41 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestfittingkiteproductionui-1-66c7dkp6sm.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 08:05:09 | production | 2025-02-03 08:03:46 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestfittingkiteproductionui-1-66c7dkp6sm.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 08:05:09 | production | 2025-02-03 08:03:57 +0000 UTC Normal Pod production-ui-collector-5648c49786-jkqp2 Binding Scheduled Successfully assigned kuttl-test-fitting-kite/production-ui-collector-5648c49786-jkqp2 to ip-10-0-66-98.us-east-2.compute.internal default-scheduler logger.go:42: 08:05:09 | production | 2025-02-03 08:03:57 +0000 UTC Normal ReplicaSet.apps production-ui-collector-5648c49786 SuccessfulCreate Created pod: production-ui-collector-5648c49786-jkqp2 replicaset-controller logger.go:42: 08:05:09 | production | 2025-02-03 08:03:57 +0000 UTC Normal Deployment.apps production-ui-collector ScalingReplicaSet Scaled up replica set production-ui-collector-5648c49786 to 1 deployment-controller logger.go:42: 08:05:09 | production | 2025-02-03 08:03:57 +0000 UTC Normal Pod production-ui-query-676765dd67-f5b2m Binding Scheduled Successfully assigned kuttl-test-fitting-kite/production-ui-query-676765dd67-f5b2m to ip-10-0-26-111.us-east-2.compute.internal default-scheduler logger.go:42: 08:05:09 | production | 2025-02-03 08:03:57 +0000 UTC Warning Pod production-ui-query-676765dd67-f5b2m FailedMount MountVolume.SetUp failed for volume "production-ui-ui-oauth-proxy-tls" : secret "production-ui-ui-oauth-proxy-tls" not found kubelet logger.go:42: 08:05:09 | production | 2025-02-03 08:03:57 +0000 UTC Normal ReplicaSet.apps production-ui-query-676765dd67 SuccessfulCreate Created pod: production-ui-query-676765dd67-f5b2m replicaset-controller logger.go:42: 08:05:09 | production | 2025-02-03 08:03:57 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled up replica set production-ui-query-676765dd67 to 1 deployment-controller logger.go:42: 08:05:09 | production | 2025-02-03 08:03:58 +0000 UTC Normal Pod production-ui-collector-5648c49786-jkqp2 AddedInterface Add eth0 [10.129.2.83/23] from ovn-kubernetes multus logger.go:42: 08:05:09 | production | 2025-02-03 08:03:58 +0000 UTC Normal Pod production-ui-collector-5648c49786-jkqp2.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:40a1c7fa23aa5ffa64e8e22aa38022f5d4d7ff644c46a3da7169b713d486c3c1" already present on machine kubelet logger.go:42: 08:05:09 | production | 2025-02-03 08:03:58 +0000 UTC Normal Pod production-ui-collector-5648c49786-jkqp2.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 08:05:09 | production | 2025-02-03 08:03:58 +0000 UTC Normal Pod production-ui-collector-5648c49786-jkqp2.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 08:05:09 | production | 2025-02-03 08:03:59 +0000 UTC Normal Pod production-ui-query-676765dd67-f5b2m AddedInterface Add eth0 [10.128.2.77/23] from ovn-kubernetes multus logger.go:42: 08:05:09 | production | 2025-02-03 08:03:59 +0000 UTC Normal Pod production-ui-query-676765dd67-f5b2m.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:84020ce29bcb5a4bd018e6596188ae919c5cd600e08f78a546c0e76ea477685e" already present on machine kubelet logger.go:42: 08:05:09 | production | 2025-02-03 08:03:59 +0000 UTC Normal Pod production-ui-query-676765dd67-f5b2m.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 08:05:09 | production | 2025-02-03 08:03:59 +0000 UTC Normal Pod production-ui-query-676765dd67-f5b2m.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 08:05:09 | production | 2025-02-03 08:03:59 +0000 UTC Normal Pod production-ui-query-676765dd67-f5b2m.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" already present on machine kubelet logger.go:42: 08:05:09 | production | 2025-02-03 08:03:59 +0000 UTC Normal Pod production-ui-query-676765dd67-f5b2m.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 08:05:09 | production | 2025-02-03 08:03:59 +0000 UTC Normal Pod production-ui-query-676765dd67-f5b2m.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 08:05:09 | production | 2025-02-03 08:03:59 +0000 UTC Normal Pod production-ui-query-676765dd67-f5b2m.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:65342c7d622f336741a8ae0b9dead79c7ecabc155d02cdd42b7b49ca36680e74" already present on machine kubelet logger.go:42: 08:05:09 | production | 2025-02-03 08:03:59 +0000 UTC Normal Pod production-ui-query-676765dd67-f5b2m.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 08:05:09 | production | 2025-02-03 08:03:59 +0000 UTC Normal Pod production-ui-query-676765dd67-f5b2m.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 08:05:09 | production | 2025-02-03 08:04:12 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling production-ui-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 08:05:09 | production | 2025-02-03 08:04:12 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling production-ui-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 08:05:09 | production | 2025-02-03 08:04:12 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling production-ui-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 08:05:09 | production | 2025-02-03 08:04:14 +0000 UTC Normal Pod production-ui-query-676765dd67-f5b2m.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 08:05:09 | production | 2025-02-03 08:04:14 +0000 UTC Normal Pod production-ui-query-676765dd67-f5b2m.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 08:05:09 | production | 2025-02-03 08:04:14 +0000 UTC Normal Pod production-ui-query-676765dd67-f5b2m.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 08:05:09 | production | 2025-02-03 08:04:14 +0000 UTC Normal ReplicaSet.apps production-ui-query-676765dd67 SuccessfulDelete Deleted pod: production-ui-query-676765dd67-f5b2m replicaset-controller logger.go:42: 08:05:09 | production | 2025-02-03 08:04:14 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled down replica set production-ui-query-676765dd67 to 0 from 1 deployment-controller logger.go:42: 08:05:09 | production | 2025-02-03 08:04:15 +0000 UTC Normal Pod production-ui-query-677f877cc8-mfqbr Binding Scheduled Successfully assigned kuttl-test-fitting-kite/production-ui-query-677f877cc8-mfqbr to ip-10-0-26-111.us-east-2.compute.internal default-scheduler logger.go:42: 08:05:09 | production | 2025-02-03 08:04:15 +0000 UTC Normal ReplicaSet.apps production-ui-query-677f877cc8 SuccessfulCreate Created pod: production-ui-query-677f877cc8-mfqbr replicaset-controller logger.go:42: 08:05:09 | production | 2025-02-03 08:04:15 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled up replica set production-ui-query-677f877cc8 to 1 deployment-controller logger.go:42: 08:05:09 | production | 2025-02-03 08:04:16 +0000 UTC Normal Pod production-ui-query-677f877cc8-mfqbr AddedInterface Add eth0 [10.128.2.78/23] from ovn-kubernetes multus logger.go:42: 08:05:09 | production | 2025-02-03 08:04:16 +0000 UTC Normal Pod production-ui-query-677f877cc8-mfqbr.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:84020ce29bcb5a4bd018e6596188ae919c5cd600e08f78a546c0e76ea477685e" already present on machine kubelet logger.go:42: 08:05:09 | production | 2025-02-03 08:04:16 +0000 UTC Normal Pod production-ui-query-677f877cc8-mfqbr.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 08:05:09 | production | 2025-02-03 08:04:16 +0000 UTC Normal Pod production-ui-query-677f877cc8-mfqbr.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 08:05:09 | production | 2025-02-03 08:04:16 +0000 UTC Normal Pod production-ui-query-677f877cc8-mfqbr.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:234af927030921ab8f7333f61f967b4b4dee37a1b3cf85689e9e63240dd62800" already present on machine kubelet logger.go:42: 08:05:09 | production | 2025-02-03 08:04:16 +0000 UTC Normal Pod production-ui-query-677f877cc8-mfqbr.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 08:05:09 | production | 2025-02-03 08:04:16 +0000 UTC Normal Pod production-ui-query-677f877cc8-mfqbr.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 08:05:09 | production | 2025-02-03 08:04:16 +0000 UTC Normal Pod production-ui-query-677f877cc8-mfqbr.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:65342c7d622f336741a8ae0b9dead79c7ecabc155d02cdd42b7b49ca36680e74" already present on machine kubelet logger.go:42: 08:05:09 | production | 2025-02-03 08:04:16 +0000 UTC Normal Pod production-ui-query-677f877cc8-mfqbr.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 08:05:09 | production | 2025-02-03 08:04:16 +0000 UTC Normal Pod production-ui-query-677f877cc8-mfqbr.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 08:05:09 | production | 2025-02-03 08:04:30 +0000 UTC Normal Pod production-ui-query-677f877cc8-mfqbr.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 08:05:09 | production | 2025-02-03 08:04:30 +0000 UTC Normal Pod production-ui-query-677f877cc8-mfqbr.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 08:05:09 | production | 2025-02-03 08:04:30 +0000 UTC Normal Pod production-ui-query-677f877cc8-mfqbr.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 08:05:09 | production | 2025-02-03 08:04:30 +0000 UTC Normal ReplicaSet.apps production-ui-query-677f877cc8 SuccessfulDelete Deleted pod: production-ui-query-677f877cc8-mfqbr replicaset-controller logger.go:42: 08:05:09 | production | 2025-02-03 08:04:30 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled down replica set production-ui-query-677f877cc8 to 0 from 1 deployment-controller logger.go:42: 08:05:09 | production | 2025-02-03 08:04:31 +0000 UTC Normal Pod production-ui-query-5ffc66789d-h9h6z Binding Scheduled Successfully assigned kuttl-test-fitting-kite/production-ui-query-5ffc66789d-h9h6z to ip-10-0-26-111.us-east-2.compute.internal default-scheduler logger.go:42: 08:05:09 | production | 2025-02-03 08:04:31 +0000 UTC Normal ReplicaSet.apps production-ui-query-5ffc66789d SuccessfulCreate Created pod: production-ui-query-5ffc66789d-h9h6z replicaset-controller logger.go:42: 08:05:09 | production | 2025-02-03 08:04:31 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled up replica set production-ui-query-5ffc66789d to 1 deployment-controller logger.go:42: 08:05:09 | production | 2025-02-03 08:04:32 +0000 UTC Normal Pod production-ui-query-5ffc66789d-h9h6z AddedInterface Add eth0 [10.128.2.79/23] from ovn-kubernetes multus logger.go:42: 08:05:09 | production | 2025-02-03 08:04:32 +0000 UTC Normal Pod production-ui-query-5ffc66789d-h9h6z.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:84020ce29bcb5a4bd018e6596188ae919c5cd600e08f78a546c0e76ea477685e" already present on machine kubelet logger.go:42: 08:05:09 | production | 2025-02-03 08:04:32 +0000 UTC Normal Pod production-ui-query-5ffc66789d-h9h6z.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 08:05:09 | production | 2025-02-03 08:04:32 +0000 UTC Normal Pod production-ui-query-5ffc66789d-h9h6z.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 08:05:09 | production | 2025-02-03 08:04:32 +0000 UTC Normal Pod production-ui-query-5ffc66789d-h9h6z.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:65342c7d622f336741a8ae0b9dead79c7ecabc155d02cdd42b7b49ca36680e74" already present on machine kubelet logger.go:42: 08:05:09 | production | 2025-02-03 08:04:32 +0000 UTC Normal Pod production-ui-query-5ffc66789d-h9h6z.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 08:05:09 | production | 2025-02-03 08:04:32 +0000 UTC Normal Pod production-ui-query-5ffc66789d-h9h6z.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 08:05:09 | production | 2025-02-03 08:04:41 +0000 UTC Normal Pod production-ui-query-5ffc66789d-h9h6z.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 08:05:09 | production | 2025-02-03 08:04:41 +0000 UTC Normal Pod production-ui-query-5ffc66789d-h9h6z.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 08:05:09 | production | 2025-02-03 08:04:41 +0000 UTC Normal ReplicaSet.apps production-ui-query-5ffc66789d SuccessfulDelete Deleted pod: production-ui-query-5ffc66789d-h9h6z replicaset-controller logger.go:42: 08:05:09 | production | 2025-02-03 08:04:41 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled down replica set production-ui-query-5ffc66789d to 0 from 1 deployment-controller logger.go:42: 08:05:09 | production | 2025-02-03 08:04:42 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling production-ui-collector FailedGetResourceMetric failed to get cpu utilization: missing request for cpu in container jaeger-collector of Pod production-ui-collector-5648c49786-jkqp2 horizontal-pod-autoscaler logger.go:42: 08:05:09 | production | 2025-02-03 08:04:42 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling production-ui-collector FailedGetResourceMetric failed to get memory utilization: missing request for memory in container jaeger-collector of Pod production-ui-collector-5648c49786-jkqp2 horizontal-pod-autoscaler logger.go:42: 08:05:09 | production | 2025-02-03 08:04:42 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling production-ui-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: missing request for cpu in container jaeger-collector of Pod production-ui-collector-5648c49786-jkqp2 horizontal-pod-autoscaler logger.go:42: 08:05:09 | production | 2025-02-03 08:04:42 +0000 UTC Warning Pod production-ui-query-5ffc66789d-h9h6z.spec.containers{jaeger-query} Unhealthy Readiness probe failed: Get "http://10.128.2.79:16687/": dial tcp 10.128.2.79:16687: connect: connection refused kubelet logger.go:42: 08:05:09 | production | 2025-02-03 08:04:42 +0000 UTC Warning Pod production-ui-query-5ffc66789d-h9h6z.spec.containers{jaeger-agent} Unhealthy Readiness probe failed: Get "http://10.128.2.79:14271/": dial tcp 10.128.2.79:14271: connect: connection refused kubelet logger.go:42: 08:05:09 | production | 2025-02-03 08:04:42 +0000 UTC Normal Pod production-ui-query-cb9bcc6b7-4dfzm Binding Scheduled Successfully assigned kuttl-test-fitting-kite/production-ui-query-cb9bcc6b7-4dfzm to ip-10-0-66-98.us-east-2.compute.internal default-scheduler logger.go:42: 08:05:09 | production | 2025-02-03 08:04:42 +0000 UTC Normal ReplicaSet.apps production-ui-query-cb9bcc6b7 SuccessfulCreate Created pod: production-ui-query-cb9bcc6b7-4dfzm replicaset-controller logger.go:42: 08:05:09 | production | 2025-02-03 08:04:42 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled up replica set production-ui-query-cb9bcc6b7 to 1 deployment-controller logger.go:42: 08:05:09 | production | 2025-02-03 08:04:43 +0000 UTC Normal Pod production-ui-query-cb9bcc6b7-4dfzm AddedInterface Add eth0 [10.129.2.84/23] from ovn-kubernetes multus logger.go:42: 08:05:09 | production | 2025-02-03 08:04:43 +0000 UTC Normal Pod production-ui-query-cb9bcc6b7-4dfzm.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:84020ce29bcb5a4bd018e6596188ae919c5cd600e08f78a546c0e76ea477685e" already present on machine kubelet logger.go:42: 08:05:09 | production | 2025-02-03 08:04:43 +0000 UTC Normal Pod production-ui-query-cb9bcc6b7-4dfzm.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 08:05:09 | production | 2025-02-03 08:04:43 +0000 UTC Normal Pod production-ui-query-cb9bcc6b7-4dfzm.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 08:05:09 | production | 2025-02-03 08:04:43 +0000 UTC Normal Pod production-ui-query-cb9bcc6b7-4dfzm.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:65342c7d622f336741a8ae0b9dead79c7ecabc155d02cdd42b7b49ca36680e74" already present on machine kubelet logger.go:42: 08:05:09 | production | 2025-02-03 08:04:43 +0000 UTC Normal Pod production-ui-query-cb9bcc6b7-4dfzm.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 08:05:09 | production | 2025-02-03 08:04:43 +0000 UTC Normal Pod production-ui-query-cb9bcc6b7-4dfzm.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 08:05:09 | production | Deleting namespace: kuttl-test-fitting-kite === CONT kuttl/harness/artifacts logger.go:42: 08:05:16 | artifacts | Creating namespace: kuttl-test-proper-rooster logger.go:42: 08:05:16 | artifacts | artifacts events from ns kuttl-test-proper-rooster: logger.go:42: 08:05:16 | artifacts | Deleting namespace: kuttl-test-proper-rooster === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (192.40s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/allinone (60.26s) --- PASS: kuttl/harness/production (125.90s) --- PASS: kuttl/harness/artifacts (6.18s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name ui --report --output /logs/artifacts/ui.xml ./artifacts/kuttl-report.xml time="2025-02-03T08:05:23Z" level=debug msg="Setting a new name for the test suites" time="2025-02-03T08:05:23Z" level=debug msg="Removing 'artifacts' TestCase" time="2025-02-03T08:05:23Z" level=debug msg="normalizing test case names" time="2025-02-03T08:05:23Z" level=debug msg="ui/allinone -> ui_allinone" time="2025-02-03T08:05:23Z" level=debug msg="ui/production -> ui_production" time="2025-02-03T08:05:23Z" level=debug msg="ui/artifacts -> ui_artifacts" +---------------+--------+ | NAME | RESULT | +---------------+--------+ | ui_allinone | passed | | ui_production | passed | | ui_artifacts | passed | +---------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh upgrade false true + '[' 3 -ne 3 ']' + test_suite_name=upgrade + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/upgrade.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-upgrade make[2]: Entering directory '/tmp/jaeger-tests' make docker JAEGER_VERSION=1.62.1 IMG="quay.io//jaeger-operator:next" make[3]: Entering directory '/tmp/jaeger-tests' [ ! -z "true" ] || docker build --build-arg=GOPROXY= --build-arg=VERSION="1.62.0" --build-arg=JAEGER_VERSION=1.62.1 --build-arg=TARGETARCH= --build-arg VERSION_DATE=2025-02-03T08:05:23Z --build-arg VERSION_PKG="github.com/jaegertracing/jaeger-operator/pkg/version" -t "quay.io//jaeger-operator:next" . make[3]: Leaving directory '/tmp/jaeger-tests' touch build-e2e-upgrade-image SKIP_ES_EXTERNAL=true IMG=quay.io//jaeger-operator:"1.62.0" JAEGER_OPERATOR_VERSION="1.62.0" JAEGER_VERSION="1.62.0" ./tests/e2e/upgrade/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.18.0-0.nightly-2025-02-01-185117 True False 56m Cluster version is 4.18.0-0.nightly-2025-02-01-185117' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.18.0-0.nightly-2025-02-01-185117 True False 56m Cluster version is 4.18.0-0.nightly-2025-02-01-185117' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/upgrade/render.sh ++ export SUITE_DIR=./tests/e2e/upgrade ++ SUITE_DIR=./tests/e2e/upgrade ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/upgrade ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + export JAEGER_NAME + '[' true = true ']' + skip_test upgrade 'Test not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=upgrade + message='Test not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/upgrade/_build + '[' _build '!=' _build ']' + rm -rf upgrade + warning 'upgrade: Test not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: upgrade: Test not supported in OpenShift\e[0m' WAR: upgrade: Test not supported in OpenShift + '[' true = true ']' + skip_test upgrade-from-latest-release 'Test not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=upgrade-from-latest-release + message='Test not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/upgrade/_build + '[' _build '!=' _build ']' + rm -rf upgrade-from-latest-release + warning 'upgrade-from-latest-release: Test not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: upgrade-from-latest-release: Test not supported in OpenShift\e[0m' WAR: upgrade-from-latest-release: Test not supported in OpenShift make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running upgrade E2E tests' Running upgrade E2E tests + cd tests/e2e/upgrade/_build + set +e + KUBECONFIG=/tmp/kubeconfig-2531963025 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-d102g9k0-f1a62.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 1 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === CONT kuttl/harness/artifacts logger.go:42: 08:05:24 | artifacts | Creating namespace: kuttl-test-pure-dragon logger.go:42: 08:05:24 | artifacts | artifacts events from ns kuttl-test-pure-dragon: logger.go:42: 08:05:24 | artifacts | Deleting namespace: kuttl-test-pure-dragon === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (6.32s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (6.26s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name upgrade --report --output /logs/artifacts/upgrade.xml ./artifacts/kuttl-report.xml time="2025-02-03T08:05:31Z" level=debug msg="Setting a new name for the test suites" time="2025-02-03T08:05:31Z" level=debug msg="Removing 'artifacts' TestCase" time="2025-02-03T08:05:31Z" level=debug msg="normalizing test case names" time="2025-02-03T08:05:31Z" level=debug msg="upgrade/artifacts -> upgrade_artifacts" +-------------------+--------+ | NAME | RESULT | +-------------------+--------+ | upgrade_artifacts | passed | +-------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests'