% Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 3831 100 3831 0 0 24757 0 --:--:-- --:--:-- --:--:-- 24876 % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 9315 100 9315 0 0 45743 0 --:--:-- --:--:-- --:--:-- 45886 % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 2607 100 2607 0 0 9863 0 --:--:-- --:--:-- --:--:-- 9837 % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 1953 100 1953 0 0 13145 0 --:--:-- --:--:-- --:--:-- 13195 % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 350 100 350 0 0 2820 0 --:--:-- --:--:-- --:--:-- 2845 Installing kuttl Try 0... curl -sLo /tmp/jaeger-tests/hack/install/../../bin/kubectl-kuttl https://github.com/kudobuilder/kuttl/releases/download/v0.15.0/kubectl-kuttl_0.15.0_linux_x86_64 KUBECONFIG file is: /tmp/kubeconfig-2030369819 for suite in elasticsearch examples generate miscellaneous sidecar streaming ui upgrade; do \ make run-e2e-tests-$suite ; \ done make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh elasticsearch false true + '[' 3 -ne 3 ']' + test_suite_name=elasticsearch + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/elasticsearch.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-elasticsearch make[2]: Entering directory '/tmp/jaeger-tests' >>>> Elasticsearch image not loaded because SKIP_ES_EXTERNAL is true SKIP_ES_EXTERNAL=true \ KAFKA_VERSION=3.6.0 \ SKIP_KAFKA=false \ ./tests/e2e/elasticsearch/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.16.8 True False 11m Cluster version is 4.16.8' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.16.8 True False 11m Cluster version is 4.16.8' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/elasticsearch/render.sh ++ export SUITE_DIR=./tests/e2e/elasticsearch ++ SUITE_DIR=./tests/e2e/elasticsearch ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/elasticsearch ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + is_secured=false + '[' true = true ']' + is_secured=true + start_test es-from-aio-to-production + '[' 1 -ne 1 ']' + test_name=es-from-aio-to-production + echo =========================================================================== =========================================================================== + info 'Rendering files for test es-from-aio-to-production' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test es-from-aio-to-production\e[0m' Rendering files for test es-from-aio-to-production + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build + '[' _build '!=' _build ']' + mkdir -p es-from-aio-to-production + cd es-from-aio-to-production + jaeger_name=my-jaeger + render_install_jaeger my-jaeger allInOne 00 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=allInOne + test_step=00 + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test my-jaeger true 01 + '[' 3 -ne 3 ']' + jaeger=my-jaeger + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + jaeger_deploy_mode=production + [[ true = true ]] + [[ true = true ]] + jaeger_deploy_mode=production_autoprovisioned + render_install_jaeger my-jaeger production_autoprovisioned 03 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=production_autoprovisioned + test_step=03 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./03-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./03-assert.yaml + [[ true = true ]] + [[ true = true ]] + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.elasticsearch.redundancyPolicy="ZeroRedundancy"' ./03-install.yaml + render_smoke_test my-jaeger true 04 + '[' 3 -ne 3 ']' + jaeger=my-jaeger + is_secured=true + test_step=04 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./04-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./04-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test es-increasing-replicas + '[' 1 -ne 1 ']' + test_name=es-increasing-replicas + echo =========================================================================== =========================================================================== + info 'Rendering files for test es-increasing-replicas' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test es-increasing-replicas\e[0m' Rendering files for test es-increasing-replicas + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build/es-from-aio-to-production + '[' es-from-aio-to-production '!=' _build ']' + cd .. + mkdir -p es-increasing-replicas + cd es-increasing-replicas + jaeger_name=simple-prod + '[' true = true ']' + jaeger_deployment_mode=production_autoprovisioned + render_install_jaeger simple-prod production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=simple-prod + JAEGER_NAME=simple-prod + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + cp ./01-install.yaml ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.collector.replicas=2 ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.query.replicas=2 ./02-install.yaml + cp ./01-assert.yaml ./02-assert.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.replicas=2 ./02-assert.yaml + /tmp/jaeger-tests/bin/yq e -i .status.readyReplicas=2 ./02-assert.yaml + render_smoke_test simple-prod true 03 + '[' 3 -ne 3 ']' + jaeger=simple-prod + is_secured=true + test_step=03 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + export JAEGER_NAME=simple-prod + JAEGER_NAME=simple-prod + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./03-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./03-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + '[' true = true ']' + cp ./02-install.yaml ./04-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.storage.elasticsearch.nodeCount=2 ./04-install.yaml + /tmp/jaeger-tests/bin/gomplate -f ./openshift-check-es-nodes.yaml.template -o ./05-check-es-nodes.yaml + '[' true = true ']' + skip_test es-index-cleaner-upstream 'SKIP_ES_EXTERNAL is true' + '[' 2 -ne 2 ']' + test_name=es-index-cleaner-upstream + message='SKIP_ES_EXTERNAL is true' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build/es-increasing-replicas + '[' es-increasing-replicas '!=' _build ']' + cd .. + rm -rf es-index-cleaner-upstream + warning 'es-index-cleaner-upstream: SKIP_ES_EXTERNAL is true' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: es-index-cleaner-upstream: SKIP_ES_EXTERNAL is true\e[0m' WAR: es-index-cleaner-upstream: SKIP_ES_EXTERNAL is true + '[' true = true ']' + es_index_cleaner -autoprov production_autoprovisioned + '[' 2 -ne 2 ']' + postfix=-autoprov + jaeger_deployment_strategy=production_autoprovisioned + start_test es-index-cleaner-autoprov + '[' 1 -ne 1 ']' + test_name=es-index-cleaner-autoprov + echo =========================================================================== =========================================================================== + info 'Rendering files for test es-index-cleaner-autoprov' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test es-index-cleaner-autoprov\e[0m' Rendering files for test es-index-cleaner-autoprov + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build + '[' _build '!=' _build ']' + mkdir -p es-index-cleaner-autoprov + cd es-index-cleaner-autoprov + jaeger_name=test-es-index-cleaner-with-prefix + cronjob_name=test-es-index-cleaner-with-prefix-es-index-cleaner + secured_es_connection=false + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_managed_es ']' + ELASTICSEARCH_URL=https://elasticsearch + secured_es_connection=true + cp ../../es-index-cleaner-upstream/04-assert.yaml ../../es-index-cleaner-upstream/README.md . + render_install_jaeger test-es-index-cleaner-with-prefix production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=test-es-index-cleaner-with-prefix + JAEGER_NAME=test-es-index-cleaner-with-prefix + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.options.es.index-prefix=""' ./01-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.storage.esIndexCleaner.enabled=false ./01-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.storage.esIndexCleaner.numberOfDays=0 ./01-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.esIndexCleaner.schedule="*/1 * * * *"' ./01-install.yaml + render_report_spans test-es-index-cleaner-with-prefix true 5 00 true 02 + '[' 6 -ne 6 ']' + jaeger=test-es-index-cleaner-with-prefix + is_secured=true + number_of_spans=5 + job_number=00 + ensure_reported_spans=true + test_step=02 + export JAEGER_NAME=test-es-index-cleaner-with-prefix + JAEGER_NAME=test-es-index-cleaner-with-prefix + export JAEGER_COLLECTOR_ENDPOINT=http://test-es-index-cleaner-with-prefix-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://test-es-index-cleaner-with-prefix-collector-headless:14268 + export JOB_NUMBER=00 + JOB_NUMBER=00 + export DAYS=5 + DAYS=5 + '[' true = true ']' + protocol=https:// + query_port= + template=/tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template + '[' true = true ']' + export ENSURE_REPORTED_SPANS=true + ENSURE_REPORTED_SPANS=true + export JAEGER_QUERY_ENDPOINT=https://test-es-index-cleaner-with-prefix-query + JAEGER_QUERY_ENDPOINT=https://test-es-index-cleaner-with-prefix-query + params= + '[' true = true ']' + '[' true = true ']' + '[' '' '!=' allInOne ']' + params='-t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template -t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template -o ./02-report-spans.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-report-spans.yaml.template -o ./02-assert.yaml + unset JAEGER_COLLECTOR_ENDPOINT + unset JAEGER_QUERY_ENDPOINT + unset JOB_NUMBER + unset DAYS + unset ENSURE_REPORTED_SPANS + sed 's~enabled: false~enabled: true~gi' ./01-install.yaml + CRONJOB_NAME=test-es-index-cleaner-with-prefix-es-index-cleaner + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/wait-for-cronjob-execution.yaml.template -o ./04-wait-es-index-cleaner.yaml + /tmp/jaeger-tests/bin/gomplate -f ./01-install.yaml -o ./05-install.yaml + render_check_indices true ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' 00 06 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + job_number=00 + test_step=06 + escape_command ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + '[' 1 -ne 1 ']' + command=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' ++ echo ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=test-es-index-cleaner-with-prefix-curator + JOB_NUMBER=00 + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + MOUNT_SECRET=test-es-index-cleaner-with-prefix-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./06-check-indices.yaml + JOB_NUMBER=00 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./06-assert.yaml + '[' true = true ']' + get_elasticsearch_openshift_operator_version + export ESO_OPERATOR_VERSION + '[' true = true ']' ++ kubectl get pods -l name=elasticsearch-operator --all-namespaces '-o=jsonpath={.items[0].metadata.annotations.operatorframework\.io/properties}' error: error executing jsonpath "{.items[0].metadata.annotations.operatorframework\\.io/properties}": Error executing template: array index out of bounds: index 0, length 0. Printing more information for debugging the template: template was: {.items[0].metadata.annotations.operatorframework\.io/properties} object given to jsonpath engine was: map[string]interface {}{"apiVersion":"v1", "items":[]interface {}{}, "kind":"List", "metadata":map[string]interface {}{"resourceVersion":""}} + properties= make[2]: *** [tests/e2e/elasticsearch/Makefile:2: render-e2e-tests-elasticsearch] Error 1 make[2]: Leaving directory '/tmp/jaeger-tests' make[1]: *** [tests/e2e/elasticsearch/Makefile:8: run-e2e-tests-elasticsearch] Error 2 make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh examples false true + '[' 3 -ne 3 ']' + test_suite_name=examples + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/examples.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-examples make[2]: Entering directory '/tmp/jaeger-tests' >>>> Elasticsearch image not loaded because SKIP_ES_EXTERNAL is true KAFKA_VERSION=3.6.0 \ SKIP_KAFKA=false \ VERTX_IMG=jaegertracing/vertx-create-span:operator-e2e-tests \ ./tests/e2e/examples/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.16.8 True False 11m Cluster version is 4.16.8' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.16.8 True False 11m Cluster version is 4.16.8' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/examples/render.sh ++ export SUITE_DIR=./tests/e2e/examples ++ SUITE_DIR=./tests/e2e/examples ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/examples ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + start_test examples-agent-with-priority-class + '[' 1 -ne 1 ']' + test_name=examples-agent-with-priority-class + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-agent-with-priority-class' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-agent-with-priority-class\e[0m' Rendering files for test examples-agent-with-priority-class + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build + '[' _build '!=' _build ']' + mkdir -p examples-agent-with-priority-class + cd examples-agent-with-priority-class + example_name=agent-with-priority-class + prepare_daemonset 00 + '[' 1 -ne 1 ']' + test_step=00 + '[' true = true ']' + cat /tmp/jaeger-tests/examples/openshift/hostport-scc-daemonset.yaml + echo --- + cat /tmp/jaeger-tests/examples/openshift/service_account_jaeger-agent-daemonset.yaml + '[' true '!=' true ']' + render_install_example agent-with-priority-class 02 + '[' 2 -ne 2 ']' + example_name=agent-with-priority-class + test_step=02 + install_file=./02-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/agent-with-priority-class.yaml -o ./02-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./02-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./02-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./02-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./02-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./02-install.yaml ++ jaeger_name=agent-as-daemonset ++ '[' -z agent-as-daemonset ']' ++ echo agent-as-daemonset ++ return 0 + JAEGER_NAME=agent-as-daemonset + local jaeger_strategy ++ get_jaeger_strategy ./02-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./02-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./02-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./02-install.yaml ++ strategy=DaemonSet ++ '[' DaemonSet = null ']' ++ echo DaemonSet ++ return 0 + jaeger_strategy=DaemonSet + '[' DaemonSet = DaemonSet ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./02-assert.yaml + render_smoke_test_example agent-with-priority-class 02 + '[' 2 -ne 2 ']' + example_name=agent-with-priority-class + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/agent-with-priority-class.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/agent-with-priority-class.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/agent-with-priority-class.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/agent-with-priority-class.yaml ++ jaeger_name=agent-as-daemonset ++ '[' -z agent-as-daemonset ']' ++ echo agent-as-daemonset ++ return 0 + jaeger_name=agent-as-daemonset + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test agent-as-daemonset true 02 + '[' 3 -ne 3 ']' + jaeger=agent-as-daemonset + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://agent-as-daemonset-query:443 + JAEGER_QUERY_ENDPOINT=https://agent-as-daemonset-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://agent-as-daemonset-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://agent-as-daemonset-collector-headless:14268 + export JAEGER_NAME=agent-as-daemonset + JAEGER_NAME=agent-as-daemonset + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-all-in-one-with-options + '[' 1 -ne 1 ']' + test_name=examples-all-in-one-with-options + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-all-in-one-with-options' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-all-in-one-with-options\e[0m' Rendering files for test examples-all-in-one-with-options + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-agent-with-priority-class + '[' examples-agent-with-priority-class '!=' _build ']' + cd .. + mkdir -p examples-all-in-one-with-options + cd examples-all-in-one-with-options + example_name=all-in-one-with-options + render_install_example all-in-one-with-options 00 + '[' 2 -ne 2 ']' + example_name=all-in-one-with-options + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/all-in-one-with-options.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=my-jaeger ++ '[' -z my-jaeger ']' ++ echo my-jaeger ++ return 0 + JAEGER_NAME=my-jaeger + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=allInOne ++ '[' allInOne = production ']' ++ '[' allInOne = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + /tmp/jaeger-tests/bin/yq e -i '.metadata.name="my-jaeger"' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i 'del(.spec.allInOne.image)' ./00-install.yaml + render_smoke_test_example all-in-one-with-options 01 + '[' 2 -ne 2 ']' + example_name=all-in-one-with-options + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/all-in-one-with-options.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/all-in-one-with-options.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/all-in-one-with-options.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/all-in-one-with-options.yaml ++ jaeger_name=my-jaeger ++ '[' -z my-jaeger ']' ++ echo my-jaeger ++ return 0 + jaeger_name=my-jaeger + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test my-jaeger true 01 + '[' 3 -ne 3 ']' + jaeger=my-jaeger + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + '[' true = true ']' + sed -i s~my-jaeger-query:443~my-jaeger-query:443/jaeger~gi ./01-smoke-test.yaml + start_test examples-business-application-injected-sidecar + '[' 1 -ne 1 ']' + test_name=examples-business-application-injected-sidecar + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-business-application-injected-sidecar' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-business-application-injected-sidecar\e[0m' Rendering files for test examples-business-application-injected-sidecar + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-all-in-one-with-options + '[' examples-all-in-one-with-options '!=' _build ']' + cd .. + mkdir -p examples-business-application-injected-sidecar + cd examples-business-application-injected-sidecar + example_name=simplest + cp /tmp/jaeger-tests/examples/business-application-injected-sidecar.yaml ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].image=strenv(VERTX_IMG)' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.httpGet.path="/"' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.httpGet.port=8080' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.initialDelaySeconds=1' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.failureThreshold=3' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.periodSeconds=10' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.successThreshold=1' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.timeoutSeconds=1' ./00-install.yaml + render_install_example simplest 01 + '[' 2 -ne 2 ']' + example_name=simplest + test_step=01 + install_file=./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/simplest.yaml -o ./01-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./01-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./01-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./01-install.yaml ++ jaeger_name=simplest ++ '[' -z simplest ']' ++ echo simplest ++ return 0 + JAEGER_NAME=simplest + local jaeger_strategy ++ get_jaeger_strategy ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./01-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./01-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./01-assert.yaml + render_smoke_test_example simplest 02 + '[' 2 -ne 2 ']' + example_name=simplest + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/simplest.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/simplest.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/simplest.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/simplest.yaml ++ jaeger_name=simplest ++ '[' -z simplest ']' ++ echo simplest ++ return 0 + jaeger_name=simplest + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test simplest true 02 + '[' 3 -ne 3 ']' + jaeger=simplest + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simplest-query:443 + JAEGER_QUERY_ENDPOINT=https://simplest-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 + export JAEGER_NAME=simplest + JAEGER_NAME=simplest + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-collector-with-priority-class + '[' 1 -ne 1 ']' + test_name=examples-collector-with-priority-class + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-collector-with-priority-class' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-collector-with-priority-class\e[0m' Rendering files for test examples-collector-with-priority-class + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-business-application-injected-sidecar + '[' examples-business-application-injected-sidecar '!=' _build ']' + cd .. + mkdir -p examples-collector-with-priority-class + cd examples-collector-with-priority-class + example_name=collector-with-priority-class + render_install_example collector-with-priority-class 00 + '[' 2 -ne 2 ']' + example_name=collector-with-priority-class + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/collector-with-priority-class.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=collector-with-high-priority ++ '[' -z collector-with-high-priority ']' ++ echo collector-with-high-priority ++ return 0 + JAEGER_NAME=collector-with-high-priority + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test_example collector-with-priority-class 01 + '[' 2 -ne 2 ']' + example_name=collector-with-priority-class + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/collector-with-priority-class.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/collector-with-priority-class.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/collector-with-priority-class.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/collector-with-priority-class.yaml ++ jaeger_name=collector-with-high-priority ++ '[' -z collector-with-high-priority ']' ++ echo collector-with-high-priority ++ return 0 + jaeger_name=collector-with-high-priority + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test collector-with-high-priority true 01 + '[' 3 -ne 3 ']' + jaeger=collector-with-high-priority + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://collector-with-high-priority-query:443 + JAEGER_QUERY_ENDPOINT=https://collector-with-high-priority-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://collector-with-high-priority-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://collector-with-high-priority-collector-headless:14268 + export JAEGER_NAME=collector-with-high-priority + JAEGER_NAME=collector-with-high-priority + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-service-types + '[' 1 -ne 1 ']' + test_name=examples-service-types + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-service-types' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-service-types\e[0m' Rendering files for test examples-service-types + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-collector-with-priority-class + '[' examples-collector-with-priority-class '!=' _build ']' + cd .. + mkdir -p examples-service-types + cd examples-service-types + example_name=service-types + render_install_example service-types 00 + '[' 2 -ne 2 ']' + example_name=service-types + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/service-types.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=service-types ++ '[' -z service-types ']' ++ echo service-types ++ return 0 + JAEGER_NAME=service-types + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test_example service-types 01 + '[' 2 -ne 2 ']' + example_name=service-types + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/service-types.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/service-types.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/service-types.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/service-types.yaml ++ jaeger_name=service-types ++ '[' -z service-types ']' ++ echo service-types ++ return 0 + jaeger_name=service-types + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test service-types true 01 + '[' 3 -ne 3 ']' + jaeger=service-types + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://service-types-query:443 + JAEGER_QUERY_ENDPOINT=https://service-types-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://service-types-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://service-types-collector-headless:14268 + export JAEGER_NAME=service-types + JAEGER_NAME=service-types + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-simple-prod + '[' 1 -ne 1 ']' + test_name=examples-simple-prod + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-simple-prod' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-simple-prod\e[0m' Rendering files for test examples-simple-prod + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-service-types + '[' examples-service-types '!=' _build ']' + cd .. + mkdir -p examples-simple-prod + cd examples-simple-prod + example_name=simple-prod + render_install_example simple-prod 01 + '[' 2 -ne 2 ']' + example_name=simple-prod + test_step=01 + install_file=./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/simple-prod.yaml -o ./01-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./01-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./01-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./01-install.yaml ++ jaeger_name=simple-prod ++ '[' -z simple-prod ']' ++ echo simple-prod ++ return 0 + JAEGER_NAME=simple-prod + local jaeger_strategy ++ get_jaeger_strategy ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./01-install.yaml ++ strategy=production ++ '[' production = production ']' ++ echo production ++ return 0 + jaeger_strategy=production + '[' production = DaemonSet ']' + '[' production = allInOne ']' + '[' production = production ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + [[ true = true ]] + [[ true = true ]] + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.options={}' ./01-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.elasticsearch={"nodeCount":1,"resources":{"limits":{"memory":"2Gi"}}}' ./01-install.yaml + render_smoke_test_example simple-prod 02 + '[' 2 -ne 2 ']' + example_name=simple-prod + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/simple-prod.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/simple-prod.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/simple-prod.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/simple-prod.yaml ++ jaeger_name=simple-prod ++ '[' -z simple-prod ']' ++ echo simple-prod ++ return 0 + jaeger_name=simple-prod + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test simple-prod true 02 + '[' 3 -ne 3 ']' + jaeger=simple-prod + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + export JAEGER_NAME=simple-prod + JAEGER_NAME=simple-prod + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-simple-prod-with-volumes + '[' 1 -ne 1 ']' + test_name=examples-simple-prod-with-volumes + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-simple-prod-with-volumes' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-simple-prod-with-volumes\e[0m' Rendering files for test examples-simple-prod-with-volumes + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-simple-prod + '[' examples-simple-prod '!=' _build ']' + cd .. + mkdir -p examples-simple-prod-with-volumes + cd examples-simple-prod-with-volumes + example_name=simple-prod-with-volumes + render_install_example simple-prod-with-volumes 01 + '[' 2 -ne 2 ']' + example_name=simple-prod-with-volumes + test_step=01 + install_file=./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/simple-prod-with-volumes.yaml -o ./01-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./01-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./01-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./01-install.yaml ++ jaeger_name=simple-prod ++ '[' -z simple-prod ']' ++ echo simple-prod ++ return 0 + JAEGER_NAME=simple-prod + local jaeger_strategy ++ get_jaeger_strategy ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./01-install.yaml ++ strategy=production ++ '[' production = production ']' ++ echo production ++ return 0 + jaeger_strategy=production + '[' production = DaemonSet ']' + '[' production = allInOne ']' + '[' production = production ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + [[ true = true ]] + [[ true = true ]] + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.options={}' ./01-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.elasticsearch={"nodeCount":1,"resources":{"limits":{"memory":"2Gi"}}}' ./01-install.yaml + render_smoke_test_example simple-prod-with-volumes 02 + '[' 2 -ne 2 ']' + example_name=simple-prod-with-volumes + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/simple-prod-with-volumes.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/simple-prod-with-volumes.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/simple-prod-with-volumes.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/simple-prod-with-volumes.yaml ++ jaeger_name=simple-prod ++ '[' -z simple-prod ']' ++ echo simple-prod ++ return 0 + jaeger_name=simple-prod + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test simple-prod true 02 + '[' 3 -ne 3 ']' + jaeger=simple-prod + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + export JAEGER_NAME=simple-prod + JAEGER_NAME=simple-prod + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + /tmp/jaeger-tests/bin/gomplate -f ./03-check-volume.yaml.template -o 03-check-volume.yaml + start_test examples-simplest + '[' 1 -ne 1 ']' + test_name=examples-simplest + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-simplest' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-simplest\e[0m' Rendering files for test examples-simplest + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-simple-prod-with-volumes + '[' examples-simple-prod-with-volumes '!=' _build ']' + cd .. + mkdir -p examples-simplest + cd examples-simplest + example_name=simplest + render_install_example simplest 00 + '[' 2 -ne 2 ']' + example_name=simplest + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/simplest.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=simplest ++ '[' -z simplest ']' ++ echo simplest ++ return 0 + JAEGER_NAME=simplest + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test_example simplest 01 + '[' 2 -ne 2 ']' + example_name=simplest + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/simplest.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/simplest.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/simplest.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/simplest.yaml ++ jaeger_name=simplest ++ '[' -z simplest ']' ++ echo simplest ++ return 0 + jaeger_name=simplest + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test simplest true 01 + '[' 3 -ne 3 ']' + jaeger=simplest + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simplest-query:443 + JAEGER_QUERY_ENDPOINT=https://simplest-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 + export JAEGER_NAME=simplest + JAEGER_NAME=simplest + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-with-badger + '[' 1 -ne 1 ']' + test_name=examples-with-badger + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-with-badger' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-with-badger\e[0m' Rendering files for test examples-with-badger + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-simplest + '[' examples-simplest '!=' _build ']' + cd .. + mkdir -p examples-with-badger + cd examples-with-badger + example_name=with-badger + render_install_example with-badger 00 + '[' 2 -ne 2 ']' + example_name=with-badger + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/with-badger.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=with-badger ++ '[' -z with-badger ']' ++ echo with-badger ++ return 0 + JAEGER_NAME=with-badger + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test_example with-badger 01 + '[' 2 -ne 2 ']' + example_name=with-badger + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/with-badger.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/with-badger.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/with-badger.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/with-badger.yaml ++ jaeger_name=with-badger ++ '[' -z with-badger ']' ++ echo with-badger ++ return 0 + jaeger_name=with-badger + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test with-badger true 01 + '[' 3 -ne 3 ']' + jaeger=with-badger + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://with-badger-query:443 + JAEGER_QUERY_ENDPOINT=https://with-badger-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://with-badger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://with-badger-collector-headless:14268 + export JAEGER_NAME=with-badger + JAEGER_NAME=with-badger + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-with-badger-and-volume + '[' 1 -ne 1 ']' + test_name=examples-with-badger-and-volume + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-with-badger-and-volume' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-with-badger-and-volume\e[0m' Rendering files for test examples-with-badger-and-volume + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-with-badger + '[' examples-with-badger '!=' _build ']' + cd .. + mkdir -p examples-with-badger-and-volume + cd examples-with-badger-and-volume + example_name=with-badger-and-volume + render_install_example with-badger-and-volume 00 + '[' 2 -ne 2 ']' + example_name=with-badger-and-volume + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/with-badger-and-volume.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=with-badger-and-volume ++ '[' -z with-badger-and-volume ']' ++ echo with-badger-and-volume ++ return 0 + JAEGER_NAME=with-badger-and-volume + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test_example with-badger-and-volume 01 + '[' 2 -ne 2 ']' + example_name=with-badger-and-volume + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/with-badger-and-volume.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/with-badger-and-volume.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/with-badger-and-volume.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/with-badger-and-volume.yaml ++ jaeger_name=with-badger-and-volume ++ '[' -z with-badger-and-volume ']' ++ echo with-badger-and-volume ++ return 0 + jaeger_name=with-badger-and-volume + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test with-badger-and-volume true 01 + '[' 3 -ne 3 ']' + jaeger=with-badger-and-volume + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://with-badger-and-volume-query:443 + JAEGER_QUERY_ENDPOINT=https://with-badger-and-volume-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://with-badger-and-volume-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://with-badger-and-volume-collector-headless:14268 + export JAEGER_NAME=with-badger-and-volume + JAEGER_NAME=with-badger-and-volume + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-with-cassandra + '[' 1 -ne 1 ']' + test_name=examples-with-cassandra + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-with-cassandra' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-with-cassandra\e[0m' Rendering files for test examples-with-cassandra + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-with-badger-and-volume + '[' examples-with-badger-and-volume '!=' _build ']' + cd .. + mkdir -p examples-with-cassandra + cd examples-with-cassandra + example_name=with-cassandra + render_install_cassandra 00 + '[' 1 -ne 1 ']' + test_step=00 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/cassandra-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/cassandra-assert.yaml.template -o ./00-assert.yaml + render_install_example with-cassandra 01 + '[' 2 -ne 2 ']' + example_name=with-cassandra + test_step=01 + install_file=./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/with-cassandra.yaml -o ./01-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./01-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./01-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./01-install.yaml ++ jaeger_name=with-cassandra ++ '[' -z with-cassandra ']' ++ echo with-cassandra ++ return 0 + JAEGER_NAME=with-cassandra + local jaeger_strategy ++ get_jaeger_strategy ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./01-install.yaml ++ strategy=allInOne ++ '[' allInOne = production ']' ++ '[' allInOne = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./01-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./01-assert.yaml + render_smoke_test_example with-cassandra 02 + '[' 2 -ne 2 ']' + example_name=with-cassandra + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/with-cassandra.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/with-cassandra.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/with-cassandra.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/with-cassandra.yaml ++ jaeger_name=with-cassandra ++ '[' -z with-cassandra ']' ++ echo with-cassandra ++ return 0 + jaeger_name=with-cassandra + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test with-cassandra true 02 + '[' 3 -ne 3 ']' + jaeger=with-cassandra + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://with-cassandra-query:443 + JAEGER_QUERY_ENDPOINT=https://with-cassandra-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://with-cassandra-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://with-cassandra-collector-headless:14268 + export JAEGER_NAME=with-cassandra + JAEGER_NAME=with-cassandra + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-with-sampling + '[' 1 -ne 1 ']' + test_name=examples-with-sampling + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-with-sampling' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-with-sampling\e[0m' Rendering files for test examples-with-sampling + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-with-cassandra + '[' examples-with-cassandra '!=' _build ']' + cd .. + mkdir -p examples-with-sampling + cd examples-with-sampling + export example_name=with-sampling + example_name=with-sampling + render_install_cassandra 00 + '[' 1 -ne 1 ']' + test_step=00 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/cassandra-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/cassandra-assert.yaml.template -o ./00-assert.yaml + render_install_example with-sampling 01 + '[' 2 -ne 2 ']' + example_name=with-sampling + test_step=01 + install_file=./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/with-sampling.yaml -o ./01-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./01-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./01-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./01-install.yaml ++ jaeger_name=with-sampling ++ '[' -z with-sampling ']' ++ echo with-sampling ++ return 0 + JAEGER_NAME=with-sampling + local jaeger_strategy ++ get_jaeger_strategy ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./01-install.yaml ++ strategy=allInOne ++ '[' allInOne = production ']' ++ '[' allInOne = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./01-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./01-assert.yaml + render_smoke_test_example with-sampling 02 + '[' 2 -ne 2 ']' + example_name=with-sampling + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/with-sampling.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/with-sampling.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/with-sampling.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/with-sampling.yaml ++ jaeger_name=with-sampling ++ '[' -z with-sampling ']' ++ echo with-sampling ++ return 0 + jaeger_name=with-sampling + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test with-sampling true 02 + '[' 3 -ne 3 ']' + jaeger=with-sampling + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://with-sampling-query:443 + JAEGER_QUERY_ENDPOINT=https://with-sampling-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://with-sampling-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://with-sampling-collector-headless:14268 + export JAEGER_NAME=with-sampling + JAEGER_NAME=with-sampling + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-agent-as-daemonset + '[' 1 -ne 1 ']' + test_name=examples-agent-as-daemonset + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-agent-as-daemonset' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-agent-as-daemonset\e[0m' Rendering files for test examples-agent-as-daemonset + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-with-sampling + '[' examples-with-sampling '!=' _build ']' + cd .. + mkdir -p examples-agent-as-daemonset + cd examples-agent-as-daemonset + '[' true = true ']' + prepare_daemonset 00 + '[' 1 -ne 1 ']' + test_step=00 + '[' true = true ']' + cat /tmp/jaeger-tests/examples/openshift/hostport-scc-daemonset.yaml + echo --- + cat /tmp/jaeger-tests/examples/openshift/service_account_jaeger-agent-daemonset.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/openshift/agent-as-daemonset.yaml -o 02-install.yaml + '[' true = true ']' + start_test examples-openshift-with-htpasswd + '[' 1 -ne 1 ']' + test_name=examples-openshift-with-htpasswd + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-openshift-with-htpasswd' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-openshift-with-htpasswd\e[0m' Rendering files for test examples-openshift-with-htpasswd + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-agent-as-daemonset + '[' examples-agent-as-daemonset '!=' _build ']' + cd .. + mkdir -p examples-openshift-with-htpasswd + cd examples-openshift-with-htpasswd + export JAEGER_NAME=with-htpasswd + JAEGER_NAME=with-htpasswd + export JAEGER_USERNAME=awesomeuser + JAEGER_USERNAME=awesomeuser + export JAEGER_PASSWORD=awesomepassword + JAEGER_PASSWORD=awesomepassword + export 'JAEGER_USER_PASSWORD_HASH=awesomeuser:{SHA}uUdqPVUyqNBmERU0Qxj3KFaZnjw=' + JAEGER_USER_PASSWORD_HASH='awesomeuser:{SHA}uUdqPVUyqNBmERU0Qxj3KFaZnjw=' ++ echo 'awesomeuser:{SHA}uUdqPVUyqNBmERU0Qxj3KFaZnjw=' ++ base64 + SECRET=YXdlc29tZXVzZXI6e1NIQX11VWRxUFZVeXFOQm1FUlUwUXhqM0tGYVpuanc9Cg== + /tmp/jaeger-tests/bin/gomplate -f ./00-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/openshift/with-htpasswd.yaml -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./01-assert.yaml + export 'GET_URL_COMMAND=kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE' + GET_URL_COMMAND='kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE' + export 'URL=https://$(kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE)/search' + URL='https://$(kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE)/search' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/ensure-ingress-host.sh.template -o ./ensure-ingress-host.sh + chmod +x ./ensure-ingress-host.sh + INSECURE=true + JAEGER_USERNAME= + JAEGER_PASSWORD= + EXPECTED_CODE=403 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./02-check-unsecured.yaml + JAEGER_USERNAME=wronguser + JAEGER_PASSWORD=wrongpassword + EXPECTED_CODE=403 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./03-check-unauthorized.yaml + EXPECTED_CODE=200 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./04-check-authorized.yaml + skip_test examples-agent-as-daemonset 'This test is flaky in Prow CI' + '[' 2 -ne 2 ']' + test_name=examples-agent-as-daemonset + message='This test is flaky in Prow CI' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-openshift-with-htpasswd + '[' examples-openshift-with-htpasswd '!=' _build ']' + cd .. + rm -rf examples-agent-as-daemonset + warning 'examples-agent-as-daemonset: This test is flaky in Prow CI' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: examples-agent-as-daemonset: This test is flaky in Prow CI\e[0m' WAR: examples-agent-as-daemonset: This test is flaky in Prow CI + skip_test examples-with-badger-and-volume 'This test is flaky in Prow CI' + '[' 2 -ne 2 ']' + test_name=examples-with-badger-and-volume + message='This test is flaky in Prow CI' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build + '[' _build '!=' _build ']' + rm -rf examples-with-badger-and-volume + warning 'examples-with-badger-and-volume: This test is flaky in Prow CI' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: examples-with-badger-and-volume: This test is flaky in Prow CI\e[0m' WAR: examples-with-badger-and-volume: This test is flaky in Prow CI + skip_test examples-collector-with-priority-class 'This test is flaky in Prow CI' + '[' 2 -ne 2 ']' + test_name=examples-collector-with-priority-class + message='This test is flaky in Prow CI' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build + '[' _build '!=' _build ']' + rm -rf examples-collector-with-priority-class + warning 'examples-collector-with-priority-class: This test is flaky in Prow CI' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: examples-collector-with-priority-class: This test is flaky in Prow CI\e[0m' WAR: examples-collector-with-priority-class: This test is flaky in Prow CI make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running examples E2E tests' Running examples E2E tests + cd tests/e2e/examples/_build + set +e + KUBECONFIG=/tmp/kubeconfig-2030369819 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-rosa-h-yd3i.24xa.s3.devshift.org:443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 12 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/examples-agent-with-priority-class === PAUSE kuttl/harness/examples-agent-with-priority-class === RUN kuttl/harness/examples-all-in-one-with-options === PAUSE kuttl/harness/examples-all-in-one-with-options === RUN kuttl/harness/examples-business-application-injected-sidecar === PAUSE kuttl/harness/examples-business-application-injected-sidecar === RUN kuttl/harness/examples-openshift-with-htpasswd === PAUSE kuttl/harness/examples-openshift-with-htpasswd === RUN kuttl/harness/examples-service-types === PAUSE kuttl/harness/examples-service-types === RUN kuttl/harness/examples-simple-prod === PAUSE kuttl/harness/examples-simple-prod === RUN kuttl/harness/examples-simple-prod-with-volumes === PAUSE kuttl/harness/examples-simple-prod-with-volumes === RUN kuttl/harness/examples-simplest === PAUSE kuttl/harness/examples-simplest === RUN kuttl/harness/examples-with-badger === PAUSE kuttl/harness/examples-with-badger === RUN kuttl/harness/examples-with-cassandra === PAUSE kuttl/harness/examples-with-cassandra === RUN kuttl/harness/examples-with-sampling === PAUSE kuttl/harness/examples-with-sampling === CONT kuttl/harness/artifacts logger.go:42: 19:16:32 | artifacts | Creating namespace: kuttl-test-prime-parrot logger.go:42: 19:16:32 | artifacts | artifacts events from ns kuttl-test-prime-parrot: logger.go:42: 19:16:32 | artifacts | Deleting namespace: kuttl-test-prime-parrot === CONT kuttl/harness/examples-simple-prod logger.go:42: 19:16:38 | examples-simple-prod | Creating namespace: kuttl-test-enjoyed-hare logger.go:42: 19:16:38 | examples-simple-prod/1-install | starting test step 1-install logger.go:42: 19:16:38 | examples-simple-prod/1-install | Jaeger:kuttl-test-enjoyed-hare/simple-prod created logger.go:42: 19:26:39 | examples-simple-prod/1-install | test step failed 1-install case.go:364: failed in step 1-install case.go:366: --- Deployment:kuttl-test-enjoyed-hare/simple-prod-collector +++ Deployment:kuttl-test-enjoyed-hare/simple-prod-collector @@ -1,10 +1,412 @@ apiVersion: apps/v1 kind: Deployment metadata: + annotations: + linkerd.io/inject: disabled + prometheus.io/port: "14269" + prometheus.io/scrape: "true" + labels: + app: jaeger + app.kubernetes.io/component: collector + app.kubernetes.io/instance: simple-prod + app.kubernetes.io/managed-by: jaeger-operator + app.kubernetes.io/name: simple-prod-collector + app.kubernetes.io/part-of: jaeger + managedFields: + - apiVersion: apps/v1 + fieldsType: FieldsV1 + fieldsV1: + f:metadata: + f:annotations: + .: {} + f:linkerd.io/inject: {} + f:prometheus.io/port: {} + f:prometheus.io/scrape: {} + f:labels: + .: {} + f:app: {} + f:app.kubernetes.io/component: {} + f:app.kubernetes.io/instance: {} + f:app.kubernetes.io/managed-by: {} + f:app.kubernetes.io/name: {} + f:app.kubernetes.io/part-of: {} + f:ownerReferences: + .: {} + k:{"uid":"6b093f37-febc-4a17-9a07-7e75f83dac2f"}: {} + f:spec: + f:progressDeadlineSeconds: {} + f:replicas: {} + f:revisionHistoryLimit: {} + f:selector: {} + f:strategy: + f:type: {} + f:template: + f:metadata: + f:annotations: + .: {} + f:linkerd.io/inject: {} + f:prometheus.io/port: {} + f:prometheus.io/scrape: {} + f:sidecar.istio.io/inject: {} + f:labels: + .: {} + f:app: {} + f:app.kubernetes.io/component: {} + f:app.kubernetes.io/instance: {} + f:app.kubernetes.io/managed-by: {} + f:app.kubernetes.io/name: {} + f:app.kubernetes.io/part-of: {} + f:spec: + f:containers: + k:{"name":"jaeger-collector"}: + .: {} + f:args: {} + f:env: + .: {} + k:{"name":"COLLECTOR_OTLP_ENABLED"}: + .: {} + f:name: {} + f:value: {} + k:{"name":"COLLECTOR_ZIPKIN_HOST_PORT"}: + .: {} + f:name: {} + f:value: {} + k:{"name":"SPAN_STORAGE_TYPE"}: + .: {} + f:name: {} + f:value: {} + f:image: {} + f:imagePullPolicy: {} + f:livenessProbe: + .: {} + f:failureThreshold: {} + f:httpGet: + .: {} + f:path: {} + f:port: {} + f:scheme: {} + f:initialDelaySeconds: {} + f:periodSeconds: {} + f:successThreshold: {} + f:timeoutSeconds: {} + f:name: {} + f:ports: + .: {} + k:{"containerPort":4317,"protocol":"TCP"}: + .: {} + f:containerPort: {} + f:name: {} + f:protocol: {} + k:{"containerPort":4318,"protocol":"TCP"}: + .: {} + f:containerPort: {} + f:name: {} + f:protocol: {} + k:{"containerPort":9411,"protocol":"TCP"}: + .: {} + f:containerPort: {} + f:name: {} + f:protocol: {} + k:{"containerPort":14250,"protocol":"TCP"}: + .: {} + f:containerPort: {} + f:name: {} + f:protocol: {} + k:{"containerPort":14267,"protocol":"TCP"}: + .: {} + f:containerPort: {} + f:name: {} + f:protocol: {} + k:{"containerPort":14268,"protocol":"TCP"}: + .: {} + f:containerPort: {} + f:name: {} + f:protocol: {} + k:{"containerPort":14269,"protocol":"TCP"}: + .: {} + f:containerPort: {} + f:name: {} + f:protocol: {} + f:readinessProbe: + .: {} + f:failureThreshold: {} + f:httpGet: + .: {} + f:path: {} + f:port: {} + f:scheme: {} + f:initialDelaySeconds: {} + f:periodSeconds: {} + f:successThreshold: {} + f:timeoutSeconds: {} + f:resources: {} + f:terminationMessagePath: {} + f:terminationMessagePolicy: {} + f:volumeMounts: + .: {} + k:{"mountPath":"/certs"}: + .: {} + f:mountPath: {} + f:name: {} + f:readOnly: {} + k:{"mountPath":"/etc/jaeger/sampling"}: + .: {} + f:mountPath: {} + f:name: {} + f:readOnly: {} + k:{"mountPath":"/etc/pki/ca-trust/extracted/pem"}: + .: {} + f:mountPath: {} + f:name: {} + f:readOnly: {} + k:{"mountPath":"/etc/tls-config"}: + .: {} + f:mountPath: {} + f:name: {} + f:readOnly: {} + f:dnsPolicy: {} + f:enableServiceLinks: {} + f:restartPolicy: {} + f:schedulerName: {} + f:securityContext: {} + f:serviceAccount: {} + f:serviceAccountName: {} + f:terminationGracePeriodSeconds: {} + f:volumes: + .: {} + k:{"name":"certs"}: + .: {} + f:name: {} + f:secret: + .: {} + f:defaultMode: {} + f:secretName: {} + k:{"name":"simple-prod-collector-tls-config-volume"}: + .: {} + f:name: {} + f:secret: + .: {} + f:defaultMode: {} + f:secretName: {} + k:{"name":"simple-prod-sampling-configuration-volume"}: + .: {} + f:configMap: + .: {} + f:defaultMode: {} + f:items: {} + f:name: {} + f:name: {} + k:{"name":"simple-prod-trusted-ca"}: + .: {} + f:configMap: + .: {} + f:defaultMode: {} + f:items: {} + f:name: {} + f:name: {} + manager: jaeger-operator + operation: Update + time: "2024-08-29T19:16:43Z" + - apiVersion: apps/v1 + fieldsType: FieldsV1 + fieldsV1: + f:metadata: + f:annotations: + f:deployment.kubernetes.io/revision: {} + f:status: + f:conditions: + .: {} + k:{"type":"Available"}: + .: {} + f:lastTransitionTime: {} + f:lastUpdateTime: {} + f:message: {} + f:reason: {} + f:status: {} + f:type: {} + k:{"type":"Progressing"}: + .: {} + f:lastTransitionTime: {} + f:lastUpdateTime: {} + f:message: {} + f:reason: {} + f:status: {} + f:type: {} + f:observedGeneration: {} + f:replicas: {} + f:unavailableReplicas: {} + f:updatedReplicas: {} + manager: kube-controller-manager + operation: Update + subresource: status + time: "2024-08-29T19:16:43Z" name: simple-prod-collector namespace: kuttl-test-enjoyed-hare + ownerReferences: + - apiVersion: jaegertracing.io/v1 + controller: true + kind: Jaeger + name: simple-prod + uid: 6b093f37-febc-4a17-9a07-7e75f83dac2f spec: + progressDeadlineSeconds: 600 replicas: 1 + revisionHistoryLimit: 10 + selector: + matchLabels: + app: jaeger + app.kubernetes.io/component: collector + app.kubernetes.io/instance: simple-prod + app.kubernetes.io/managed-by: jaeger-operator + app.kubernetes.io/name: simple-prod-collector + app.kubernetes.io/part-of: jaeger + strategy: + type: Recreate + template: + metadata: + annotations: + linkerd.io/inject: disabled + prometheus.io/port: "14269" + prometheus.io/scrape: "true" + sidecar.istio.io/inject: "false" + creationTimestamp: null + labels: + app: jaeger + app.kubernetes.io/component: collector + app.kubernetes.io/instance: simple-prod + app.kubernetes.io/managed-by: jaeger-operator + app.kubernetes.io/name: simple-prod-collector + app.kubernetes.io/part-of: jaeger + spec: + containers: + - args: + - --collector.grpc.tls.cert=/etc/tls-config/tls.crt + - --collector.grpc.tls.enabled=true + - --collector.grpc.tls.key=/etc/tls-config/tls.key + - --sampling.strategies-file=/etc/jaeger/sampling/sampling.json + - --es.server-urls=https://elasticsearch.kuttl-test-enjoyed-hare.svc.cluster.local:9200 + - --es.tls.enabled=true + - --es.tls.ca=/certs/ca + - --es.tls.cert=/certs/cert + - --es.tls.key=/certs/key + - --es.timeout=15s + - --es.num-shards=1 + - --es.num-replicas=0 + env: + - name: SPAN_STORAGE_TYPE + value: elasticsearch + - name: COLLECTOR_ZIPKIN_HOST_PORT + value: :9411 + - name: COLLECTOR_OTLP_ENABLED + value: "true" + image: registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:b90e0d264bb0b051c2a4b9e6d499f407fed58f834ab8ebf30263ee1cf7c7ea3c + imagePullPolicy: IfNotPresent + livenessProbe: + failureThreshold: 5 + httpGet: + path: / + port: 14269 + scheme: HTTP + initialDelaySeconds: 5 + periodSeconds: 15 + successThreshold: 1 + timeoutSeconds: 1 + name: jaeger-collector + ports: + - containerPort: 9411 + name: zipkin + protocol: TCP + - containerPort: 14267 + name: c-tchan-trft + protocol: TCP + - containerPort: 14268 + name: c-binary-trft + protocol: TCP + - containerPort: 14269 + name: admin-http + protocol: TCP + - containerPort: 14250 + name: grpc + protocol: TCP + - containerPort: 4317 + name: grpc-otlp + protocol: TCP + - containerPort: 4318 + name: http-otlp + protocol: TCP + readinessProbe: + failureThreshold: 3 + httpGet: + path: / + port: 14269 + scheme: HTTP + initialDelaySeconds: 1 + periodSeconds: 10 + successThreshold: 1 + timeoutSeconds: 1 + resources: {} + terminationMessagePath: /dev/termination-log + terminationMessagePolicy: File + volumeMounts: + - mountPath: /etc/jaeger/sampling + name: simple-prod-sampling-configuration-volume + readOnly: true + - mountPath: /etc/tls-config + name: simple-prod-collector-tls-config-volume + readOnly: true + - mountPath: /etc/pki/ca-trust/extracted/pem + name: simple-prod-trusted-ca + readOnly: true + - mountPath: /certs + name: certs + readOnly: true + dnsPolicy: ClusterFirst + enableServiceLinks: false + restartPolicy: Always + schedulerName: default-scheduler + securityContext: {} + serviceAccount: simple-prod + serviceAccountName: simple-prod + terminationGracePeriodSeconds: 30 + volumes: + - configMap: + defaultMode: 420 + items: + - key: sampling + path: sampling.json + name: simple-prod-sampling-configuration + name: simple-prod-sampling-configuration-volume + - name: simple-prod-collector-tls-config-volume + secret: + defaultMode: 420 + secretName: simple-prod-collector-headless-tls + - configMap: + defaultMode: 420 + items: + - key: ca-bundle.crt + path: tls-ca-bundle.pem + name: simple-prod-trusted-ca + name: simple-prod-trusted-ca + - name: certs + secret: + defaultMode: 420 + secretName: simple-prod-jaeger-elasticsearch status: - readyReplicas: 1 + conditions: + - lastTransitionTime: "2024-08-29T19:16:43Z" + lastUpdateTime: "2024-08-29T19:16:43Z" + message: Deployment does not have minimum availability. + reason: MinimumReplicasUnavailable + status: "False" + type: Available + - lastTransitionTime: "2024-08-29T19:16:43Z" + lastUpdateTime: "2024-08-29T19:16:43Z" + message: ReplicaSet "simple-prod-collector-77854cb9f7" is progressing. + reason: ReplicaSetUpdated + status: "True" + type: Progressing + observedGeneration: 1 + replicas: 1 + unavailableReplicas: 1 + updatedReplicas: 1 case.go:366: resource Deployment:kuttl-test-enjoyed-hare/simple-prod-collector: .status.readyReplicas: key is missing from map case.go:366: --- Deployment:kuttl-test-enjoyed-hare/simple-prod-query +++ Deployment:kuttl-test-enjoyed-hare/simple-prod-query @@ -1,10 +1,508 @@ apiVersion: apps/v1 kind: Deployment metadata: + annotations: + linkerd.io/inject: disabled + prometheus.io/port: "16687" + prometheus.io/scrape: "true" + sidecar.jaegertracing.io/inject: simple-prod + sidecar.jaegertracing.io/revision: "1" + labels: + app: jaeger + app.kubernetes.io/component: query + app.kubernetes.io/instance: simple-prod + app.kubernetes.io/managed-by: jaeger-operator + app.kubernetes.io/name: simple-prod-query + app.kubernetes.io/part-of: jaeger + sidecar.jaegertracing.io/injected: simple-prod + managedFields: + - apiVersion: apps/v1 + fieldsType: FieldsV1 + fieldsV1: + f:metadata: + f:annotations: + .: {} + f:linkerd.io/inject: {} + f:prometheus.io/port: {} + f:prometheus.io/scrape: {} + f:sidecar.jaegertracing.io/inject: {} + f:sidecar.jaegertracing.io/revision: {} + f:labels: + .: {} + f:app: {} + f:app.kubernetes.io/component: {} + f:app.kubernetes.io/instance: {} + f:app.kubernetes.io/managed-by: {} + f:app.kubernetes.io/name: {} + f:app.kubernetes.io/part-of: {} + f:ownerReferences: + .: {} + k:{"uid":"6b093f37-febc-4a17-9a07-7e75f83dac2f"}: {} + f:spec: + f:progressDeadlineSeconds: {} + f:replicas: {} + f:revisionHistoryLimit: {} + f:selector: {} + f:strategy: + f:type: {} + f:template: + f:metadata: + f:annotations: + .: {} + f:linkerd.io/inject: {} + f:prometheus.io/port: {} + f:prometheus.io/scrape: {} + f:sidecar.istio.io/inject: {} + f:sidecar.jaegertracing.io/inject: {} + f:labels: + .: {} + f:app: {} + f:app.kubernetes.io/component: {} + f:app.kubernetes.io/instance: {} + f:app.kubernetes.io/managed-by: {} + f:app.kubernetes.io/name: {} + f:app.kubernetes.io/part-of: {} + f:spec: + f:containers: + k:{"name":"jaeger-query"}: + .: {} + f:args: {} + f:env: + .: {} + k:{"name":"JAEGER_DISABLED"}: + .: {} + f:name: {} + f:value: {} + k:{"name":"METRICS_STORAGE_TYPE"}: + .: {} + f:name: {} + k:{"name":"SPAN_STORAGE_TYPE"}: + .: {} + f:name: {} + f:value: {} + f:image: {} + f:imagePullPolicy: {} + f:livenessProbe: + .: {} + f:failureThreshold: {} + f:httpGet: + .: {} + f:path: {} + f:port: {} + f:scheme: {} + f:initialDelaySeconds: {} + f:periodSeconds: {} + f:successThreshold: {} + f:timeoutSeconds: {} + f:name: {} + f:ports: + .: {} + k:{"containerPort":16685,"protocol":"TCP"}: + .: {} + f:containerPort: {} + f:name: {} + f:protocol: {} + k:{"containerPort":16686,"protocol":"TCP"}: + .: {} + f:containerPort: {} + f:name: {} + f:protocol: {} + k:{"containerPort":16687,"protocol":"TCP"}: + .: {} + f:containerPort: {} + f:name: {} + f:protocol: {} + f:readinessProbe: + .: {} + f:failureThreshold: {} + f:httpGet: + .: {} + f:path: {} + f:port: {} + f:scheme: {} + f:initialDelaySeconds: {} + f:periodSeconds: {} + f:successThreshold: {} + f:timeoutSeconds: {} + f:resources: {} + f:terminationMessagePath: {} + f:terminationMessagePolicy: {} + f:volumeMounts: + .: {} + k:{"mountPath":"/certs"}: + .: {} + f:mountPath: {} + f:name: {} + f:readOnly: {} + k:{"mountPath":"/etc/config"}: + .: {} + f:mountPath: {} + f:name: {} + f:readOnly: {} + k:{"mountPath":"/etc/pki/ca-trust/extracted/pem"}: + .: {} + f:mountPath: {} + f:name: {} + f:readOnly: {} + k:{"name":"oauth-proxy"}: + .: {} + f:args: {} + f:image: {} + f:imagePullPolicy: {} + f:name: {} + f:ports: + .: {} + k:{"containerPort":8443,"protocol":"TCP"}: + .: {} + f:containerPort: {} + f:name: {} + f:protocol: {} + f:resources: {} + f:terminationMessagePath: {} + f:terminationMessagePolicy: {} + f:volumeMounts: + .: {} + k:{"mountPath":"/etc/pki/ca-trust/extracted/pem"}: + .: {} + f:mountPath: {} + f:name: {} + f:readOnly: {} + k:{"mountPath":"/etc/tls/private"}: + .: {} + f:mountPath: {} + f:name: {} + f:dnsPolicy: {} + f:enableServiceLinks: {} + f:restartPolicy: {} + f:schedulerName: {} + f:securityContext: {} + f:serviceAccount: {} + f:serviceAccountName: {} + f:terminationGracePeriodSeconds: {} + f:volumes: + .: {} + k:{"name":"certs"}: + .: {} + f:name: {} + f:secret: + .: {} + f:defaultMode: {} + f:secretName: {} + k:{"name":"simple-prod-trusted-ca"}: + .: {} + f:configMap: + .: {} + f:defaultMode: {} + f:items: {} + f:name: {} + f:name: {} + k:{"name":"simple-prod-ui-configuration-volume"}: + .: {} + f:configMap: + .: {} + f:defaultMode: {} + f:items: {} + f:name: {} + f:name: {} + k:{"name":"simple-prod-ui-oauth-proxy-tls"}: + .: {} + f:name: {} + f:secret: + .: {} + f:defaultMode: {} + f:secretName: {} + manager: jaeger-operator + operation: Update + time: "2024-08-29T19:21:44Z" + - apiVersion: apps/v1 + fieldsType: FieldsV1 + fieldsV1: + f:metadata: + f:annotations: + f:deployment.kubernetes.io/revision: {} + f:status: + f:conditions: + .: {} + k:{"type":"Available"}: + .: {} + f:lastTransitionTime: {} + f:lastUpdateTime: {} + f:message: {} + f:reason: {} + f:status: {} + f:type: {} + k:{"type":"Progressing"}: + .: {} + f:lastTransitionTime: {} + f:lastUpdateTime: {} + f:message: {} + f:reason: {} + f:status: {} + f:type: {} + f:observedGeneration: {} + f:replicas: {} + f:unavailableReplicas: {} + f:updatedReplicas: {} + manager: kube-controller-manager + operation: Update + subresource: status + time: "2024-08-29T19:21:44Z" name: simple-prod-query namespace: kuttl-test-enjoyed-hare + ownerReferences: + - apiVersion: jaegertracing.io/v1 + controller: true + kind: Jaeger + name: simple-prod + uid: 6b093f37-febc-4a17-9a07-7e75f83dac2f spec: + progressDeadlineSeconds: 600 replicas: 1 + revisionHistoryLimit: 10 + selector: + matchLabels: + app: jaeger + app.kubernetes.io/component: query + app.kubernetes.io/instance: simple-prod + app.kubernetes.io/managed-by: jaeger-operator + app.kubernetes.io/name: simple-prod-query + app.kubernetes.io/part-of: jaeger + strategy: + type: Recreate + template: + metadata: + annotations: + linkerd.io/inject: disabled + prometheus.io/port: "16687" + prometheus.io/scrape: "true" + sidecar.istio.io/inject: "false" + sidecar.jaegertracing.io/inject: simple-prod + creationTimestamp: null + labels: + app: jaeger + app.kubernetes.io/component: query + app.kubernetes.io/instance: simple-prod + app.kubernetes.io/managed-by: jaeger-operator + app.kubernetes.io/name: simple-prod-query + app.kubernetes.io/part-of: jaeger + spec: + containers: + - args: + - --query.ui-config=/etc/config/ui.json + - --es.server-urls=https://elasticsearch.kuttl-test-enjoyed-hare.svc.cluster.local:9200 + - --es.tls.enabled=true + - --es.tls.ca=/certs/ca + - --es.tls.cert=/certs/cert + - --es.tls.key=/certs/key + - --es.timeout=15s + - --es.num-shards=1 + - --es.num-replicas=0 + env: + - name: SPAN_STORAGE_TYPE + value: elasticsearch + - name: METRICS_STORAGE_TYPE + - name: JAEGER_DISABLED + value: "false" + - name: JAEGER_SERVICE_NAME + value: simple-prod.kuttl-test-enjoyed-hare + - name: JAEGER_PROPAGATION + value: jaeger,b3,w3c + image: registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:831f83e8396d61822fb2dd576e711e0f10c1f8c7bd03bf0d4bec49dfb90cacca + imagePullPolicy: IfNotPresent + livenessProbe: + failureThreshold: 5 + httpGet: + path: / + port: 16687 + scheme: HTTP + initialDelaySeconds: 5 + periodSeconds: 15 + successThreshold: 1 + timeoutSeconds: 1 + name: jaeger-query + ports: + - containerPort: 16685 + name: grpc-query + protocol: TCP + - containerPort: 16686 + name: query + protocol: TCP + - containerPort: 16687 + name: admin-http + protocol: TCP + readinessProbe: + failureThreshold: 3 + httpGet: + path: / + port: 16687 + scheme: HTTP + initialDelaySeconds: 1 + periodSeconds: 10 + successThreshold: 1 + timeoutSeconds: 1 + resources: {} + terminationMessagePath: /dev/termination-log + terminationMessagePolicy: File + volumeMounts: + - mountPath: /etc/config + name: simple-prod-ui-configuration-volume + readOnly: true + - mountPath: /etc/pki/ca-trust/extracted/pem + name: simple-prod-trusted-ca + readOnly: true + - mountPath: /certs + name: certs + readOnly: true + - args: + - --cookie-secret=p1eNIWJwUuh4KfiXWoVhED + - --https-address=:8443 + - '--openshift-sar={"namespace": "kuttl-test-enjoyed-hare", "resource": "pods", + "verb": "get"}' + - --openshift-service-account=simple-prod-ui-proxy + - --provider=openshift + - --tls-cert=/etc/tls/private/tls.crt + - --tls-key=/etc/tls/private/tls.key + - --upstream=http://localhost:16686 + env: + - name: JAEGER_SERVICE_NAME + value: simple-prod.kuttl-test-enjoyed-hare + - name: JAEGER_PROPAGATION + value: jaeger,b3,w3c + image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1613032062a68faa414cd5754a666b7b78ad0e52cf328eaaed355719423a55dd + imagePullPolicy: IfNotPresent + name: oauth-proxy + ports: + - containerPort: 8443 + name: public + protocol: TCP + resources: {} + terminationMessagePath: /dev/termination-log + terminationMessagePolicy: File + volumeMounts: + - mountPath: /etc/tls/private + name: simple-prod-ui-oauth-proxy-tls + - mountPath: /etc/pki/ca-trust/extracted/pem + name: simple-prod-trusted-ca + readOnly: true + - args: + - --agent.tags=cluster=undefined,deployment.name=simple-prod-query,host.ip=${HOST_IP:},pod.name=${POD_NAME:},pod.namespace=kuttl-test-enjoyed-hare + - --reporter.grpc.host-port=dns:///simple-prod-collector-headless.kuttl-test-enjoyed-hare.svc:14250 + - --reporter.grpc.tls.ca=/etc/pki/ca-trust/source/service-ca/service-ca.crt + - --reporter.grpc.tls.enabled=true + env: + - name: POD_NAME + valueFrom: + fieldRef: + apiVersion: v1 + fieldPath: metadata.name + - name: HOST_IP + valueFrom: + fieldRef: + apiVersion: v1 + fieldPath: status.hostIP + image: registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:5fcc9d365678f54c0f52005318dbbc32b87f743cae7e1a84568e769653f3b128 + imagePullPolicy: IfNotPresent + livenessProbe: + failureThreshold: 5 + httpGet: + path: / + port: 14271 + scheme: HTTP + initialDelaySeconds: 5 + periodSeconds: 15 + successThreshold: 1 + timeoutSeconds: 1 + name: jaeger-agent + ports: + - containerPort: 5775 + name: zk-compact-trft + protocol: UDP + - containerPort: 5778 + name: config-rest + protocol: TCP + - containerPort: 6831 + name: jg-compact-trft + protocol: UDP + - containerPort: 6832 + name: jg-binary-trft + protocol: UDP + - containerPort: 14271 + name: admin-http + protocol: TCP + readinessProbe: + failureThreshold: 3 + httpGet: + path: / + port: 14271 + scheme: HTTP + initialDelaySeconds: 1 + periodSeconds: 10 + successThreshold: 1 + timeoutSeconds: 1 + resources: {} + terminationMessagePath: /dev/termination-log + terminationMessagePolicy: File + volumeMounts: + - mountPath: /etc/pki/ca-trust/extracted/pem + name: simple-prod-trusted-ca + readOnly: true + - mountPath: /etc/pki/ca-trust/source/service-ca + name: simple-prod-service-ca + readOnly: true + dnsPolicy: ClusterFirst + enableServiceLinks: false + restartPolicy: Always + schedulerName: default-scheduler + securityContext: {} + serviceAccount: simple-prod-ui-proxy + serviceAccountName: simple-prod-ui-proxy + terminationGracePeriodSeconds: 30 + volumes: + - configMap: + defaultMode: 420 + items: + - key: ui + path: ui.json + name: simple-prod-ui-configuration + name: simple-prod-ui-configuration-volume + - configMap: + defaultMode: 420 + items: + - key: ca-bundle.crt + path: tls-ca-bundle.pem + name: simple-prod-trusted-ca + name: simple-prod-trusted-ca + - name: simple-prod-ui-oauth-proxy-tls + secret: + defaultMode: 420 + secretName: simple-prod-ui-oauth-proxy-tls + - name: certs + secret: + defaultMode: 420 + secretName: simple-prod-jaeger-elasticsearch + - configMap: + defaultMode: 420 + items: + - key: service-ca.crt + path: service-ca.crt + name: simple-prod-service-ca + name: simple-prod-service-ca status: - readyReplicas: 1 + conditions: + - lastTransitionTime: "2024-08-29T19:16:43Z" + lastUpdateTime: "2024-08-29T19:16:43Z" + message: Deployment does not have minimum availability. + reason: MinimumReplicasUnavailable + status: "False" + type: Available + - lastTransitionTime: "2024-08-29T19:16:43Z" + lastUpdateTime: "2024-08-29T19:16:43Z" + message: ReplicaSet "simple-prod-query-75c44cf6fc" is progressing. + reason: ReplicaSetUpdated + status: "True" + type: Progressing + observedGeneration: 3 + replicas: 1 + unavailableReplicas: 1 + updatedReplicas: 1 case.go:366: resource Deployment:kuttl-test-enjoyed-hare/simple-prod-query: .status.readyReplicas: key is missing from map logger.go:42: 19:26:39 | examples-simple-prod | examples-simple-prod events from ns kuttl-test-enjoyed-hare: logger.go:42: 19:26:39 | examples-simple-prod | 2024-08-29 19:16:43 +0000 UTC Normal Pod simple-prod-collector-77854cb9f7-gvj55 Binding Scheduled Successfully assigned kuttl-test-enjoyed-hare/simple-prod-collector-77854cb9f7-gvj55 to ip-10-0-51-237.us-west-2.compute.internal default-scheduler logger.go:42: 19:26:39 | examples-simple-prod | 2024-08-29 19:16:43 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-77854cb9f7 SuccessfulCreate Created pod: simple-prod-collector-77854cb9f7-gvj55 replicaset-controller logger.go:42: 19:26:39 | examples-simple-prod | 2024-08-29 19:16:43 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-77854cb9f7 to 1 deployment-controller logger.go:42: 19:26:39 | examples-simple-prod | 2024-08-29 19:16:43 +0000 UTC Normal Pod simple-prod-query-75c44cf6fc-dp5kb Binding Scheduled Successfully assigned kuttl-test-enjoyed-hare/simple-prod-query-75c44cf6fc-dp5kb to ip-10-0-51-237.us-west-2.compute.internal default-scheduler logger.go:42: 19:26:39 | examples-simple-prod | 2024-08-29 19:16:43 +0000 UTC Normal ReplicaSet.apps simple-prod-query-75c44cf6fc SuccessfulCreate Created pod: simple-prod-query-75c44cf6fc-dp5kb replicaset-controller logger.go:42: 19:26:39 | examples-simple-prod | 2024-08-29 19:16:43 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-75c44cf6fc to 1 deployment-controller logger.go:42: 19:26:39 | examples-simple-prod | 2024-08-29 19:16:44 +0000 UTC Normal Pod simple-prod-collector-77854cb9f7-gvj55 AddedInterface Add eth0 [10.130.0.12/23] from ovn-kubernetes multus logger.go:42: 19:26:39 | examples-simple-prod | 2024-08-29 19:16:44 +0000 UTC Normal Pod simple-prod-collector-77854cb9f7-gvj55.spec.containers{jaeger-collector} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:b90e0d264bb0b051c2a4b9e6d499f407fed58f834ab8ebf30263ee1cf7c7ea3c" kubelet logger.go:42: 19:26:39 | examples-simple-prod | 2024-08-29 19:16:44 +0000 UTC Normal Pod simple-prod-query-75c44cf6fc-dp5kb AddedInterface Add eth0 [10.130.0.11/23] from ovn-kubernetes multus logger.go:42: 19:26:39 | examples-simple-prod | 2024-08-29 19:16:44 +0000 UTC Normal Pod simple-prod-query-75c44cf6fc-dp5kb.spec.containers{jaeger-query} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:831f83e8396d61822fb2dd576e711e0f10c1f8c7bd03bf0d4bec49dfb90cacca" kubelet logger.go:42: 19:26:39 | examples-simple-prod | 2024-08-29 19:16:50 +0000 UTC Normal Pod simple-prod-query-75c44cf6fc-dp5kb.spec.containers{jaeger-query} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:831f83e8396d61822fb2dd576e711e0f10c1f8c7bd03bf0d4bec49dfb90cacca" in 5.977s (5.977s including waiting) kubelet logger.go:42: 19:26:39 | examples-simple-prod | 2024-08-29 19:16:50 +0000 UTC Normal Pod simple-prod-query-75c44cf6fc-dp5kb.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 19:26:39 | examples-simple-prod | 2024-08-29 19:16:50 +0000 UTC Normal Pod simple-prod-query-75c44cf6fc-dp5kb.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 19:26:39 | examples-simple-prod | 2024-08-29 19:16:50 +0000 UTC Normal Pod simple-prod-query-75c44cf6fc-dp5kb.spec.containers{oauth-proxy} Pulling Pulling image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1613032062a68faa414cd5754a666b7b78ad0e52cf328eaaed355719423a55dd" kubelet logger.go:42: 19:26:39 | examples-simple-prod | 2024-08-29 19:16:51 +0000 UTC Normal Pod simple-prod-collector-77854cb9f7-gvj55.spec.containers{jaeger-collector} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:b90e0d264bb0b051c2a4b9e6d499f407fed58f834ab8ebf30263ee1cf7c7ea3c" in 6.92s (6.92s including waiting) kubelet logger.go:42: 19:26:39 | examples-simple-prod | 2024-08-29 19:16:51 +0000 UTC Normal Pod simple-prod-collector-77854cb9f7-gvj55.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 19:26:39 | examples-simple-prod | 2024-08-29 19:16:51 +0000 UTC Normal Pod simple-prod-collector-77854cb9f7-gvj55.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 19:26:39 | examples-simple-prod | 2024-08-29 19:16:52 +0000 UTC Warning Pod simple-prod-collector-77854cb9f7-gvj55.spec.containers{jaeger-collector} Unhealthy Readiness probe failed: HTTP probe failed with statuscode: 503 kubelet logger.go:42: 19:26:39 | examples-simple-prod | 2024-08-29 19:16:58 +0000 UTC Normal Pod simple-prod-collector-77854cb9f7-gvj55.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:b90e0d264bb0b051c2a4b9e6d499f407fed58f834ab8ebf30263ee1cf7c7ea3c" already present on machine kubelet logger.go:42: 19:26:39 | examples-simple-prod | 2024-08-29 19:16:58 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 19:26:39 | examples-simple-prod | 2024-08-29 19:16:58 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 19:26:39 | examples-simple-prod | 2024-08-29 19:16:58 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 19:26:39 | examples-simple-prod | 2024-08-29 19:16:58 +0000 UTC Normal Pod simple-prod-query-75c44cf6fc-dp5kb.spec.containers{oauth-proxy} Pulled Successfully pulled image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1613032062a68faa414cd5754a666b7b78ad0e52cf328eaaed355719423a55dd" in 7.958s (7.958s including waiting) kubelet logger.go:42: 19:26:39 | examples-simple-prod | 2024-08-29 19:16:58 +0000 UTC Normal Pod simple-prod-query-75c44cf6fc-dp5kb.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 19:26:39 | examples-simple-prod | 2024-08-29 19:16:58 +0000 UTC Normal Pod simple-prod-query-75c44cf6fc-dp5kb.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 19:26:39 | examples-simple-prod | 2024-08-29 19:16:58 +0000 UTC Normal Pod simple-prod-query-75c44cf6fc-dp5kb.spec.containers{jaeger-agent} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:5fcc9d365678f54c0f52005318dbbc32b87f743cae7e1a84568e769653f3b128" kubelet logger.go:42: 19:26:39 | examples-simple-prod | 2024-08-29 19:17:03 +0000 UTC Warning Pod simple-prod-collector-77854cb9f7-gvj55.spec.containers{jaeger-collector} BackOff Back-off restarting failed container jaeger-collector in pod simple-prod-collector-77854cb9f7-gvj55_kuttl-test-enjoyed-hare(9d14f6ad-9c24-4782-b0b2-faa20d43e434) kubelet logger.go:42: 19:26:39 | examples-simple-prod | 2024-08-29 19:17:03 +0000 UTC Normal Pod simple-prod-query-75c44cf6fc-dp5kb.spec.containers{jaeger-agent} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:5fcc9d365678f54c0f52005318dbbc32b87f743cae7e1a84568e769653f3b128" in 4.452s (4.452s including waiting) kubelet logger.go:42: 19:26:39 | examples-simple-prod | 2024-08-29 19:17:03 +0000 UTC Normal Pod simple-prod-query-75c44cf6fc-dp5kb.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 19:26:39 | examples-simple-prod | 2024-08-29 19:17:03 +0000 UTC Normal Pod simple-prod-query-75c44cf6fc-dp5kb.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 19:26:39 | examples-simple-prod | 2024-08-29 19:17:03 +0000 UTC Normal Pod simple-prod-query-75c44cf6fc-dp5kb.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:831f83e8396d61822fb2dd576e711e0f10c1f8c7bd03bf0d4bec49dfb90cacca" already present on machine kubelet logger.go:42: 19:26:39 | examples-simple-prod | 2024-08-29 19:17:04 +0000 UTC Warning Pod simple-prod-query-75c44cf6fc-dp5kb.spec.containers{jaeger-query} Unhealthy Readiness probe failed: HTTP probe failed with statuscode: 503 kubelet logger.go:42: 19:26:39 | examples-simple-prod | 2024-08-29 19:17:09 +0000 UTC Warning Pod simple-prod-query-75c44cf6fc-dp5kb.spec.containers{jaeger-query} BackOff Back-off restarting failed container jaeger-query in pod simple-prod-query-75c44cf6fc-dp5kb_kuttl-test-enjoyed-hare(5f7256c6-7fdb-48ae-b9ba-cd50fc0c69c5) kubelet logger.go:42: 19:26:39 | examples-simple-prod | Deleting namespace: kuttl-test-enjoyed-hare === CONT kuttl/harness/examples-with-sampling logger.go:42: 19:26:45 | examples-with-sampling | Creating namespace: kuttl-test-elegant-grouper logger.go:42: 19:26:46 | examples-with-sampling/0-install | starting test step 0-install logger.go:42: 19:26:46 | examples-with-sampling/0-install | running command: [sh -c cd /tmp/jaeger-tests && make cassandra STORAGE_NAMESPACE=$NAMESPACE] logger.go:42: 19:26:46 | examples-with-sampling/0-install | make[2]: Entering directory '/tmp/jaeger-tests' logger.go:42: 19:26:46 | examples-with-sampling/0-install | >>>> Creating namespace kuttl-test-elegant-grouper logger.go:42: 19:26:46 | examples-with-sampling/0-install | kubectl create namespace kuttl-test-elegant-grouper 2>&1 | grep -v "already exists" || true logger.go:42: 19:26:46 | examples-with-sampling/0-install | kubectl create -f ./tests/cassandra.yml --namespace kuttl-test-elegant-grouper 2>&1 | grep -v "already exists" || true logger.go:42: 19:26:49 | examples-with-sampling/0-install | service/cassandra created logger.go:42: 19:26:49 | examples-with-sampling/0-install | statefulset.apps/cassandra created logger.go:42: 19:26:49 | examples-with-sampling/0-install | make[2]: Leaving directory '/tmp/jaeger-tests' logger.go:42: 19:27:03 | examples-with-sampling/0-install | test step completed 0-install logger.go:42: 19:27:03 | examples-with-sampling/1-install | starting test step 1-install logger.go:42: 19:27:03 | examples-with-sampling/1-install | Jaeger:kuttl-test-elegant-grouper/with-sampling created logger.go:42: 19:27:10 | examples-with-sampling/1-install | test step completed 1-install logger.go:42: 19:27:10 | examples-with-sampling/2-smoke-test | starting test step 2-smoke-test logger.go:42: 19:27:10 | examples-with-sampling/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE with-sampling /dev/null] logger.go:42: 19:27:13 | examples-with-sampling/2-smoke-test | Warning: resource jaegers/with-sampling is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 19:27:22 | examples-with-sampling/2-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://with-sampling-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://with-sampling-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 19:27:22 | examples-with-sampling/2-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 19:27:23 | examples-with-sampling/2-smoke-test | job.batch/report-span created logger.go:42: 19:27:23 | examples-with-sampling/2-smoke-test | job.batch/check-span created logger.go:42: 19:27:38 | examples-with-sampling/2-smoke-test | test step completed 2-smoke-test logger.go:42: 19:27:38 | examples-with-sampling/3- | starting test step 3- logger.go:42: 19:27:39 | examples-with-sampling/3- | test step completed 3- logger.go:42: 19:27:39 | examples-with-sampling | examples-with-sampling events from ns kuttl-test-elegant-grouper: logger.go:42: 19:27:39 | examples-with-sampling | 2024-08-29 19:26:49 +0000 UTC Normal Pod cassandra-0 Binding Scheduled Successfully assigned kuttl-test-elegant-grouper/cassandra-0 to ip-10-0-51-237.us-west-2.compute.internal default-scheduler logger.go:42: 19:27:39 | examples-with-sampling | 2024-08-29 19:26:49 +0000 UTC Normal Pod cassandra-0 AddedInterface Add eth0 [10.130.0.13/23] from ovn-kubernetes multus logger.go:42: 19:27:39 | examples-with-sampling | 2024-08-29 19:26:49 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Pulling Pulling image "cassandra:3.11" kubelet logger.go:42: 19:27:39 | examples-with-sampling | 2024-08-29 19:26:49 +0000 UTC Normal StatefulSet.apps cassandra SuccessfulCreate create Pod cassandra-0 in StatefulSet cassandra successful statefulset-controller logger.go:42: 19:27:39 | examples-with-sampling | 2024-08-29 19:26:55 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Pulled Successfully pulled image "cassandra:3.11" in 5.514s (5.515s including waiting) kubelet logger.go:42: 19:27:39 | examples-with-sampling | 2024-08-29 19:26:55 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Created Created container cassandra kubelet logger.go:42: 19:27:39 | examples-with-sampling | 2024-08-29 19:26:55 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Started Started container cassandra kubelet logger.go:42: 19:27:39 | examples-with-sampling | 2024-08-29 19:26:56 +0000 UTC Normal Pod cassandra-1 Binding Scheduled Successfully assigned kuttl-test-elegant-grouper/cassandra-1 to ip-10-0-57-246.us-west-2.compute.internal default-scheduler logger.go:42: 19:27:39 | examples-with-sampling | 2024-08-29 19:26:56 +0000 UTC Normal Pod cassandra-1 AddedInterface Add eth0 [10.129.0.30/23] from ovn-kubernetes multus logger.go:42: 19:27:39 | examples-with-sampling | 2024-08-29 19:26:56 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Pulling Pulling image "cassandra:3.11" kubelet logger.go:42: 19:27:39 | examples-with-sampling | 2024-08-29 19:26:56 +0000 UTC Normal StatefulSet.apps cassandra SuccessfulCreate create Pod cassandra-1 in StatefulSet cassandra successful statefulset-controller logger.go:42: 19:27:39 | examples-with-sampling | 2024-08-29 19:27:02 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Pulled Successfully pulled image "cassandra:3.11" in 5.64s (5.64s including waiting) kubelet logger.go:42: 19:27:39 | examples-with-sampling | 2024-08-29 19:27:02 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Created Created container cassandra kubelet logger.go:42: 19:27:39 | examples-with-sampling | 2024-08-29 19:27:02 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Started Started container cassandra kubelet logger.go:42: 19:27:39 | examples-with-sampling | 2024-08-29 19:27:04 +0000 UTC Normal Pod with-sampling-7b8549646-t7lkt Binding Scheduled Successfully assigned kuttl-test-elegant-grouper/with-sampling-7b8549646-t7lkt to ip-10-0-51-237.us-west-2.compute.internal default-scheduler logger.go:42: 19:27:39 | examples-with-sampling | 2024-08-29 19:27:04 +0000 UTC Normal Pod with-sampling-7b8549646-t7lkt AddedInterface Add eth0 [10.130.0.14/23] from ovn-kubernetes multus logger.go:42: 19:27:39 | examples-with-sampling | 2024-08-29 19:27:04 +0000 UTC Normal Pod with-sampling-7b8549646-t7lkt.spec.containers{jaeger} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:345d3b6c507df9f8a18948a594e7a03698085689c6965faf79b8aadc74a0ceaa" kubelet logger.go:42: 19:27:39 | examples-with-sampling | 2024-08-29 19:27:04 +0000 UTC Normal ReplicaSet.apps with-sampling-7b8549646 SuccessfulCreate Created pod: with-sampling-7b8549646-t7lkt replicaset-controller logger.go:42: 19:27:39 | examples-with-sampling | 2024-08-29 19:27:04 +0000 UTC Normal Deployment.apps with-sampling ScalingReplicaSet Scaled up replica set with-sampling-7b8549646 to 1 deployment-controller logger.go:42: 19:27:39 | examples-with-sampling | 2024-08-29 19:27:08 +0000 UTC Normal Pod with-sampling-7b8549646-t7lkt.spec.containers{jaeger} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:345d3b6c507df9f8a18948a594e7a03698085689c6965faf79b8aadc74a0ceaa" in 4.005s (4.005s including waiting) kubelet logger.go:42: 19:27:39 | examples-with-sampling | 2024-08-29 19:27:08 +0000 UTC Normal Pod with-sampling-7b8549646-t7lkt.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 19:27:39 | examples-with-sampling | 2024-08-29 19:27:08 +0000 UTC Normal Pod with-sampling-7b8549646-t7lkt.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 19:27:39 | examples-with-sampling | 2024-08-29 19:27:08 +0000 UTC Normal Pod with-sampling-7b8549646-t7lkt.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1613032062a68faa414cd5754a666b7b78ad0e52cf328eaaed355719423a55dd" already present on machine kubelet logger.go:42: 19:27:39 | examples-with-sampling | 2024-08-29 19:27:09 +0000 UTC Normal Pod with-sampling-7b8549646-t7lkt.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 19:27:39 | examples-with-sampling | 2024-08-29 19:27:09 +0000 UTC Normal Pod with-sampling-7b8549646-t7lkt.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 19:27:39 | examples-with-sampling | 2024-08-29 19:27:15 +0000 UTC Normal Pod with-sampling-7b8549646-t7lkt.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 19:27:39 | examples-with-sampling | 2024-08-29 19:27:15 +0000 UTC Normal Pod with-sampling-7b8549646-t7lkt.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 19:27:39 | examples-with-sampling | 2024-08-29 19:27:15 +0000 UTC Normal ReplicaSet.apps with-sampling-7b8549646 SuccessfulDelete Deleted pod: with-sampling-7b8549646-t7lkt replicaset-controller logger.go:42: 19:27:39 | examples-with-sampling | 2024-08-29 19:27:15 +0000 UTC Normal Deployment.apps with-sampling ScalingReplicaSet Scaled down replica set with-sampling-7b8549646 to 0 from 1 deployment-controller logger.go:42: 19:27:39 | examples-with-sampling | 2024-08-29 19:27:16 +0000 UTC Normal Pod with-sampling-7b7cd758bb-vvts2 Binding Scheduled Successfully assigned kuttl-test-elegant-grouper/with-sampling-7b7cd758bb-vvts2 to ip-10-0-51-237.us-west-2.compute.internal default-scheduler logger.go:42: 19:27:39 | examples-with-sampling | 2024-08-29 19:27:16 +0000 UTC Normal Pod with-sampling-7b7cd758bb-vvts2 AddedInterface Add eth0 [10.130.0.15/23] from ovn-kubernetes multus logger.go:42: 19:27:39 | examples-with-sampling | 2024-08-29 19:27:16 +0000 UTC Normal Pod with-sampling-7b7cd758bb-vvts2.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:345d3b6c507df9f8a18948a594e7a03698085689c6965faf79b8aadc74a0ceaa" already present on machine kubelet logger.go:42: 19:27:39 | examples-with-sampling | 2024-08-29 19:27:16 +0000 UTC Normal Pod with-sampling-7b7cd758bb-vvts2.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 19:27:39 | examples-with-sampling | 2024-08-29 19:27:16 +0000 UTC Normal ReplicaSet.apps with-sampling-7b7cd758bb SuccessfulCreate Created pod: with-sampling-7b7cd758bb-vvts2 replicaset-controller logger.go:42: 19:27:39 | examples-with-sampling | 2024-08-29 19:27:16 +0000 UTC Normal Deployment.apps with-sampling ScalingReplicaSet Scaled up replica set with-sampling-7b7cd758bb to 1 deployment-controller logger.go:42: 19:27:39 | examples-with-sampling | 2024-08-29 19:27:17 +0000 UTC Normal Pod with-sampling-7b7cd758bb-vvts2.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 19:27:39 | examples-with-sampling | 2024-08-29 19:27:17 +0000 UTC Normal Pod with-sampling-7b7cd758bb-vvts2.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1613032062a68faa414cd5754a666b7b78ad0e52cf328eaaed355719423a55dd" already present on machine kubelet logger.go:42: 19:27:39 | examples-with-sampling | 2024-08-29 19:27:17 +0000 UTC Normal Pod with-sampling-7b7cd758bb-vvts2.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 19:27:39 | examples-with-sampling | 2024-08-29 19:27:17 +0000 UTC Normal Pod with-sampling-7b7cd758bb-vvts2.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 19:27:39 | examples-with-sampling | 2024-08-29 19:27:23 +0000 UTC Normal Pod check-span-fnl8r Binding Scheduled Successfully assigned kuttl-test-elegant-grouper/check-span-fnl8r to ip-10-0-51-237.us-west-2.compute.internal default-scheduler logger.go:42: 19:27:39 | examples-with-sampling | 2024-08-29 19:27:23 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-fnl8r job-controller logger.go:42: 19:27:39 | examples-with-sampling | 2024-08-29 19:27:23 +0000 UTC Normal Pod report-span-rnzxr Binding Scheduled Successfully assigned kuttl-test-elegant-grouper/report-span-rnzxr to ip-10-0-51-237.us-west-2.compute.internal default-scheduler logger.go:42: 19:27:39 | examples-with-sampling | 2024-08-29 19:27:23 +0000 UTC Normal Pod report-span-rnzxr AddedInterface Add eth0 [10.130.0.16/23] from ovn-kubernetes multus logger.go:42: 19:27:39 | examples-with-sampling | 2024-08-29 19:27:23 +0000 UTC Normal Pod report-span-rnzxr.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 19:27:39 | examples-with-sampling | 2024-08-29 19:27:23 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-rnzxr job-controller logger.go:42: 19:27:39 | examples-with-sampling | 2024-08-29 19:27:24 +0000 UTC Normal Pod check-span-fnl8r AddedInterface Add eth0 [10.130.0.17/23] from ovn-kubernetes multus logger.go:42: 19:27:39 | examples-with-sampling | 2024-08-29 19:27:24 +0000 UTC Normal Pod check-span-fnl8r.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 19:27:39 | examples-with-sampling | 2024-08-29 19:27:27 +0000 UTC Normal Pod check-span-fnl8r.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 3.454s (3.454s including waiting) kubelet logger.go:42: 19:27:39 | examples-with-sampling | 2024-08-29 19:27:27 +0000 UTC Normal Pod report-span-rnzxr.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 3.948s (3.948s including waiting) kubelet logger.go:42: 19:27:39 | examples-with-sampling | 2024-08-29 19:27:28 +0000 UTC Normal Pod check-span-fnl8r.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 19:27:39 | examples-with-sampling | 2024-08-29 19:27:28 +0000 UTC Normal Pod check-span-fnl8r.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 19:27:39 | examples-with-sampling | 2024-08-29 19:27:28 +0000 UTC Normal Pod report-span-rnzxr.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 19:27:39 | examples-with-sampling | 2024-08-29 19:27:28 +0000 UTC Normal Pod report-span-rnzxr.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 19:27:39 | examples-with-sampling | 2024-08-29 19:27:38 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 19:27:39 | examples-with-sampling | Deleting namespace: kuttl-test-elegant-grouper === CONT kuttl/harness/examples-with-cassandra logger.go:42: 19:27:51 | examples-with-cassandra | Creating namespace: kuttl-test-peaceful-goat logger.go:42: 19:27:51 | examples-with-cassandra/0-install | starting test step 0-install logger.go:42: 19:27:51 | examples-with-cassandra/0-install | running command: [sh -c cd /tmp/jaeger-tests && make cassandra STORAGE_NAMESPACE=$NAMESPACE] logger.go:42: 19:27:51 | examples-with-cassandra/0-install | make[2]: Entering directory '/tmp/jaeger-tests' logger.go:42: 19:27:51 | examples-with-cassandra/0-install | >>>> Creating namespace kuttl-test-peaceful-goat logger.go:42: 19:27:51 | examples-with-cassandra/0-install | kubectl create namespace kuttl-test-peaceful-goat 2>&1 | grep -v "already exists" || true logger.go:42: 19:27:51 | examples-with-cassandra/0-install | kubectl create -f ./tests/cassandra.yml --namespace kuttl-test-peaceful-goat 2>&1 | grep -v "already exists" || true logger.go:42: 19:27:52 | examples-with-cassandra/0-install | service/cassandra created logger.go:42: 19:27:52 | examples-with-cassandra/0-install | statefulset.apps/cassandra created logger.go:42: 19:27:52 | examples-with-cassandra/0-install | make[2]: Leaving directory '/tmp/jaeger-tests' logger.go:42: 19:27:54 | examples-with-cassandra/0-install | test step completed 0-install logger.go:42: 19:27:54 | examples-with-cassandra/1-install | starting test step 1-install logger.go:42: 19:27:55 | examples-with-cassandra/1-install | Jaeger:kuttl-test-peaceful-goat/with-cassandra created logger.go:42: 19:28:13 | examples-with-cassandra/1-install | test step completed 1-install logger.go:42: 19:28:13 | examples-with-cassandra/2-smoke-test | starting test step 2-smoke-test logger.go:42: 19:28:13 | examples-with-cassandra/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE with-cassandra /dev/null] logger.go:42: 19:28:15 | examples-with-cassandra/2-smoke-test | Warning: resource jaegers/with-cassandra is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 19:28:22 | examples-with-cassandra/2-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://with-cassandra-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://with-cassandra-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 19:28:22 | examples-with-cassandra/2-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 19:28:23 | examples-with-cassandra/2-smoke-test | job.batch/report-span created logger.go:42: 19:28:23 | examples-with-cassandra/2-smoke-test | job.batch/check-span created logger.go:42: 19:28:36 | examples-with-cassandra/2-smoke-test | test step completed 2-smoke-test logger.go:42: 19:28:36 | examples-with-cassandra | examples-with-cassandra events from ns kuttl-test-peaceful-goat: logger.go:42: 19:28:36 | examples-with-cassandra | 2024-08-29 19:27:52 +0000 UTC Normal Pod cassandra-0 Binding Scheduled Successfully assigned kuttl-test-peaceful-goat/cassandra-0 to ip-10-0-51-237.us-west-2.compute.internal default-scheduler logger.go:42: 19:28:36 | examples-with-cassandra | 2024-08-29 19:27:52 +0000 UTC Normal Pod cassandra-0 AddedInterface Add eth0 [10.130.0.18/23] from ovn-kubernetes multus logger.go:42: 19:28:36 | examples-with-cassandra | 2024-08-29 19:27:52 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Pulled Container image "cassandra:3.11" already present on machine kubelet logger.go:42: 19:28:36 | examples-with-cassandra | 2024-08-29 19:27:52 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Created Created container cassandra kubelet logger.go:42: 19:28:36 | examples-with-cassandra | 2024-08-29 19:27:52 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Started Started container cassandra kubelet logger.go:42: 19:28:36 | examples-with-cassandra | 2024-08-29 19:27:52 +0000 UTC Normal StatefulSet.apps cassandra SuccessfulCreate create Pod cassandra-0 in StatefulSet cassandra successful statefulset-controller logger.go:42: 19:28:36 | examples-with-cassandra | 2024-08-29 19:27:53 +0000 UTC Normal Pod cassandra-1 Binding Scheduled Successfully assigned kuttl-test-peaceful-goat/cassandra-1 to ip-10-0-57-246.us-west-2.compute.internal default-scheduler logger.go:42: 19:28:36 | examples-with-cassandra | 2024-08-29 19:27:53 +0000 UTC Normal Pod cassandra-1 AddedInterface Add eth0 [10.129.0.31/23] from ovn-kubernetes multus logger.go:42: 19:28:36 | examples-with-cassandra | 2024-08-29 19:27:53 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Pulled Container image "cassandra:3.11" already present on machine kubelet logger.go:42: 19:28:36 | examples-with-cassandra | 2024-08-29 19:27:53 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Created Created container cassandra kubelet logger.go:42: 19:28:36 | examples-with-cassandra | 2024-08-29 19:27:53 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Started Started container cassandra kubelet logger.go:42: 19:28:36 | examples-with-cassandra | 2024-08-29 19:27:53 +0000 UTC Normal StatefulSet.apps cassandra SuccessfulCreate create Pod cassandra-1 in StatefulSet cassandra successful statefulset-controller logger.go:42: 19:28:36 | examples-with-cassandra | 2024-08-29 19:27:55 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-dxmzj Binding Scheduled Successfully assigned kuttl-test-peaceful-goat/with-cassandra-cassandra-schema-job-dxmzj to ip-10-0-51-237.us-west-2.compute.internal default-scheduler logger.go:42: 19:28:36 | examples-with-cassandra | 2024-08-29 19:27:55 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-dxmzj AddedInterface Add eth0 [10.130.0.19/23] from ovn-kubernetes multus logger.go:42: 19:28:36 | examples-with-cassandra | 2024-08-29 19:27:55 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-dxmzj.spec.containers{with-cassandra-cassandra-schema-job} Pulling Pulling image "jaegertracing/jaeger-cassandra-schema:1.57.0" kubelet logger.go:42: 19:28:36 | examples-with-cassandra | 2024-08-29 19:27:55 +0000 UTC Normal Job.batch with-cassandra-cassandra-schema-job SuccessfulCreate Created pod: with-cassandra-cassandra-schema-job-dxmzj job-controller logger.go:42: 19:28:36 | examples-with-cassandra | 2024-08-29 19:28:02 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-dxmzj.spec.containers{with-cassandra-cassandra-schema-job} Pulled Successfully pulled image "jaegertracing/jaeger-cassandra-schema:1.57.0" in 6.271s (6.271s including waiting) kubelet logger.go:42: 19:28:36 | examples-with-cassandra | 2024-08-29 19:28:02 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-dxmzj.spec.containers{with-cassandra-cassandra-schema-job} Created Created container with-cassandra-cassandra-schema-job kubelet logger.go:42: 19:28:36 | examples-with-cassandra | 2024-08-29 19:28:02 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-dxmzj.spec.containers{with-cassandra-cassandra-schema-job} Started Started container with-cassandra-cassandra-schema-job kubelet logger.go:42: 19:28:36 | examples-with-cassandra | 2024-08-29 19:28:10 +0000 UTC Normal Pod with-cassandra-7f69d999f7-f4m47 Binding Scheduled Successfully assigned kuttl-test-peaceful-goat/with-cassandra-7f69d999f7-f4m47 to ip-10-0-51-237.us-west-2.compute.internal default-scheduler logger.go:42: 19:28:36 | examples-with-cassandra | 2024-08-29 19:28:10 +0000 UTC Normal ReplicaSet.apps with-cassandra-7f69d999f7 SuccessfulCreate Created pod: with-cassandra-7f69d999f7-f4m47 replicaset-controller logger.go:42: 19:28:36 | examples-with-cassandra | 2024-08-29 19:28:10 +0000 UTC Normal Job.batch with-cassandra-cassandra-schema-job Completed Job completed job-controller logger.go:42: 19:28:36 | examples-with-cassandra | 2024-08-29 19:28:10 +0000 UTC Normal Deployment.apps with-cassandra ScalingReplicaSet Scaled up replica set with-cassandra-7f69d999f7 to 1 deployment-controller logger.go:42: 19:28:36 | examples-with-cassandra | 2024-08-29 19:28:11 +0000 UTC Normal Pod with-cassandra-7f69d999f7-f4m47 AddedInterface Add eth0 [10.130.0.20/23] from ovn-kubernetes multus logger.go:42: 19:28:36 | examples-with-cassandra | 2024-08-29 19:28:11 +0000 UTC Normal Pod with-cassandra-7f69d999f7-f4m47.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:345d3b6c507df9f8a18948a594e7a03698085689c6965faf79b8aadc74a0ceaa" already present on machine kubelet logger.go:42: 19:28:36 | examples-with-cassandra | 2024-08-29 19:28:11 +0000 UTC Normal Pod with-cassandra-7f69d999f7-f4m47.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 19:28:36 | examples-with-cassandra | 2024-08-29 19:28:11 +0000 UTC Normal Pod with-cassandra-7f69d999f7-f4m47.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 19:28:36 | examples-with-cassandra | 2024-08-29 19:28:11 +0000 UTC Normal Pod with-cassandra-7f69d999f7-f4m47.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1613032062a68faa414cd5754a666b7b78ad0e52cf328eaaed355719423a55dd" already present on machine kubelet logger.go:42: 19:28:36 | examples-with-cassandra | 2024-08-29 19:28:11 +0000 UTC Normal Pod with-cassandra-7f69d999f7-f4m47.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 19:28:36 | examples-with-cassandra | 2024-08-29 19:28:11 +0000 UTC Normal Pod with-cassandra-7f69d999f7-f4m47.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 19:28:36 | examples-with-cassandra | 2024-08-29 19:28:15 +0000 UTC Normal Pod with-cassandra-7f69d999f7-f4m47.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 19:28:36 | examples-with-cassandra | 2024-08-29 19:28:15 +0000 UTC Normal Pod with-cassandra-7f69d999f7-f4m47.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 19:28:36 | examples-with-cassandra | 2024-08-29 19:28:15 +0000 UTC Normal ReplicaSet.apps with-cassandra-7f69d999f7 SuccessfulDelete Deleted pod: with-cassandra-7f69d999f7-f4m47 replicaset-controller logger.go:42: 19:28:36 | examples-with-cassandra | 2024-08-29 19:28:15 +0000 UTC Normal Deployment.apps with-cassandra ScalingReplicaSet Scaled down replica set with-cassandra-7f69d999f7 to 0 from 1 deployment-controller logger.go:42: 19:28:36 | examples-with-cassandra | 2024-08-29 19:28:16 +0000 UTC Normal Pod with-cassandra-f8677bcd7-tchxd Binding Scheduled Successfully assigned kuttl-test-peaceful-goat/with-cassandra-f8677bcd7-tchxd to ip-10-0-51-237.us-west-2.compute.internal default-scheduler logger.go:42: 19:28:36 | examples-with-cassandra | 2024-08-29 19:28:16 +0000 UTC Normal Pod with-cassandra-f8677bcd7-tchxd AddedInterface Add eth0 [10.130.0.21/23] from ovn-kubernetes multus logger.go:42: 19:28:36 | examples-with-cassandra | 2024-08-29 19:28:16 +0000 UTC Normal Pod with-cassandra-f8677bcd7-tchxd.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:345d3b6c507df9f8a18948a594e7a03698085689c6965faf79b8aadc74a0ceaa" already present on machine kubelet logger.go:42: 19:28:36 | examples-with-cassandra | 2024-08-29 19:28:16 +0000 UTC Normal Pod with-cassandra-f8677bcd7-tchxd.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 19:28:36 | examples-with-cassandra | 2024-08-29 19:28:16 +0000 UTC Normal Pod with-cassandra-f8677bcd7-tchxd.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 19:28:36 | examples-with-cassandra | 2024-08-29 19:28:16 +0000 UTC Normal Pod with-cassandra-f8677bcd7-tchxd.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1613032062a68faa414cd5754a666b7b78ad0e52cf328eaaed355719423a55dd" already present on machine kubelet logger.go:42: 19:28:36 | examples-with-cassandra | 2024-08-29 19:28:16 +0000 UTC Normal ReplicaSet.apps with-cassandra-f8677bcd7 SuccessfulCreate Created pod: with-cassandra-f8677bcd7-tchxd replicaset-controller logger.go:42: 19:28:36 | examples-with-cassandra | 2024-08-29 19:28:16 +0000 UTC Normal Deployment.apps with-cassandra ScalingReplicaSet Scaled up replica set with-cassandra-f8677bcd7 to 1 deployment-controller logger.go:42: 19:28:36 | examples-with-cassandra | 2024-08-29 19:28:17 +0000 UTC Normal Pod with-cassandra-f8677bcd7-tchxd.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 19:28:36 | examples-with-cassandra | 2024-08-29 19:28:17 +0000 UTC Normal Pod with-cassandra-f8677bcd7-tchxd.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 19:28:36 | examples-with-cassandra | 2024-08-29 19:28:23 +0000 UTC Normal Pod check-span-d4qs4 Binding Scheduled Successfully assigned kuttl-test-peaceful-goat/check-span-d4qs4 to ip-10-0-51-237.us-west-2.compute.internal default-scheduler logger.go:42: 19:28:36 | examples-with-cassandra | 2024-08-29 19:28:23 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-d4qs4 job-controller logger.go:42: 19:28:36 | examples-with-cassandra | 2024-08-29 19:28:23 +0000 UTC Normal Pod report-span-k49d7 Binding Scheduled Successfully assigned kuttl-test-peaceful-goat/report-span-k49d7 to ip-10-0-51-237.us-west-2.compute.internal default-scheduler logger.go:42: 19:28:36 | examples-with-cassandra | 2024-08-29 19:28:23 +0000 UTC Normal Pod report-span-k49d7 AddedInterface Add eth0 [10.130.0.22/23] from ovn-kubernetes multus logger.go:42: 19:28:36 | examples-with-cassandra | 2024-08-29 19:28:23 +0000 UTC Normal Pod report-span-k49d7.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 19:28:36 | examples-with-cassandra | 2024-08-29 19:28:23 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-k49d7 job-controller logger.go:42: 19:28:36 | examples-with-cassandra | 2024-08-29 19:28:24 +0000 UTC Normal Pod check-span-d4qs4 AddedInterface Add eth0 [10.130.0.23/23] from ovn-kubernetes multus logger.go:42: 19:28:36 | examples-with-cassandra | 2024-08-29 19:28:24 +0000 UTC Normal Pod check-span-d4qs4.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 19:28:36 | examples-with-cassandra | 2024-08-29 19:28:24 +0000 UTC Normal Pod report-span-k49d7.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 835ms (835ms including waiting) kubelet logger.go:42: 19:28:36 | examples-with-cassandra | 2024-08-29 19:28:24 +0000 UTC Normal Pod report-span-k49d7.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 19:28:36 | examples-with-cassandra | 2024-08-29 19:28:24 +0000 UTC Normal Pod report-span-k49d7.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 19:28:36 | examples-with-cassandra | 2024-08-29 19:28:25 +0000 UTC Normal Pod check-span-d4qs4.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 1.052s (1.052s including waiting) kubelet logger.go:42: 19:28:36 | examples-with-cassandra | 2024-08-29 19:28:25 +0000 UTC Normal Pod check-span-d4qs4.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 19:28:36 | examples-with-cassandra | 2024-08-29 19:28:25 +0000 UTC Normal Pod check-span-d4qs4.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 19:28:36 | examples-with-cassandra | 2024-08-29 19:28:36 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 19:28:36 | examples-with-cassandra | Deleting namespace: kuttl-test-peaceful-goat === CONT kuttl/harness/examples-with-badger logger.go:42: 19:28:48 | examples-with-badger | Creating namespace: kuttl-test-balanced-jackass logger.go:42: 19:28:48 | examples-with-badger/0-install | starting test step 0-install logger.go:42: 19:28:49 | examples-with-badger/0-install | Jaeger:kuttl-test-balanced-jackass/with-badger created logger.go:42: 19:28:51 | examples-with-badger/0-install | test step completed 0-install logger.go:42: 19:28:51 | examples-with-badger/1-smoke-test | starting test step 1-smoke-test logger.go:42: 19:28:51 | examples-with-badger/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE with-badger /dev/null] logger.go:42: 19:28:53 | examples-with-badger/1-smoke-test | Warning: resource jaegers/with-badger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 19:29:00 | examples-with-badger/1-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://with-badger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://with-badger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 19:29:01 | examples-with-badger/1-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 19:29:01 | examples-with-badger/1-smoke-test | job.batch/report-span created logger.go:42: 19:29:01 | examples-with-badger/1-smoke-test | job.batch/check-span created logger.go:42: 19:29:13 | examples-with-badger/1-smoke-test | test step completed 1-smoke-test logger.go:42: 19:29:13 | examples-with-badger | examples-with-badger events from ns kuttl-test-balanced-jackass: logger.go:42: 19:29:13 | examples-with-badger | 2024-08-29 19:28:49 +0000 UTC Normal Pod with-badger-6845f684bb-pw5tf Binding Scheduled Successfully assigned kuttl-test-balanced-jackass/with-badger-6845f684bb-pw5tf to ip-10-0-51-237.us-west-2.compute.internal default-scheduler logger.go:42: 19:29:13 | examples-with-badger | 2024-08-29 19:28:49 +0000 UTC Normal ReplicaSet.apps with-badger-6845f684bb SuccessfulCreate Created pod: with-badger-6845f684bb-pw5tf replicaset-controller logger.go:42: 19:29:13 | examples-with-badger | 2024-08-29 19:28:49 +0000 UTC Normal Deployment.apps with-badger ScalingReplicaSet Scaled up replica set with-badger-6845f684bb to 1 deployment-controller logger.go:42: 19:29:13 | examples-with-badger | 2024-08-29 19:28:50 +0000 UTC Normal Pod with-badger-6845f684bb-pw5tf AddedInterface Add eth0 [10.130.0.24/23] from ovn-kubernetes multus logger.go:42: 19:29:13 | examples-with-badger | 2024-08-29 19:28:50 +0000 UTC Normal Pod with-badger-6845f684bb-pw5tf.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:345d3b6c507df9f8a18948a594e7a03698085689c6965faf79b8aadc74a0ceaa" already present on machine kubelet logger.go:42: 19:29:13 | examples-with-badger | 2024-08-29 19:28:50 +0000 UTC Normal Pod with-badger-6845f684bb-pw5tf.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 19:29:13 | examples-with-badger | 2024-08-29 19:28:50 +0000 UTC Normal Pod with-badger-6845f684bb-pw5tf.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 19:29:13 | examples-with-badger | 2024-08-29 19:28:50 +0000 UTC Normal Pod with-badger-6845f684bb-pw5tf.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1613032062a68faa414cd5754a666b7b78ad0e52cf328eaaed355719423a55dd" already present on machine kubelet logger.go:42: 19:29:13 | examples-with-badger | 2024-08-29 19:28:50 +0000 UTC Normal Pod with-badger-6845f684bb-pw5tf.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 19:29:13 | examples-with-badger | 2024-08-29 19:28:50 +0000 UTC Normal Pod with-badger-6845f684bb-pw5tf.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 19:29:13 | examples-with-badger | 2024-08-29 19:28:53 +0000 UTC Normal ReplicaSet.apps with-badger-6845f684bb SuccessfulDelete Deleted pod: with-badger-6845f684bb-pw5tf replicaset-controller logger.go:42: 19:29:13 | examples-with-badger | 2024-08-29 19:28:53 +0000 UTC Normal Deployment.apps with-badger ScalingReplicaSet Scaled down replica set with-badger-6845f684bb to 0 from 1 deployment-controller logger.go:42: 19:29:13 | examples-with-badger | 2024-08-29 19:28:54 +0000 UTC Normal Pod with-badger-6845f684bb-pw5tf.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 19:29:13 | examples-with-badger | 2024-08-29 19:28:54 +0000 UTC Normal Pod with-badger-6845f684bb-pw5tf.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 19:29:13 | examples-with-badger | 2024-08-29 19:28:55 +0000 UTC Normal Pod with-badger-65cc5466d4-f4b55 Binding Scheduled Successfully assigned kuttl-test-balanced-jackass/with-badger-65cc5466d4-f4b55 to ip-10-0-51-237.us-west-2.compute.internal default-scheduler logger.go:42: 19:29:13 | examples-with-badger | 2024-08-29 19:28:55 +0000 UTC Normal Pod with-badger-65cc5466d4-f4b55 AddedInterface Add eth0 [10.130.0.25/23] from ovn-kubernetes multus logger.go:42: 19:29:13 | examples-with-badger | 2024-08-29 19:28:55 +0000 UTC Normal Pod with-badger-65cc5466d4-f4b55.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:345d3b6c507df9f8a18948a594e7a03698085689c6965faf79b8aadc74a0ceaa" already present on machine kubelet logger.go:42: 19:29:13 | examples-with-badger | 2024-08-29 19:28:55 +0000 UTC Normal ReplicaSet.apps with-badger-65cc5466d4 SuccessfulCreate Created pod: with-badger-65cc5466d4-f4b55 replicaset-controller logger.go:42: 19:29:13 | examples-with-badger | 2024-08-29 19:28:55 +0000 UTC Normal Deployment.apps with-badger ScalingReplicaSet Scaled up replica set with-badger-65cc5466d4 to 1 deployment-controller logger.go:42: 19:29:13 | examples-with-badger | 2024-08-29 19:28:56 +0000 UTC Normal Pod with-badger-65cc5466d4-f4b55.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 19:29:13 | examples-with-badger | 2024-08-29 19:28:56 +0000 UTC Normal Pod with-badger-65cc5466d4-f4b55.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 19:29:13 | examples-with-badger | 2024-08-29 19:28:56 +0000 UTC Normal Pod with-badger-65cc5466d4-f4b55.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1613032062a68faa414cd5754a666b7b78ad0e52cf328eaaed355719423a55dd" already present on machine kubelet logger.go:42: 19:29:13 | examples-with-badger | 2024-08-29 19:28:56 +0000 UTC Normal Pod with-badger-65cc5466d4-f4b55.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 19:29:13 | examples-with-badger | 2024-08-29 19:28:56 +0000 UTC Normal Pod with-badger-65cc5466d4-f4b55.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 19:29:13 | examples-with-badger | 2024-08-29 19:29:01 +0000 UTC Normal Pod check-span-g4bv4 Binding Scheduled Successfully assigned kuttl-test-balanced-jackass/check-span-g4bv4 to ip-10-0-51-237.us-west-2.compute.internal default-scheduler logger.go:42: 19:29:13 | examples-with-badger | 2024-08-29 19:29:01 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-g4bv4 job-controller logger.go:42: 19:29:13 | examples-with-badger | 2024-08-29 19:29:01 +0000 UTC Normal Pod report-span-frf9h Binding Scheduled Successfully assigned kuttl-test-balanced-jackass/report-span-frf9h to ip-10-0-51-237.us-west-2.compute.internal default-scheduler logger.go:42: 19:29:13 | examples-with-badger | 2024-08-29 19:29:01 +0000 UTC Normal Pod report-span-frf9h AddedInterface Add eth0 [10.130.0.26/23] from ovn-kubernetes multus logger.go:42: 19:29:13 | examples-with-badger | 2024-08-29 19:29:01 +0000 UTC Normal Pod report-span-frf9h.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 19:29:13 | examples-with-badger | 2024-08-29 19:29:01 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-frf9h job-controller logger.go:42: 19:29:13 | examples-with-badger | 2024-08-29 19:29:02 +0000 UTC Normal Pod check-span-g4bv4 AddedInterface Add eth0 [10.130.0.27/23] from ovn-kubernetes multus logger.go:42: 19:29:13 | examples-with-badger | 2024-08-29 19:29:02 +0000 UTC Normal Pod check-span-g4bv4.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 19:29:13 | examples-with-badger | 2024-08-29 19:29:02 +0000 UTC Normal Pod check-span-g4bv4.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 813ms (813ms including waiting) kubelet logger.go:42: 19:29:13 | examples-with-badger | 2024-08-29 19:29:02 +0000 UTC Normal Pod report-span-frf9h.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 820ms (820ms including waiting) kubelet logger.go:42: 19:29:13 | examples-with-badger | 2024-08-29 19:29:02 +0000 UTC Normal Pod report-span-frf9h.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 19:29:13 | examples-with-badger | 2024-08-29 19:29:02 +0000 UTC Normal Pod report-span-frf9h.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 19:29:13 | examples-with-badger | 2024-08-29 19:29:03 +0000 UTC Normal Pod check-span-g4bv4.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 19:29:13 | examples-with-badger | 2024-08-29 19:29:03 +0000 UTC Normal Pod check-span-g4bv4.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 19:29:13 | examples-with-badger | 2024-08-29 19:29:13 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 19:29:13 | examples-with-badger | Deleting namespace: kuttl-test-balanced-jackass === CONT kuttl/harness/examples-simplest logger.go:42: 19:29:25 | examples-simplest | Creating namespace: kuttl-test-willing-halibut logger.go:42: 19:29:25 | examples-simplest/0-install | starting test step 0-install logger.go:42: 19:29:26 | examples-simplest/0-install | Jaeger:kuttl-test-willing-halibut/simplest created logger.go:42: 19:29:28 | examples-simplest/0-install | test step completed 0-install logger.go:42: 19:29:28 | examples-simplest/1-smoke-test | starting test step 1-smoke-test logger.go:42: 19:29:28 | examples-simplest/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE simplest /dev/null] logger.go:42: 19:29:30 | examples-simplest/1-smoke-test | Warning: resource jaegers/simplest is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 19:29:37 | examples-simplest/1-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://simplest-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 19:29:38 | examples-simplest/1-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 19:29:38 | examples-simplest/1-smoke-test | job.batch/report-span created logger.go:42: 19:29:38 | examples-simplest/1-smoke-test | job.batch/check-span created logger.go:42: 19:29:50 | examples-simplest/1-smoke-test | test step completed 1-smoke-test logger.go:42: 19:29:50 | examples-simplest | examples-simplest events from ns kuttl-test-willing-halibut: logger.go:42: 19:29:50 | examples-simplest | 2024-08-29 19:29:26 +0000 UTC Normal Pod simplest-57d8f659c9-fxrbq Binding Scheduled Successfully assigned kuttl-test-willing-halibut/simplest-57d8f659c9-fxrbq to ip-10-0-51-237.us-west-2.compute.internal default-scheduler logger.go:42: 19:29:50 | examples-simplest | 2024-08-29 19:29:26 +0000 UTC Normal ReplicaSet.apps simplest-57d8f659c9 SuccessfulCreate Created pod: simplest-57d8f659c9-fxrbq replicaset-controller logger.go:42: 19:29:50 | examples-simplest | 2024-08-29 19:29:26 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled up replica set simplest-57d8f659c9 to 1 deployment-controller logger.go:42: 19:29:50 | examples-simplest | 2024-08-29 19:29:27 +0000 UTC Normal Pod simplest-57d8f659c9-fxrbq AddedInterface Add eth0 [10.130.0.28/23] from ovn-kubernetes multus logger.go:42: 19:29:50 | examples-simplest | 2024-08-29 19:29:27 +0000 UTC Normal Pod simplest-57d8f659c9-fxrbq.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:345d3b6c507df9f8a18948a594e7a03698085689c6965faf79b8aadc74a0ceaa" already present on machine kubelet logger.go:42: 19:29:50 | examples-simplest | 2024-08-29 19:29:27 +0000 UTC Normal Pod simplest-57d8f659c9-fxrbq.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 19:29:50 | examples-simplest | 2024-08-29 19:29:27 +0000 UTC Normal Pod simplest-57d8f659c9-fxrbq.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 19:29:50 | examples-simplest | 2024-08-29 19:29:27 +0000 UTC Normal Pod simplest-57d8f659c9-fxrbq.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1613032062a68faa414cd5754a666b7b78ad0e52cf328eaaed355719423a55dd" already present on machine kubelet logger.go:42: 19:29:50 | examples-simplest | 2024-08-29 19:29:27 +0000 UTC Normal Pod simplest-57d8f659c9-fxrbq.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 19:29:50 | examples-simplest | 2024-08-29 19:29:27 +0000 UTC Normal Pod simplest-57d8f659c9-fxrbq.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 19:29:50 | examples-simplest | 2024-08-29 19:29:30 +0000 UTC Normal ReplicaSet.apps simplest-57d8f659c9 SuccessfulDelete Deleted pod: simplest-57d8f659c9-fxrbq replicaset-controller logger.go:42: 19:29:50 | examples-simplest | 2024-08-29 19:29:30 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled down replica set simplest-57d8f659c9 to 0 from 1 deployment-controller logger.go:42: 19:29:50 | examples-simplest | 2024-08-29 19:29:31 +0000 UTC Normal Pod simplest-57d8f659c9-fxrbq.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 19:29:50 | examples-simplest | 2024-08-29 19:29:31 +0000 UTC Normal Pod simplest-57d8f659c9-fxrbq.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 19:29:50 | examples-simplest | 2024-08-29 19:29:32 +0000 UTC Normal Pod simplest-65c7f7fb4-p9q6v Binding Scheduled Successfully assigned kuttl-test-willing-halibut/simplest-65c7f7fb4-p9q6v to ip-10-0-51-237.us-west-2.compute.internal default-scheduler logger.go:42: 19:29:50 | examples-simplest | 2024-08-29 19:29:32 +0000 UTC Normal Pod simplest-65c7f7fb4-p9q6v AddedInterface Add eth0 [10.130.0.29/23] from ovn-kubernetes multus logger.go:42: 19:29:50 | examples-simplest | 2024-08-29 19:29:32 +0000 UTC Normal Pod simplest-65c7f7fb4-p9q6v.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:345d3b6c507df9f8a18948a594e7a03698085689c6965faf79b8aadc74a0ceaa" already present on machine kubelet logger.go:42: 19:29:50 | examples-simplest | 2024-08-29 19:29:32 +0000 UTC Normal ReplicaSet.apps simplest-65c7f7fb4 SuccessfulCreate Created pod: simplest-65c7f7fb4-p9q6v replicaset-controller logger.go:42: 19:29:50 | examples-simplest | 2024-08-29 19:29:32 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled up replica set simplest-65c7f7fb4 to 1 deployment-controller logger.go:42: 19:29:50 | examples-simplest | 2024-08-29 19:29:33 +0000 UTC Normal Pod simplest-65c7f7fb4-p9q6v.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 19:29:50 | examples-simplest | 2024-08-29 19:29:33 +0000 UTC Normal Pod simplest-65c7f7fb4-p9q6v.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 19:29:50 | examples-simplest | 2024-08-29 19:29:33 +0000 UTC Normal Pod simplest-65c7f7fb4-p9q6v.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1613032062a68faa414cd5754a666b7b78ad0e52cf328eaaed355719423a55dd" already present on machine kubelet logger.go:42: 19:29:50 | examples-simplest | 2024-08-29 19:29:33 +0000 UTC Normal Pod simplest-65c7f7fb4-p9q6v.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 19:29:50 | examples-simplest | 2024-08-29 19:29:33 +0000 UTC Normal Pod simplest-65c7f7fb4-p9q6v.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 19:29:50 | examples-simplest | 2024-08-29 19:29:38 +0000 UTC Normal Pod check-span-pzrg7 Binding Scheduled Successfully assigned kuttl-test-willing-halibut/check-span-pzrg7 to ip-10-0-51-237.us-west-2.compute.internal default-scheduler logger.go:42: 19:29:50 | examples-simplest | 2024-08-29 19:29:38 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-pzrg7 job-controller logger.go:42: 19:29:50 | examples-simplest | 2024-08-29 19:29:38 +0000 UTC Normal Pod report-span-n9vfw Binding Scheduled Successfully assigned kuttl-test-willing-halibut/report-span-n9vfw to ip-10-0-51-237.us-west-2.compute.internal default-scheduler logger.go:42: 19:29:50 | examples-simplest | 2024-08-29 19:29:38 +0000 UTC Normal Pod report-span-n9vfw AddedInterface Add eth0 [10.130.0.30/23] from ovn-kubernetes multus logger.go:42: 19:29:50 | examples-simplest | 2024-08-29 19:29:38 +0000 UTC Normal Pod report-span-n9vfw.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 19:29:50 | examples-simplest | 2024-08-29 19:29:38 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-n9vfw job-controller logger.go:42: 19:29:50 | examples-simplest | 2024-08-29 19:29:39 +0000 UTC Normal Pod check-span-pzrg7 AddedInterface Add eth0 [10.130.0.31/23] from ovn-kubernetes multus logger.go:42: 19:29:50 | examples-simplest | 2024-08-29 19:29:39 +0000 UTC Normal Pod check-span-pzrg7.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 19:29:50 | examples-simplest | 2024-08-29 19:29:39 +0000 UTC Normal Pod check-span-pzrg7.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 844ms (844ms including waiting) kubelet logger.go:42: 19:29:50 | examples-simplest | 2024-08-29 19:29:39 +0000 UTC Normal Pod report-span-n9vfw.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 861ms (861ms including waiting) kubelet logger.go:42: 19:29:50 | examples-simplest | 2024-08-29 19:29:39 +0000 UTC Normal Pod report-span-n9vfw.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 19:29:50 | examples-simplest | 2024-08-29 19:29:39 +0000 UTC Normal Pod report-span-n9vfw.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 19:29:50 | examples-simplest | 2024-08-29 19:29:40 +0000 UTC Normal Pod check-span-pzrg7.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 19:29:50 | examples-simplest | 2024-08-29 19:29:40 +0000 UTC Normal Pod check-span-pzrg7.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 19:29:50 | examples-simplest | 2024-08-29 19:29:50 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 19:29:50 | examples-simplest | Deleting namespace: kuttl-test-willing-halibut === CONT kuttl/harness/examples-simple-prod-with-volumes logger.go:42: 19:30:02 | examples-simple-prod-with-volumes | Ignoring 03-check-volume.yaml.template as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 19:30:02 | examples-simple-prod-with-volumes | Creating namespace: kuttl-test-dear-weasel logger.go:42: 19:30:02 | examples-simple-prod-with-volumes/1-install | starting test step 1-install logger.go:42: 19:30:03 | examples-simple-prod-with-volumes/1-install | Jaeger:kuttl-test-dear-weasel/simple-prod created logger.go:42: 19:40:03 | examples-simple-prod-with-volumes/1-install | test step failed 1-install case.go:364: failed in step 1-install case.go:366: --- Deployment:kuttl-test-dear-weasel/simple-prod-collector +++ Deployment:kuttl-test-dear-weasel/simple-prod-collector @@ -1,10 +1,424 @@ apiVersion: apps/v1 kind: Deployment metadata: + annotations: + linkerd.io/inject: disabled + prometheus.io/port: "14269" + prometheus.io/scrape: "true" + labels: + app: jaeger + app.kubernetes.io/component: collector + app.kubernetes.io/instance: simple-prod + app.kubernetes.io/managed-by: jaeger-operator + app.kubernetes.io/name: simple-prod-collector + app.kubernetes.io/part-of: jaeger + managedFields: + - apiVersion: apps/v1 + fieldsType: FieldsV1 + fieldsV1: + f:metadata: + f:annotations: + .: {} + f:linkerd.io/inject: {} + f:prometheus.io/port: {} + f:prometheus.io/scrape: {} + f:labels: + .: {} + f:app: {} + f:app.kubernetes.io/component: {} + f:app.kubernetes.io/instance: {} + f:app.kubernetes.io/managed-by: {} + f:app.kubernetes.io/name: {} + f:app.kubernetes.io/part-of: {} + f:ownerReferences: + .: {} + k:{"uid":"4873482e-d0aa-4d58-b49a-985e7050146c"}: {} + f:spec: + f:progressDeadlineSeconds: {} + f:replicas: {} + f:revisionHistoryLimit: {} + f:selector: {} + f:strategy: + f:type: {} + f:template: + f:metadata: + f:annotations: + .: {} + f:linkerd.io/inject: {} + f:prometheus.io/port: {} + f:prometheus.io/scrape: {} + f:sidecar.istio.io/inject: {} + f:labels: + .: {} + f:app: {} + f:app.kubernetes.io/component: {} + f:app.kubernetes.io/instance: {} + f:app.kubernetes.io/managed-by: {} + f:app.kubernetes.io/name: {} + f:app.kubernetes.io/part-of: {} + f:spec: + f:containers: + k:{"name":"jaeger-collector"}: + .: {} + f:args: {} + f:env: + .: {} + k:{"name":"COLLECTOR_OTLP_ENABLED"}: + .: {} + f:name: {} + f:value: {} + k:{"name":"COLLECTOR_ZIPKIN_HOST_PORT"}: + .: {} + f:name: {} + f:value: {} + k:{"name":"SPAN_STORAGE_TYPE"}: + .: {} + f:name: {} + f:value: {} + f:image: {} + f:imagePullPolicy: {} + f:livenessProbe: + .: {} + f:failureThreshold: {} + f:httpGet: + .: {} + f:path: {} + f:port: {} + f:scheme: {} + f:initialDelaySeconds: {} + f:periodSeconds: {} + f:successThreshold: {} + f:timeoutSeconds: {} + f:name: {} + f:ports: + .: {} + k:{"containerPort":4317,"protocol":"TCP"}: + .: {} + f:containerPort: {} + f:name: {} + f:protocol: {} + k:{"containerPort":4318,"protocol":"TCP"}: + .: {} + f:containerPort: {} + f:name: {} + f:protocol: {} + k:{"containerPort":9411,"protocol":"TCP"}: + .: {} + f:containerPort: {} + f:name: {} + f:protocol: {} + k:{"containerPort":14250,"protocol":"TCP"}: + .: {} + f:containerPort: {} + f:name: {} + f:protocol: {} + k:{"containerPort":14267,"protocol":"TCP"}: + .: {} + f:containerPort: {} + f:name: {} + f:protocol: {} + k:{"containerPort":14268,"protocol":"TCP"}: + .: {} + f:containerPort: {} + f:name: {} + f:protocol: {} + k:{"containerPort":14269,"protocol":"TCP"}: + .: {} + f:containerPort: {} + f:name: {} + f:protocol: {} + f:readinessProbe: + .: {} + f:failureThreshold: {} + f:httpGet: + .: {} + f:path: {} + f:port: {} + f:scheme: {} + f:initialDelaySeconds: {} + f:periodSeconds: {} + f:successThreshold: {} + f:timeoutSeconds: {} + f:resources: {} + f:terminationMessagePath: {} + f:terminationMessagePolicy: {} + f:volumeMounts: + .: {} + k:{"mountPath":"/certs"}: + .: {} + f:mountPath: {} + f:name: {} + f:readOnly: {} + k:{"mountPath":"/etc/jaeger/sampling"}: + .: {} + f:mountPath: {} + f:name: {} + f:readOnly: {} + k:{"mountPath":"/etc/pki/ca-trust/extracted/pem"}: + .: {} + f:mountPath: {} + f:name: {} + f:readOnly: {} + k:{"mountPath":"/etc/tls-config"}: + .: {} + f:mountPath: {} + f:name: {} + f:readOnly: {} + k:{"mountPath":"/usr/share/elasticsearch/data"}: + .: {} + f:mountPath: {} + f:name: {} + f:dnsPolicy: {} + f:enableServiceLinks: {} + f:restartPolicy: {} + f:schedulerName: {} + f:securityContext: {} + f:serviceAccount: {} + f:serviceAccountName: {} + f:terminationGracePeriodSeconds: {} + f:volumes: + .: {} + k:{"name":"certs"}: + .: {} + f:name: {} + f:secret: + .: {} + f:defaultMode: {} + f:secretName: {} + k:{"name":"elastic-data"}: + .: {} + f:emptyDir: {} + f:name: {} + k:{"name":"simple-prod-collector-tls-config-volume"}: + .: {} + f:name: {} + f:secret: + .: {} + f:defaultMode: {} + f:secretName: {} + k:{"name":"simple-prod-sampling-configuration-volume"}: + .: {} + f:configMap: + .: {} + f:defaultMode: {} + f:items: {} + f:name: {} + f:name: {} + k:{"name":"simple-prod-trusted-ca"}: + .: {} + f:configMap: + .: {} + f:defaultMode: {} + f:items: {} + f:name: {} + f:name: {} + manager: jaeger-operator + operation: Update + time: "2024-08-29T19:30:06Z" + - apiVersion: apps/v1 + fieldsType: FieldsV1 + fieldsV1: + f:metadata: + f:annotations: + f:deployment.kubernetes.io/revision: {} + f:status: + f:conditions: + .: {} + k:{"type":"Available"}: + .: {} + f:lastTransitionTime: {} + f:lastUpdateTime: {} + f:message: {} + f:reason: {} + f:status: {} + f:type: {} + k:{"type":"Progressing"}: + .: {} + f:lastTransitionTime: {} + f:lastUpdateTime: {} + f:message: {} + f:reason: {} + f:status: {} + f:type: {} + f:observedGeneration: {} + f:replicas: {} + f:unavailableReplicas: {} + f:updatedReplicas: {} + manager: kube-controller-manager + operation: Update + subresource: status + time: "2024-08-29T19:30:06Z" name: simple-prod-collector namespace: kuttl-test-dear-weasel + ownerReferences: + - apiVersion: jaegertracing.io/v1 + controller: true + kind: Jaeger + name: simple-prod + uid: 4873482e-d0aa-4d58-b49a-985e7050146c spec: + progressDeadlineSeconds: 600 replicas: 1 + revisionHistoryLimit: 10 + selector: + matchLabels: + app: jaeger + app.kubernetes.io/component: collector + app.kubernetes.io/instance: simple-prod + app.kubernetes.io/managed-by: jaeger-operator + app.kubernetes.io/name: simple-prod-collector + app.kubernetes.io/part-of: jaeger + strategy: + type: Recreate + template: + metadata: + annotations: + linkerd.io/inject: disabled + prometheus.io/port: "14269" + prometheus.io/scrape: "true" + sidecar.istio.io/inject: "false" + creationTimestamp: null + labels: + app: jaeger + app.kubernetes.io/component: collector + app.kubernetes.io/instance: simple-prod + app.kubernetes.io/managed-by: jaeger-operator + app.kubernetes.io/name: simple-prod-collector + app.kubernetes.io/part-of: jaeger + spec: + containers: + - args: + - --collector.grpc.tls.cert=/etc/tls-config/tls.crt + - --collector.grpc.tls.enabled=true + - --collector.grpc.tls.key=/etc/tls-config/tls.key + - --sampling.strategies-file=/etc/jaeger/sampling/sampling.json + - --es.server-urls=https://elasticsearch.kuttl-test-dear-weasel.svc.cluster.local:9200 + - --es.tls.enabled=true + - --es.tls.ca=/certs/ca + - --es.tls.cert=/certs/cert + - --es.tls.key=/certs/key + - --es.timeout=15s + - --es.num-shards=1 + - --es.num-replicas=0 + env: + - name: SPAN_STORAGE_TYPE + value: elasticsearch + - name: COLLECTOR_ZIPKIN_HOST_PORT + value: :9411 + - name: COLLECTOR_OTLP_ENABLED + value: "true" + image: registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:b90e0d264bb0b051c2a4b9e6d499f407fed58f834ab8ebf30263ee1cf7c7ea3c + imagePullPolicy: IfNotPresent + livenessProbe: + failureThreshold: 5 + httpGet: + path: / + port: 14269 + scheme: HTTP + initialDelaySeconds: 5 + periodSeconds: 15 + successThreshold: 1 + timeoutSeconds: 1 + name: jaeger-collector + ports: + - containerPort: 9411 + name: zipkin + protocol: TCP + - containerPort: 14267 + name: c-tchan-trft + protocol: TCP + - containerPort: 14268 + name: c-binary-trft + protocol: TCP + - containerPort: 14269 + name: admin-http + protocol: TCP + - containerPort: 14250 + name: grpc + protocol: TCP + - containerPort: 4317 + name: grpc-otlp + protocol: TCP + - containerPort: 4318 + name: http-otlp + protocol: TCP + readinessProbe: + failureThreshold: 3 + httpGet: + path: / + port: 14269 + scheme: HTTP + initialDelaySeconds: 1 + periodSeconds: 10 + successThreshold: 1 + timeoutSeconds: 1 + resources: {} + terminationMessagePath: /dev/termination-log + terminationMessagePolicy: File + volumeMounts: + - mountPath: /usr/share/elasticsearch/data + name: elastic-data + - mountPath: /etc/jaeger/sampling + name: simple-prod-sampling-configuration-volume + readOnly: true + - mountPath: /etc/tls-config + name: simple-prod-collector-tls-config-volume + readOnly: true + - mountPath: /etc/pki/ca-trust/extracted/pem + name: simple-prod-trusted-ca + readOnly: true + - mountPath: /certs + name: certs + readOnly: true + dnsPolicy: ClusterFirst + enableServiceLinks: false + restartPolicy: Always + schedulerName: default-scheduler + securityContext: {} + serviceAccount: simple-prod + serviceAccountName: simple-prod + terminationGracePeriodSeconds: 30 + volumes: + - emptyDir: {} + name: elastic-data + - configMap: + defaultMode: 420 + items: + - key: sampling + path: sampling.json + name: simple-prod-sampling-configuration + name: simple-prod-sampling-configuration-volume + - name: simple-prod-collector-tls-config-volume + secret: + defaultMode: 420 + secretName: simple-prod-collector-headless-tls + - configMap: + defaultMode: 420 + items: + - key: ca-bundle.crt + path: tls-ca-bundle.pem + name: simple-prod-trusted-ca + name: simple-prod-trusted-ca + - name: certs + secret: + defaultMode: 420 + secretName: simple-prod-jaeger-elasticsearch status: - readyReplicas: 1 + conditions: + - lastTransitionTime: "2024-08-29T19:30:06Z" + lastUpdateTime: "2024-08-29T19:30:06Z" + message: Deployment does not have minimum availability. + reason: MinimumReplicasUnavailable + status: "False" + type: Available + - lastTransitionTime: "2024-08-29T19:30:06Z" + lastUpdateTime: "2024-08-29T19:30:06Z" + message: ReplicaSet "simple-prod-collector-dc4588778" is progressing. + reason: ReplicaSetUpdated + status: "True" + type: Progressing + observedGeneration: 1 + replicas: 1 + unavailableReplicas: 1 + updatedReplicas: 1 case.go:366: resource Deployment:kuttl-test-dear-weasel/simple-prod-collector: .status.readyReplicas: key is missing from map case.go:366: --- Deployment:kuttl-test-dear-weasel/simple-prod-query +++ Deployment:kuttl-test-dear-weasel/simple-prod-query @@ -1,10 +1,520 @@ apiVersion: apps/v1 kind: Deployment metadata: + annotations: + linkerd.io/inject: disabled + prometheus.io/port: "16687" + prometheus.io/scrape: "true" + sidecar.jaegertracing.io/inject: simple-prod + sidecar.jaegertracing.io/revision: "1" + labels: + app: jaeger + app.kubernetes.io/component: query + app.kubernetes.io/instance: simple-prod + app.kubernetes.io/managed-by: jaeger-operator + app.kubernetes.io/name: simple-prod-query + app.kubernetes.io/part-of: jaeger + sidecar.jaegertracing.io/injected: simple-prod + managedFields: + - apiVersion: apps/v1 + fieldsType: FieldsV1 + fieldsV1: + f:metadata: + f:annotations: + .: {} + f:linkerd.io/inject: {} + f:prometheus.io/port: {} + f:prometheus.io/scrape: {} + f:sidecar.jaegertracing.io/inject: {} + f:sidecar.jaegertracing.io/revision: {} + f:labels: + .: {} + f:app: {} + f:app.kubernetes.io/component: {} + f:app.kubernetes.io/instance: {} + f:app.kubernetes.io/managed-by: {} + f:app.kubernetes.io/name: {} + f:app.kubernetes.io/part-of: {} + f:ownerReferences: + .: {} + k:{"uid":"4873482e-d0aa-4d58-b49a-985e7050146c"}: {} + f:spec: + f:progressDeadlineSeconds: {} + f:replicas: {} + f:revisionHistoryLimit: {} + f:selector: {} + f:strategy: + f:type: {} + f:template: + f:metadata: + f:annotations: + .: {} + f:linkerd.io/inject: {} + f:prometheus.io/port: {} + f:prometheus.io/scrape: {} + f:sidecar.istio.io/inject: {} + f:sidecar.jaegertracing.io/inject: {} + f:labels: + .: {} + f:app: {} + f:app.kubernetes.io/component: {} + f:app.kubernetes.io/instance: {} + f:app.kubernetes.io/managed-by: {} + f:app.kubernetes.io/name: {} + f:app.kubernetes.io/part-of: {} + f:spec: + f:containers: + k:{"name":"jaeger-query"}: + .: {} + f:args: {} + f:env: + .: {} + k:{"name":"JAEGER_DISABLED"}: + .: {} + f:name: {} + f:value: {} + k:{"name":"METRICS_STORAGE_TYPE"}: + .: {} + f:name: {} + k:{"name":"SPAN_STORAGE_TYPE"}: + .: {} + f:name: {} + f:value: {} + f:image: {} + f:imagePullPolicy: {} + f:livenessProbe: + .: {} + f:failureThreshold: {} + f:httpGet: + .: {} + f:path: {} + f:port: {} + f:scheme: {} + f:initialDelaySeconds: {} + f:periodSeconds: {} + f:successThreshold: {} + f:timeoutSeconds: {} + f:name: {} + f:ports: + .: {} + k:{"containerPort":16685,"protocol":"TCP"}: + .: {} + f:containerPort: {} + f:name: {} + f:protocol: {} + k:{"containerPort":16686,"protocol":"TCP"}: + .: {} + f:containerPort: {} + f:name: {} + f:protocol: {} + k:{"containerPort":16687,"protocol":"TCP"}: + .: {} + f:containerPort: {} + f:name: {} + f:protocol: {} + f:readinessProbe: + .: {} + f:failureThreshold: {} + f:httpGet: + .: {} + f:path: {} + f:port: {} + f:scheme: {} + f:initialDelaySeconds: {} + f:periodSeconds: {} + f:successThreshold: {} + f:timeoutSeconds: {} + f:resources: {} + f:terminationMessagePath: {} + f:terminationMessagePolicy: {} + f:volumeMounts: + .: {} + k:{"mountPath":"/certs"}: + .: {} + f:mountPath: {} + f:name: {} + f:readOnly: {} + k:{"mountPath":"/etc/config"}: + .: {} + f:mountPath: {} + f:name: {} + f:readOnly: {} + k:{"mountPath":"/etc/pki/ca-trust/extracted/pem"}: + .: {} + f:mountPath: {} + f:name: {} + f:readOnly: {} + k:{"mountPath":"/usr/share/elasticsearch/data"}: + .: {} + f:mountPath: {} + f:name: {} + k:{"name":"oauth-proxy"}: + .: {} + f:args: {} + f:image: {} + f:imagePullPolicy: {} + f:name: {} + f:ports: + .: {} + k:{"containerPort":8443,"protocol":"TCP"}: + .: {} + f:containerPort: {} + f:name: {} + f:protocol: {} + f:resources: {} + f:terminationMessagePath: {} + f:terminationMessagePolicy: {} + f:volumeMounts: + .: {} + k:{"mountPath":"/etc/pki/ca-trust/extracted/pem"}: + .: {} + f:mountPath: {} + f:name: {} + f:readOnly: {} + k:{"mountPath":"/etc/tls/private"}: + .: {} + f:mountPath: {} + f:name: {} + f:dnsPolicy: {} + f:enableServiceLinks: {} + f:restartPolicy: {} + f:schedulerName: {} + f:securityContext: {} + f:serviceAccount: {} + f:serviceAccountName: {} + f:terminationGracePeriodSeconds: {} + f:volumes: + .: {} + k:{"name":"certs"}: + .: {} + f:name: {} + f:secret: + .: {} + f:defaultMode: {} + f:secretName: {} + k:{"name":"elastic-data"}: + .: {} + f:emptyDir: {} + f:name: {} + k:{"name":"simple-prod-trusted-ca"}: + .: {} + f:configMap: + .: {} + f:defaultMode: {} + f:items: {} + f:name: {} + f:name: {} + k:{"name":"simple-prod-ui-configuration-volume"}: + .: {} + f:configMap: + .: {} + f:defaultMode: {} + f:items: {} + f:name: {} + f:name: {} + k:{"name":"simple-prod-ui-oauth-proxy-tls"}: + .: {} + f:name: {} + f:secret: + .: {} + f:defaultMode: {} + f:secretName: {} + manager: jaeger-operator + operation: Update + time: "2024-08-29T19:35:06Z" + - apiVersion: apps/v1 + fieldsType: FieldsV1 + fieldsV1: + f:metadata: + f:annotations: + f:deployment.kubernetes.io/revision: {} + f:status: + f:conditions: + .: {} + k:{"type":"Available"}: + .: {} + f:lastTransitionTime: {} + f:lastUpdateTime: {} + f:message: {} + f:reason: {} + f:status: {} + f:type: {} + k:{"type":"Progressing"}: + .: {} + f:lastTransitionTime: {} + f:lastUpdateTime: {} + f:message: {} + f:reason: {} + f:status: {} + f:type: {} + f:observedGeneration: {} + f:replicas: {} + f:unavailableReplicas: {} + f:updatedReplicas: {} + manager: kube-controller-manager + operation: Update + subresource: status + time: "2024-08-29T19:35:06Z" name: simple-prod-query namespace: kuttl-test-dear-weasel + ownerReferences: + - apiVersion: jaegertracing.io/v1 + controller: true + kind: Jaeger + name: simple-prod + uid: 4873482e-d0aa-4d58-b49a-985e7050146c spec: + progressDeadlineSeconds: 600 replicas: 1 + revisionHistoryLimit: 10 + selector: + matchLabels: + app: jaeger + app.kubernetes.io/component: query + app.kubernetes.io/instance: simple-prod + app.kubernetes.io/managed-by: jaeger-operator + app.kubernetes.io/name: simple-prod-query + app.kubernetes.io/part-of: jaeger + strategy: + type: Recreate + template: + metadata: + annotations: + linkerd.io/inject: disabled + prometheus.io/port: "16687" + prometheus.io/scrape: "true" + sidecar.istio.io/inject: "false" + sidecar.jaegertracing.io/inject: simple-prod + creationTimestamp: null + labels: + app: jaeger + app.kubernetes.io/component: query + app.kubernetes.io/instance: simple-prod + app.kubernetes.io/managed-by: jaeger-operator + app.kubernetes.io/name: simple-prod-query + app.kubernetes.io/part-of: jaeger + spec: + containers: + - args: + - --query.ui-config=/etc/config/ui.json + - --es.server-urls=https://elasticsearch.kuttl-test-dear-weasel.svc.cluster.local:9200 + - --es.tls.enabled=true + - --es.tls.ca=/certs/ca + - --es.tls.cert=/certs/cert + - --es.tls.key=/certs/key + - --es.timeout=15s + - --es.num-shards=1 + - --es.num-replicas=0 + env: + - name: SPAN_STORAGE_TYPE + value: elasticsearch + - name: METRICS_STORAGE_TYPE + - name: JAEGER_DISABLED + value: "false" + - name: JAEGER_SERVICE_NAME + value: simple-prod.kuttl-test-dear-weasel + - name: JAEGER_PROPAGATION + value: jaeger,b3,w3c + image: registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:831f83e8396d61822fb2dd576e711e0f10c1f8c7bd03bf0d4bec49dfb90cacca + imagePullPolicy: IfNotPresent + livenessProbe: + failureThreshold: 5 + httpGet: + path: / + port: 16687 + scheme: HTTP + initialDelaySeconds: 5 + periodSeconds: 15 + successThreshold: 1 + timeoutSeconds: 1 + name: jaeger-query + ports: + - containerPort: 16685 + name: grpc-query + protocol: TCP + - containerPort: 16686 + name: query + protocol: TCP + - containerPort: 16687 + name: admin-http + protocol: TCP + readinessProbe: + failureThreshold: 3 + httpGet: + path: / + port: 16687 + scheme: HTTP + initialDelaySeconds: 1 + periodSeconds: 10 + successThreshold: 1 + timeoutSeconds: 1 + resources: {} + terminationMessagePath: /dev/termination-log + terminationMessagePolicy: File + volumeMounts: + - mountPath: /usr/share/elasticsearch/data + name: elastic-data + - mountPath: /etc/config + name: simple-prod-ui-configuration-volume + readOnly: true + - mountPath: /etc/pki/ca-trust/extracted/pem + name: simple-prod-trusted-ca + readOnly: true + - mountPath: /certs + name: certs + readOnly: true + - args: + - --cookie-secret=kvHILP06mMfMPgbE1KWQuu + - --https-address=:8443 + - '--openshift-sar={"namespace": "kuttl-test-dear-weasel", "resource": "pods", + "verb": "get"}' + - --openshift-service-account=simple-prod-ui-proxy + - --provider=openshift + - --tls-cert=/etc/tls/private/tls.crt + - --tls-key=/etc/tls/private/tls.key + - --upstream=http://localhost:16686 + env: + - name: JAEGER_SERVICE_NAME + value: simple-prod.kuttl-test-dear-weasel + - name: JAEGER_PROPAGATION + value: jaeger,b3,w3c + image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1613032062a68faa414cd5754a666b7b78ad0e52cf328eaaed355719423a55dd + imagePullPolicy: IfNotPresent + name: oauth-proxy + ports: + - containerPort: 8443 + name: public + protocol: TCP + resources: {} + terminationMessagePath: /dev/termination-log + terminationMessagePolicy: File + volumeMounts: + - mountPath: /etc/tls/private + name: simple-prod-ui-oauth-proxy-tls + - mountPath: /etc/pki/ca-trust/extracted/pem + name: simple-prod-trusted-ca + readOnly: true + - args: + - --agent.tags=cluster=undefined,deployment.name=simple-prod-query,host.ip=${HOST_IP:},pod.name=${POD_NAME:},pod.namespace=kuttl-test-dear-weasel + - --reporter.grpc.host-port=dns:///simple-prod-collector-headless.kuttl-test-dear-weasel.svc:14250 + - --reporter.grpc.tls.ca=/etc/pki/ca-trust/source/service-ca/service-ca.crt + - --reporter.grpc.tls.enabled=true + env: + - name: POD_NAME + valueFrom: + fieldRef: + apiVersion: v1 + fieldPath: metadata.name + - name: HOST_IP + valueFrom: + fieldRef: + apiVersion: v1 + fieldPath: status.hostIP + image: registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:5fcc9d365678f54c0f52005318dbbc32b87f743cae7e1a84568e769653f3b128 + imagePullPolicy: IfNotPresent + livenessProbe: + failureThreshold: 5 + httpGet: + path: / + port: 14271 + scheme: HTTP + initialDelaySeconds: 5 + periodSeconds: 15 + successThreshold: 1 + timeoutSeconds: 1 + name: jaeger-agent + ports: + - containerPort: 5775 + name: zk-compact-trft + protocol: UDP + - containerPort: 5778 + name: config-rest + protocol: TCP + - containerPort: 6831 + name: jg-compact-trft + protocol: UDP + - containerPort: 6832 + name: jg-binary-trft + protocol: UDP + - containerPort: 14271 + name: admin-http + protocol: TCP + readinessProbe: + failureThreshold: 3 + httpGet: + path: / + port: 14271 + scheme: HTTP + initialDelaySeconds: 1 + periodSeconds: 10 + successThreshold: 1 + timeoutSeconds: 1 + resources: {} + terminationMessagePath: /dev/termination-log + terminationMessagePolicy: File + volumeMounts: + - mountPath: /etc/pki/ca-trust/extracted/pem + name: simple-prod-trusted-ca + readOnly: true + - mountPath: /etc/pki/ca-trust/source/service-ca + name: simple-prod-service-ca + readOnly: true + dnsPolicy: ClusterFirst + enableServiceLinks: false + restartPolicy: Always + schedulerName: default-scheduler + securityContext: {} + serviceAccount: simple-prod-ui-proxy + serviceAccountName: simple-prod-ui-proxy + terminationGracePeriodSeconds: 30 + volumes: + - emptyDir: {} + name: elastic-data + - configMap: + defaultMode: 420 + items: + - key: ui + path: ui.json + name: simple-prod-ui-configuration + name: simple-prod-ui-configuration-volume + - configMap: + defaultMode: 420 + items: + - key: ca-bundle.crt + path: tls-ca-bundle.pem + name: simple-prod-trusted-ca + name: simple-prod-trusted-ca + - name: simple-prod-ui-oauth-proxy-tls + secret: + defaultMode: 420 + secretName: simple-prod-ui-oauth-proxy-tls + - name: certs + secret: + defaultMode: 420 + secretName: simple-prod-jaeger-elasticsearch + - configMap: + defaultMode: 420 + items: + - key: service-ca.crt + path: service-ca.crt + name: simple-prod-service-ca + name: simple-prod-service-ca status: - readyReplicas: 1 + conditions: + - lastTransitionTime: "2024-08-29T19:30:06Z" + lastUpdateTime: "2024-08-29T19:30:06Z" + message: Deployment does not have minimum availability. + reason: MinimumReplicasUnavailable + status: "False" + type: Available + - lastTransitionTime: "2024-08-29T19:30:06Z" + lastUpdateTime: "2024-08-29T19:30:06Z" + message: ReplicaSet "simple-prod-query-778f6d88d8" is progressing. + reason: ReplicaSetUpdated + status: "True" + type: Progressing + observedGeneration: 3 + replicas: 1 + unavailableReplicas: 1 + updatedReplicas: 1 case.go:366: resource Deployment:kuttl-test-dear-weasel/simple-prod-query: .status.readyReplicas: key is missing from map logger.go:42: 19:40:03 | examples-simple-prod-with-volumes | examples-simple-prod-with-volumes events from ns kuttl-test-dear-weasel: logger.go:42: 19:40:03 | examples-simple-prod-with-volumes | 2024-08-29 19:30:06 +0000 UTC Normal Pod simple-prod-collector-dc4588778-7xtx2 Binding Scheduled Successfully assigned kuttl-test-dear-weasel/simple-prod-collector-dc4588778-7xtx2 to ip-10-0-51-237.us-west-2.compute.internal default-scheduler logger.go:42: 19:40:03 | examples-simple-prod-with-volumes | 2024-08-29 19:30:06 +0000 UTC Normal Pod simple-prod-collector-dc4588778-7xtx2 AddedInterface Add eth0 [10.130.0.32/23] from ovn-kubernetes multus logger.go:42: 19:40:03 | examples-simple-prod-with-volumes | 2024-08-29 19:30:06 +0000 UTC Normal Pod simple-prod-collector-dc4588778-7xtx2.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:b90e0d264bb0b051c2a4b9e6d499f407fed58f834ab8ebf30263ee1cf7c7ea3c" already present on machine kubelet logger.go:42: 19:40:03 | examples-simple-prod-with-volumes | 2024-08-29 19:30:06 +0000 UTC Normal Pod simple-prod-collector-dc4588778-7xtx2.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 19:40:03 | examples-simple-prod-with-volumes | 2024-08-29 19:30:06 +0000 UTC Normal Pod simple-prod-collector-dc4588778-7xtx2.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 19:40:03 | examples-simple-prod-with-volumes | 2024-08-29 19:30:06 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-dc4588778 SuccessfulCreate Created pod: simple-prod-collector-dc4588778-7xtx2 replicaset-controller logger.go:42: 19:40:03 | examples-simple-prod-with-volumes | 2024-08-29 19:30:06 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-dc4588778 to 1 deployment-controller logger.go:42: 19:40:03 | examples-simple-prod-with-volumes | 2024-08-29 19:30:06 +0000 UTC Normal Pod simple-prod-query-778f6d88d8-5mqll Binding Scheduled Successfully assigned kuttl-test-dear-weasel/simple-prod-query-778f6d88d8-5mqll to ip-10-0-51-237.us-west-2.compute.internal default-scheduler logger.go:42: 19:40:03 | examples-simple-prod-with-volumes | 2024-08-29 19:30:06 +0000 UTC Normal Pod simple-prod-query-778f6d88d8-5mqll AddedInterface Add eth0 [10.130.0.33/23] from ovn-kubernetes multus logger.go:42: 19:40:03 | examples-simple-prod-with-volumes | 2024-08-29 19:30:06 +0000 UTC Normal Pod simple-prod-query-778f6d88d8-5mqll.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:831f83e8396d61822fb2dd576e711e0f10c1f8c7bd03bf0d4bec49dfb90cacca" already present on machine kubelet logger.go:42: 19:40:03 | examples-simple-prod-with-volumes | 2024-08-29 19:30:06 +0000 UTC Normal Pod simple-prod-query-778f6d88d8-5mqll.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 19:40:03 | examples-simple-prod-with-volumes | 2024-08-29 19:30:06 +0000 UTC Normal Pod simple-prod-query-778f6d88d8-5mqll.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 19:40:03 | examples-simple-prod-with-volumes | 2024-08-29 19:30:06 +0000 UTC Normal Pod simple-prod-query-778f6d88d8-5mqll.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1613032062a68faa414cd5754a666b7b78ad0e52cf328eaaed355719423a55dd" already present on machine kubelet logger.go:42: 19:40:03 | examples-simple-prod-with-volumes | 2024-08-29 19:30:06 +0000 UTC Normal ReplicaSet.apps simple-prod-query-778f6d88d8 SuccessfulCreate Created pod: simple-prod-query-778f6d88d8-5mqll replicaset-controller logger.go:42: 19:40:03 | examples-simple-prod-with-volumes | 2024-08-29 19:30:06 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-778f6d88d8 to 1 deployment-controller logger.go:42: 19:40:03 | examples-simple-prod-with-volumes | 2024-08-29 19:30:07 +0000 UTC Normal Pod simple-prod-query-778f6d88d8-5mqll.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 19:40:03 | examples-simple-prod-with-volumes | 2024-08-29 19:30:07 +0000 UTC Normal Pod simple-prod-query-778f6d88d8-5mqll.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 19:40:03 | examples-simple-prod-with-volumes | 2024-08-29 19:30:07 +0000 UTC Normal Pod simple-prod-query-778f6d88d8-5mqll.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:5fcc9d365678f54c0f52005318dbbc32b87f743cae7e1a84568e769653f3b128" already present on machine kubelet logger.go:42: 19:40:03 | examples-simple-prod-with-volumes | 2024-08-29 19:30:07 +0000 UTC Normal Pod simple-prod-query-778f6d88d8-5mqll.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 19:40:03 | examples-simple-prod-with-volumes | 2024-08-29 19:30:07 +0000 UTC Normal Pod simple-prod-query-778f6d88d8-5mqll.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 19:40:03 | examples-simple-prod-with-volumes | 2024-08-29 19:30:08 +0000 UTC Warning Pod simple-prod-collector-dc4588778-7xtx2.spec.containers{jaeger-collector} Unhealthy Readiness probe failed: HTTP probe failed with statuscode: 503 kubelet logger.go:42: 19:40:03 | examples-simple-prod-with-volumes | 2024-08-29 19:30:08 +0000 UTC Warning Pod simple-prod-query-778f6d88d8-5mqll.spec.containers{jaeger-query} Unhealthy Readiness probe failed: HTTP probe failed with statuscode: 503 kubelet logger.go:42: 19:40:03 | examples-simple-prod-with-volumes | 2024-08-29 19:30:18 +0000 UTC Warning Pod simple-prod-collector-dc4588778-7xtx2.spec.containers{jaeger-collector} BackOff Back-off restarting failed container jaeger-collector in pod simple-prod-collector-dc4588778-7xtx2_kuttl-test-dear-weasel(f5d2d456-71f8-487f-b676-0983f76771ad) kubelet logger.go:42: 19:40:03 | examples-simple-prod-with-volumes | 2024-08-29 19:30:18 +0000 UTC Warning Pod simple-prod-query-778f6d88d8-5mqll.spec.containers{jaeger-query} BackOff Back-off restarting failed container jaeger-query in pod simple-prod-query-778f6d88d8-5mqll_kuttl-test-dear-weasel(fadb6833-b3e7-431a-ae71-23dca9448abe) kubelet logger.go:42: 19:40:03 | examples-simple-prod-with-volumes | 2024-08-29 19:30:21 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 19:40:03 | examples-simple-prod-with-volumes | 2024-08-29 19:30:21 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 19:40:03 | examples-simple-prod-with-volumes | 2024-08-29 19:30:21 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 19:40:03 | examples-simple-prod-with-volumes | Deleting namespace: kuttl-test-dear-weasel === CONT kuttl/harness/examples-business-application-injected-sidecar logger.go:42: 19:40:09 | examples-business-application-injected-sidecar | Creating namespace: kuttl-test-adapting-sawfish logger.go:42: 19:40:10 | examples-business-application-injected-sidecar/0-install | starting test step 0-install logger.go:42: 19:40:10 | examples-business-application-injected-sidecar/0-install | Deployment:kuttl-test-adapting-sawfish/myapp created logger.go:42: 19:40:10 | examples-business-application-injected-sidecar/0-install | test step completed 0-install logger.go:42: 19:40:10 | examples-business-application-injected-sidecar/1-install | starting test step 1-install logger.go:42: 19:40:10 | examples-business-application-injected-sidecar/1-install | Jaeger:kuttl-test-adapting-sawfish/simplest created logger.go:42: 19:40:14 | examples-business-application-injected-sidecar/1-install | test step completed 1-install logger.go:42: 19:40:14 | examples-business-application-injected-sidecar/2-smoke-test | starting test step 2-smoke-test logger.go:42: 19:40:14 | examples-business-application-injected-sidecar/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE simplest /dev/null] logger.go:42: 19:40:16 | examples-business-application-injected-sidecar/2-smoke-test | Warning: resource jaegers/simplest is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 19:40:23 | examples-business-application-injected-sidecar/2-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://simplest-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 19:40:23 | examples-business-application-injected-sidecar/2-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 19:40:24 | examples-business-application-injected-sidecar/2-smoke-test | job.batch/report-span created logger.go:42: 19:40:24 | examples-business-application-injected-sidecar/2-smoke-test | job.batch/check-span created logger.go:42: 19:40:31 | examples-business-application-injected-sidecar/2-smoke-test | test step completed 2-smoke-test logger.go:42: 19:40:31 | examples-business-application-injected-sidecar | examples-business-application-injected-sidecar events from ns kuttl-test-adapting-sawfish: logger.go:42: 19:40:31 | examples-business-application-injected-sidecar | 2024-08-29 19:40:10 +0000 UTC Normal Pod myapp-679f79d5f8-5sq54 Binding Scheduled Successfully assigned kuttl-test-adapting-sawfish/myapp-679f79d5f8-5sq54 to ip-10-0-51-237.us-west-2.compute.internal default-scheduler logger.go:42: 19:40:31 | examples-business-application-injected-sidecar | 2024-08-29 19:40:10 +0000 UTC Normal Pod myapp-679f79d5f8-5sq54 AddedInterface Add eth0 [10.130.0.34/23] from ovn-kubernetes multus logger.go:42: 19:40:31 | examples-business-application-injected-sidecar | 2024-08-29 19:40:10 +0000 UTC Normal Pod myapp-679f79d5f8-5sq54.spec.containers{myapp} Pulling Pulling image "jaegertracing/vertx-create-span:operator-e2e-tests" kubelet logger.go:42: 19:40:31 | examples-business-application-injected-sidecar | 2024-08-29 19:40:10 +0000 UTC Normal ReplicaSet.apps myapp-679f79d5f8 SuccessfulCreate Created pod: myapp-679f79d5f8-5sq54 replicaset-controller logger.go:42: 19:40:31 | examples-business-application-injected-sidecar | 2024-08-29 19:40:10 +0000 UTC Normal Deployment.apps myapp ScalingReplicaSet Scaled up replica set myapp-679f79d5f8 to 1 deployment-controller logger.go:42: 19:40:31 | examples-business-application-injected-sidecar | 2024-08-29 19:40:11 +0000 UTC Normal Pod myapp-5bdf4cdb49-xwg6t Binding Scheduled Successfully assigned kuttl-test-adapting-sawfish/myapp-5bdf4cdb49-xwg6t to ip-10-0-51-237.us-west-2.compute.internal default-scheduler logger.go:42: 19:40:31 | examples-business-application-injected-sidecar | 2024-08-29 19:40:11 +0000 UTC Warning Pod myapp-5bdf4cdb49-xwg6t FailedMount MountVolume.SetUp failed for volume "simplest-trusted-ca" : configmap "simplest-trusted-ca" not found kubelet logger.go:42: 19:40:31 | examples-business-application-injected-sidecar | 2024-08-29 19:40:11 +0000 UTC Warning Pod myapp-5bdf4cdb49-xwg6t FailedMount MountVolume.SetUp failed for volume "simplest-service-ca" : configmap "simplest-service-ca" not found kubelet logger.go:42: 19:40:31 | examples-business-application-injected-sidecar | 2024-08-29 19:40:11 +0000 UTC Normal ReplicaSet.apps myapp-5bdf4cdb49 SuccessfulCreate Created pod: myapp-5bdf4cdb49-xwg6t replicaset-controller logger.go:42: 19:40:31 | examples-business-application-injected-sidecar | 2024-08-29 19:40:11 +0000 UTC Normal Deployment.apps myapp ScalingReplicaSet Scaled up replica set myapp-5bdf4cdb49 to 1 deployment-controller logger.go:42: 19:40:31 | examples-business-application-injected-sidecar | 2024-08-29 19:40:12 +0000 UTC Normal Pod simplest-7d688555d7-9pxxz Binding Scheduled Successfully assigned kuttl-test-adapting-sawfish/simplest-7d688555d7-9pxxz to ip-10-0-51-237.us-west-2.compute.internal default-scheduler logger.go:42: 19:40:31 | examples-business-application-injected-sidecar | 2024-08-29 19:40:12 +0000 UTC Normal Pod simplest-7d688555d7-9pxxz AddedInterface Add eth0 [10.130.0.36/23] from ovn-kubernetes multus logger.go:42: 19:40:31 | examples-business-application-injected-sidecar | 2024-08-29 19:40:12 +0000 UTC Normal Pod simplest-7d688555d7-9pxxz.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:345d3b6c507df9f8a18948a594e7a03698085689c6965faf79b8aadc74a0ceaa" already present on machine kubelet logger.go:42: 19:40:31 | examples-business-application-injected-sidecar | 2024-08-29 19:40:12 +0000 UTC Normal Pod simplest-7d688555d7-9pxxz.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 19:40:31 | examples-business-application-injected-sidecar | 2024-08-29 19:40:12 +0000 UTC Normal Pod simplest-7d688555d7-9pxxz.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 19:40:31 | examples-business-application-injected-sidecar | 2024-08-29 19:40:12 +0000 UTC Normal Pod simplest-7d688555d7-9pxxz.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1613032062a68faa414cd5754a666b7b78ad0e52cf328eaaed355719423a55dd" already present on machine kubelet logger.go:42: 19:40:31 | examples-business-application-injected-sidecar | 2024-08-29 19:40:12 +0000 UTC Normal ReplicaSet.apps simplest-7d688555d7 SuccessfulCreate Created pod: simplest-7d688555d7-9pxxz replicaset-controller logger.go:42: 19:40:31 | examples-business-application-injected-sidecar | 2024-08-29 19:40:12 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled up replica set simplest-7d688555d7 to 1 deployment-controller logger.go:42: 19:40:31 | examples-business-application-injected-sidecar | 2024-08-29 19:40:13 +0000 UTC Normal Pod myapp-5bdf4cdb49-xwg6t AddedInterface Add eth0 [10.130.0.35/23] from ovn-kubernetes multus logger.go:42: 19:40:31 | examples-business-application-injected-sidecar | 2024-08-29 19:40:13 +0000 UTC Normal Pod myapp-5bdf4cdb49-xwg6t.spec.containers{myapp} Pulling Pulling image "jaegertracing/vertx-create-span:operator-e2e-tests" kubelet logger.go:42: 19:40:31 | examples-business-application-injected-sidecar | 2024-08-29 19:40:13 +0000 UTC Normal Pod simplest-7d688555d7-9pxxz.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 19:40:31 | examples-business-application-injected-sidecar | 2024-08-29 19:40:13 +0000 UTC Normal Pod simplest-7d688555d7-9pxxz.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 19:40:31 | examples-business-application-injected-sidecar | 2024-08-29 19:40:16 +0000 UTC Normal Pod myapp-5bdf4cdb49-xwg6t.spec.containers{myapp} Pulled Successfully pulled image "jaegertracing/vertx-create-span:operator-e2e-tests" in 3.192s (3.192s including waiting) kubelet logger.go:42: 19:40:31 | examples-business-application-injected-sidecar | 2024-08-29 19:40:16 +0000 UTC Normal Pod myapp-5bdf4cdb49-xwg6t.spec.containers{myapp} Created Created container myapp kubelet logger.go:42: 19:40:31 | examples-business-application-injected-sidecar | 2024-08-29 19:40:16 +0000 UTC Normal Pod myapp-5bdf4cdb49-xwg6t.spec.containers{myapp} Started Started container myapp kubelet logger.go:42: 19:40:31 | examples-business-application-injected-sidecar | 2024-08-29 19:40:16 +0000 UTC Normal Pod myapp-5bdf4cdb49-xwg6t.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:5fcc9d365678f54c0f52005318dbbc32b87f743cae7e1a84568e769653f3b128" already present on machine kubelet logger.go:42: 19:40:31 | examples-business-application-injected-sidecar | 2024-08-29 19:40:16 +0000 UTC Normal Pod myapp-5bdf4cdb49-xwg6t.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 19:40:31 | examples-business-application-injected-sidecar | 2024-08-29 19:40:16 +0000 UTC Normal Pod myapp-5bdf4cdb49-xwg6t.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 19:40:31 | examples-business-application-injected-sidecar | 2024-08-29 19:40:16 +0000 UTC Normal Pod myapp-679f79d5f8-5sq54.spec.containers{myapp} Pulled Successfully pulled image "jaegertracing/vertx-create-span:operator-e2e-tests" in 5.258s (5.258s including waiting) kubelet logger.go:42: 19:40:31 | examples-business-application-injected-sidecar | 2024-08-29 19:40:16 +0000 UTC Normal Pod myapp-679f79d5f8-5sq54.spec.containers{myapp} Created Created container myapp kubelet logger.go:42: 19:40:31 | examples-business-application-injected-sidecar | 2024-08-29 19:40:16 +0000 UTC Normal Pod myapp-679f79d5f8-5sq54.spec.containers{myapp} Started Started container myapp kubelet logger.go:42: 19:40:31 | examples-business-application-injected-sidecar | 2024-08-29 19:40:16 +0000 UTC Normal Pod simplest-7d688555d7-9pxxz.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 19:40:31 | examples-business-application-injected-sidecar | 2024-08-29 19:40:16 +0000 UTC Normal Pod simplest-7d688555d7-9pxxz.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 19:40:31 | examples-business-application-injected-sidecar | 2024-08-29 19:40:16 +0000 UTC Normal ReplicaSet.apps simplest-7d688555d7 SuccessfulDelete Deleted pod: simplest-7d688555d7-9pxxz replicaset-controller logger.go:42: 19:40:31 | examples-business-application-injected-sidecar | 2024-08-29 19:40:16 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled down replica set simplest-7d688555d7 to 0 from 1 deployment-controller logger.go:42: 19:40:31 | examples-business-application-injected-sidecar | 2024-08-29 19:40:18 +0000 UTC Normal Pod myapp-679f79d5f8-5sq54.spec.containers{myapp} Killing Stopping container myapp kubelet logger.go:42: 19:40:31 | examples-business-application-injected-sidecar | 2024-08-29 19:40:18 +0000 UTC Normal ReplicaSet.apps myapp-679f79d5f8 SuccessfulDelete Deleted pod: myapp-679f79d5f8-5sq54 replicaset-controller logger.go:42: 19:40:31 | examples-business-application-injected-sidecar | 2024-08-29 19:40:18 +0000 UTC Normal Deployment.apps myapp ScalingReplicaSet Scaled down replica set myapp-679f79d5f8 to 0 from 1 deployment-controller logger.go:42: 19:40:31 | examples-business-application-injected-sidecar | 2024-08-29 19:40:18 +0000 UTC Normal Pod simplest-8567dd78cb-27ptb Binding Scheduled Successfully assigned kuttl-test-adapting-sawfish/simplest-8567dd78cb-27ptb to ip-10-0-51-237.us-west-2.compute.internal default-scheduler logger.go:42: 19:40:31 | examples-business-application-injected-sidecar | 2024-08-29 19:40:18 +0000 UTC Normal Pod simplest-8567dd78cb-27ptb AddedInterface Add eth0 [10.130.0.37/23] from ovn-kubernetes multus logger.go:42: 19:40:31 | examples-business-application-injected-sidecar | 2024-08-29 19:40:18 +0000 UTC Normal Pod simplest-8567dd78cb-27ptb.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:345d3b6c507df9f8a18948a594e7a03698085689c6965faf79b8aadc74a0ceaa" already present on machine kubelet logger.go:42: 19:40:31 | examples-business-application-injected-sidecar | 2024-08-29 19:40:18 +0000 UTC Normal Pod simplest-8567dd78cb-27ptb.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 19:40:31 | examples-business-application-injected-sidecar | 2024-08-29 19:40:18 +0000 UTC Normal Pod simplest-8567dd78cb-27ptb.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 19:40:31 | examples-business-application-injected-sidecar | 2024-08-29 19:40:18 +0000 UTC Normal Pod simplest-8567dd78cb-27ptb.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1613032062a68faa414cd5754a666b7b78ad0e52cf328eaaed355719423a55dd" already present on machine kubelet logger.go:42: 19:40:31 | examples-business-application-injected-sidecar | 2024-08-29 19:40:18 +0000 UTC Normal Pod simplest-8567dd78cb-27ptb.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 19:40:31 | examples-business-application-injected-sidecar | 2024-08-29 19:40:18 +0000 UTC Normal Pod simplest-8567dd78cb-27ptb.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 19:40:31 | examples-business-application-injected-sidecar | 2024-08-29 19:40:18 +0000 UTC Normal ReplicaSet.apps simplest-8567dd78cb SuccessfulCreate Created pod: simplest-8567dd78cb-27ptb replicaset-controller logger.go:42: 19:40:31 | examples-business-application-injected-sidecar | 2024-08-29 19:40:18 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled up replica set simplest-8567dd78cb to 1 deployment-controller logger.go:42: 19:40:31 | examples-business-application-injected-sidecar | 2024-08-29 19:40:23 +0000 UTC Warning Pod myapp-5bdf4cdb49-xwg6t.spec.containers{myapp} Unhealthy Liveness probe failed: Get "http://10.130.0.35:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 19:40:31 | examples-business-application-injected-sidecar | 2024-08-29 19:40:24 +0000 UTC Normal Pod check-span-w9lcm Binding Scheduled Successfully assigned kuttl-test-adapting-sawfish/check-span-w9lcm to ip-10-0-57-246.us-west-2.compute.internal default-scheduler logger.go:42: 19:40:31 | examples-business-application-injected-sidecar | 2024-08-29 19:40:24 +0000 UTC Normal Pod check-span-w9lcm AddedInterface Add eth0 [10.129.0.32/23] from ovn-kubernetes multus logger.go:42: 19:40:31 | examples-business-application-injected-sidecar | 2024-08-29 19:40:24 +0000 UTC Normal Pod check-span-w9lcm.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 19:40:31 | examples-business-application-injected-sidecar | 2024-08-29 19:40:24 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-w9lcm job-controller logger.go:42: 19:40:31 | examples-business-application-injected-sidecar | 2024-08-29 19:40:24 +0000 UTC Normal Pod report-span-6flcr Binding Scheduled Successfully assigned kuttl-test-adapting-sawfish/report-span-6flcr to ip-10-0-51-237.us-west-2.compute.internal default-scheduler logger.go:42: 19:40:31 | examples-business-application-injected-sidecar | 2024-08-29 19:40:24 +0000 UTC Normal Pod report-span-6flcr AddedInterface Add eth0 [10.130.0.38/23] from ovn-kubernetes multus logger.go:42: 19:40:31 | examples-business-application-injected-sidecar | 2024-08-29 19:40:24 +0000 UTC Normal Pod report-span-6flcr.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 19:40:31 | examples-business-application-injected-sidecar | 2024-08-29 19:40:24 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-6flcr job-controller logger.go:42: 19:40:31 | examples-business-application-injected-sidecar | 2024-08-29 19:40:25 +0000 UTC Normal Pod report-span-6flcr.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 857ms (857ms including waiting) kubelet logger.go:42: 19:40:31 | examples-business-application-injected-sidecar | 2024-08-29 19:40:25 +0000 UTC Normal Pod report-span-6flcr.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 19:40:31 | examples-business-application-injected-sidecar | 2024-08-29 19:40:25 +0000 UTC Normal Pod report-span-6flcr.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 19:40:31 | examples-business-application-injected-sidecar | 2024-08-29 19:40:27 +0000 UTC Normal Pod check-span-w9lcm.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 3.063s (3.063s including waiting) kubelet logger.go:42: 19:40:31 | examples-business-application-injected-sidecar | 2024-08-29 19:40:28 +0000 UTC Normal Pod check-span-w9lcm.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 19:40:31 | examples-business-application-injected-sidecar | 2024-08-29 19:40:28 +0000 UTC Normal Pod check-span-w9lcm.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 19:40:31 | examples-business-application-injected-sidecar | 2024-08-29 19:40:30 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 19:40:31 | examples-business-application-injected-sidecar | Deleting namespace: kuttl-test-adapting-sawfish === CONT kuttl/harness/examples-service-types logger.go:42: 19:40:43 | examples-service-types | Creating namespace: kuttl-test-golden-satyr logger.go:42: 19:40:43 | examples-service-types/0-install | starting test step 0-install logger.go:42: 19:40:44 | examples-service-types/0-install | Jaeger:kuttl-test-golden-satyr/service-types created logger.go:42: 19:40:47 | examples-service-types/0-install | test step completed 0-install logger.go:42: 19:40:47 | examples-service-types/1-smoke-test | starting test step 1-smoke-test logger.go:42: 19:40:47 | examples-service-types/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE service-types /dev/null] logger.go:42: 19:40:49 | examples-service-types/1-smoke-test | Warning: resource jaegers/service-types is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 19:40:56 | examples-service-types/1-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://service-types-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://service-types-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 19:40:57 | examples-service-types/1-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 19:40:57 | examples-service-types/1-smoke-test | job.batch/report-span created logger.go:42: 19:40:57 | examples-service-types/1-smoke-test | job.batch/check-span created logger.go:42: 19:41:11 | examples-service-types/1-smoke-test | test step completed 1-smoke-test logger.go:42: 19:41:11 | examples-service-types/2- | starting test step 2- logger.go:42: 19:41:11 | examples-service-types/2- | test step completed 2- logger.go:42: 19:41:11 | examples-service-types | examples-service-types events from ns kuttl-test-golden-satyr: logger.go:42: 19:41:11 | examples-service-types | 2024-08-29 19:40:44 +0000 UTC Normal Pod service-types-756667574d-j7mbj Binding Scheduled Successfully assigned kuttl-test-golden-satyr/service-types-756667574d-j7mbj to ip-10-0-51-237.us-west-2.compute.internal default-scheduler logger.go:42: 19:41:11 | examples-service-types | 2024-08-29 19:40:44 +0000 UTC Normal Pod service-types-756667574d-j7mbj AddedInterface Add eth0 [10.130.0.39/23] from ovn-kubernetes multus logger.go:42: 19:41:11 | examples-service-types | 2024-08-29 19:40:44 +0000 UTC Normal ReplicaSet.apps service-types-756667574d SuccessfulCreate Created pod: service-types-756667574d-j7mbj replicaset-controller logger.go:42: 19:41:11 | examples-service-types | 2024-08-29 19:40:44 +0000 UTC Normal Service service-types-collector EnsuringLoadBalancer Ensuring load balancer service-controller logger.go:42: 19:41:11 | examples-service-types | 2024-08-29 19:40:44 +0000 UTC Normal Deployment.apps service-types ScalingReplicaSet Scaled up replica set service-types-756667574d to 1 deployment-controller logger.go:42: 19:41:11 | examples-service-types | 2024-08-29 19:40:45 +0000 UTC Normal Pod service-types-756667574d-j7mbj.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:345d3b6c507df9f8a18948a594e7a03698085689c6965faf79b8aadc74a0ceaa" already present on machine kubelet logger.go:42: 19:41:11 | examples-service-types | 2024-08-29 19:40:45 +0000 UTC Normal Pod service-types-756667574d-j7mbj.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 19:41:11 | examples-service-types | 2024-08-29 19:40:45 +0000 UTC Normal Pod service-types-756667574d-j7mbj.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 19:41:11 | examples-service-types | 2024-08-29 19:40:45 +0000 UTC Normal Pod service-types-756667574d-j7mbj.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1613032062a68faa414cd5754a666b7b78ad0e52cf328eaaed355719423a55dd" already present on machine kubelet logger.go:42: 19:41:11 | examples-service-types | 2024-08-29 19:40:45 +0000 UTC Normal Pod service-types-756667574d-j7mbj.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 19:41:11 | examples-service-types | 2024-08-29 19:40:45 +0000 UTC Normal Pod service-types-756667574d-j7mbj.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 19:41:11 | examples-service-types | 2024-08-29 19:40:47 +0000 UTC Normal Service service-types-collector EnsuredLoadBalancer Ensured load balancer service-controller logger.go:42: 19:41:11 | examples-service-types | 2024-08-29 19:40:47 +0000 UTC Normal Service service-types-query EnsuringLoadBalancer Ensuring load balancer service-controller logger.go:42: 19:41:11 | examples-service-types | 2024-08-29 19:40:50 +0000 UTC Normal Pod service-types-756667574d-j7mbj.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 19:41:11 | examples-service-types | 2024-08-29 19:40:50 +0000 UTC Normal Pod service-types-756667574d-j7mbj.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 19:41:11 | examples-service-types | 2024-08-29 19:40:50 +0000 UTC Normal ReplicaSet.apps service-types-756667574d SuccessfulDelete Deleted pod: service-types-756667574d-j7mbj replicaset-controller logger.go:42: 19:41:11 | examples-service-types | 2024-08-29 19:40:50 +0000 UTC Normal Service service-types-query EnsuredLoadBalancer Ensured load balancer service-controller logger.go:42: 19:41:11 | examples-service-types | 2024-08-29 19:40:50 +0000 UTC Normal Deployment.apps service-types ScalingReplicaSet Scaled down replica set service-types-756667574d to 0 from 1 deployment-controller logger.go:42: 19:41:11 | examples-service-types | 2024-08-29 19:40:51 +0000 UTC Normal Pod service-types-947cbc988-h55jp Binding Scheduled Successfully assigned kuttl-test-golden-satyr/service-types-947cbc988-h55jp to ip-10-0-51-237.us-west-2.compute.internal default-scheduler logger.go:42: 19:41:11 | examples-service-types | 2024-08-29 19:40:51 +0000 UTC Normal Pod service-types-947cbc988-h55jp AddedInterface Add eth0 [10.130.0.40/23] from ovn-kubernetes multus logger.go:42: 19:41:11 | examples-service-types | 2024-08-29 19:40:51 +0000 UTC Normal Pod service-types-947cbc988-h55jp.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:345d3b6c507df9f8a18948a594e7a03698085689c6965faf79b8aadc74a0ceaa" already present on machine kubelet logger.go:42: 19:41:11 | examples-service-types | 2024-08-29 19:40:51 +0000 UTC Normal Pod service-types-947cbc988-h55jp.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 19:41:11 | examples-service-types | 2024-08-29 19:40:51 +0000 UTC Normal Pod service-types-947cbc988-h55jp.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 19:41:11 | examples-service-types | 2024-08-29 19:40:51 +0000 UTC Normal Pod service-types-947cbc988-h55jp.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1613032062a68faa414cd5754a666b7b78ad0e52cf328eaaed355719423a55dd" already present on machine kubelet logger.go:42: 19:41:11 | examples-service-types | 2024-08-29 19:40:51 +0000 UTC Normal Pod service-types-947cbc988-h55jp.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 19:41:11 | examples-service-types | 2024-08-29 19:40:51 +0000 UTC Normal Pod service-types-947cbc988-h55jp.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 19:41:11 | examples-service-types | 2024-08-29 19:40:51 +0000 UTC Normal ReplicaSet.apps service-types-947cbc988 SuccessfulCreate Created pod: service-types-947cbc988-h55jp replicaset-controller logger.go:42: 19:41:11 | examples-service-types | 2024-08-29 19:40:51 +0000 UTC Normal Deployment.apps service-types ScalingReplicaSet Scaled up replica set service-types-947cbc988 to 1 deployment-controller logger.go:42: 19:41:11 | examples-service-types | 2024-08-29 19:40:57 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-r4zkq job-controller logger.go:42: 19:41:11 | examples-service-types | 2024-08-29 19:40:57 +0000 UTC Normal Pod report-span-thndc Binding Scheduled Successfully assigned kuttl-test-golden-satyr/report-span-thndc to ip-10-0-51-237.us-west-2.compute.internal default-scheduler logger.go:42: 19:41:11 | examples-service-types | 2024-08-29 19:40:57 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-thndc job-controller logger.go:42: 19:41:11 | examples-service-types | 2024-08-29 19:40:58 +0000 UTC Normal Pod check-span-r4zkq Binding Scheduled Successfully assigned kuttl-test-golden-satyr/check-span-r4zkq to ip-10-0-51-237.us-west-2.compute.internal default-scheduler logger.go:42: 19:41:11 | examples-service-types | 2024-08-29 19:40:58 +0000 UTC Normal Pod check-span-r4zkq AddedInterface Add eth0 [10.130.0.42/23] from ovn-kubernetes multus logger.go:42: 19:41:11 | examples-service-types | 2024-08-29 19:40:58 +0000 UTC Normal Pod check-span-r4zkq.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 19:41:11 | examples-service-types | 2024-08-29 19:40:58 +0000 UTC Normal Pod report-span-thndc AddedInterface Add eth0 [10.130.0.41/23] from ovn-kubernetes multus logger.go:42: 19:41:11 | examples-service-types | 2024-08-29 19:40:58 +0000 UTC Normal Pod report-span-thndc.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 19:41:11 | examples-service-types | 2024-08-29 19:40:59 +0000 UTC Normal Pod check-span-r4zkq.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 836ms (836ms including waiting) kubelet logger.go:42: 19:41:11 | examples-service-types | 2024-08-29 19:40:59 +0000 UTC Normal Pod check-span-r4zkq.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 19:41:11 | examples-service-types | 2024-08-29 19:40:59 +0000 UTC Normal Pod check-span-r4zkq.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 19:41:11 | examples-service-types | 2024-08-29 19:40:59 +0000 UTC Normal Pod report-span-thndc.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 807ms (807ms including waiting) kubelet logger.go:42: 19:41:11 | examples-service-types | 2024-08-29 19:40:59 +0000 UTC Normal Pod report-span-thndc.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 19:41:11 | examples-service-types | 2024-08-29 19:40:59 +0000 UTC Normal Pod report-span-thndc.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 19:41:11 | examples-service-types | 2024-08-29 19:41:10 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 19:41:11 | examples-service-types | Deleting namespace: kuttl-test-golden-satyr === CONT kuttl/harness/examples-openshift-with-htpasswd logger.go:42: 19:41:35 | examples-openshift-with-htpasswd | Ignoring 00-install.yaml.template as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 19:41:35 | examples-openshift-with-htpasswd | Ignoring ensure-ingress-host.sh as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 19:41:35 | examples-openshift-with-htpasswd | Creating namespace: kuttl-test-worthy-kangaroo logger.go:42: 19:41:35 | examples-openshift-with-htpasswd/0-install | starting test step 0-install logger.go:42: 19:41:35 | examples-openshift-with-htpasswd/0-install | Secret:kuttl-test-worthy-kangaroo/htpasswd created logger.go:42: 19:41:35 | examples-openshift-with-htpasswd/0-install | test step completed 0-install logger.go:42: 19:41:35 | examples-openshift-with-htpasswd/1-install | starting test step 1-install logger.go:42: 19:41:35 | examples-openshift-with-htpasswd/1-install | Jaeger:kuttl-test-worthy-kangaroo/with-htpasswd created logger.go:42: 19:41:38 | examples-openshift-with-htpasswd/1-install | test step completed 1-install logger.go:42: 19:41:38 | examples-openshift-with-htpasswd/2-check-unsecured | starting test step 2-check-unsecured logger.go:42: 19:41:38 | examples-openshift-with-htpasswd/2-check-unsecured | running command: [./ensure-ingress-host.sh] logger.go:42: 19:41:38 | examples-openshift-with-htpasswd/2-check-unsecured | Checking the Ingress host value was populated logger.go:42: 19:41:38 | examples-openshift-with-htpasswd/2-check-unsecured | Try number 0 logger.go:42: 19:41:38 | examples-openshift-with-htpasswd/2-check-unsecured | Hostname is with-htpasswd-kuttl-test-worthy-kangaroo.apps.rosa.ci-rosa-h-yd3i.24xa.s3.devshift.org logger.go:42: 19:41:38 | examples-openshift-with-htpasswd/2-check-unsecured | running command: [sh -c INSECURE=true ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 403 true $NAMESPACE with-htpasswd] logger.go:42: 19:41:38 | examples-openshift-with-htpasswd/2-check-unsecured | Checking an expected HTTP response logger.go:42: 19:41:38 | examples-openshift-with-htpasswd/2-check-unsecured | Running in OpenShift logger.go:42: 19:41:38 | examples-openshift-with-htpasswd/2-check-unsecured | Not using any secret logger.go:42: 19:41:38 | examples-openshift-with-htpasswd/2-check-unsecured | Try number 1/30 the https://with-htpasswd-kuttl-test-worthy-kangaroo.apps.rosa.ci-rosa-h-yd3i.24xa.s3.devshift.org/search logger.go:42: 19:41:38 | examples-openshift-with-htpasswd/2-check-unsecured | HTTP response is 503. 403 expected. Waiting 10 s logger.go:42: 19:41:48 | examples-openshift-with-htpasswd/2-check-unsecured | Try number 2/30 the https://with-htpasswd-kuttl-test-worthy-kangaroo.apps.rosa.ci-rosa-h-yd3i.24xa.s3.devshift.org/search logger.go:42: 19:41:49 | examples-openshift-with-htpasswd/2-check-unsecured | curl response asserted properly logger.go:42: 19:41:49 | examples-openshift-with-htpasswd/2-check-unsecured | test step completed 2-check-unsecured logger.go:42: 19:41:49 | examples-openshift-with-htpasswd/3-check-unauthorized | starting test step 3-check-unauthorized logger.go:42: 19:41:49 | examples-openshift-with-htpasswd/3-check-unauthorized | running command: [./ensure-ingress-host.sh] logger.go:42: 19:41:49 | examples-openshift-with-htpasswd/3-check-unauthorized | Checking the Ingress host value was populated logger.go:42: 19:41:49 | examples-openshift-with-htpasswd/3-check-unauthorized | Try number 0 logger.go:42: 19:41:49 | examples-openshift-with-htpasswd/3-check-unauthorized | Hostname is with-htpasswd-kuttl-test-worthy-kangaroo.apps.rosa.ci-rosa-h-yd3i.24xa.s3.devshift.org logger.go:42: 19:41:49 | examples-openshift-with-htpasswd/3-check-unauthorized | running command: [sh -c JAEGER_USERNAME=wronguser JAEGER_PASSWORD=wrongpassword ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 403 true $NAMESPACE with-htpasswd] logger.go:42: 19:41:49 | examples-openshift-with-htpasswd/3-check-unauthorized | Checking an expected HTTP response logger.go:42: 19:41:49 | examples-openshift-with-htpasswd/3-check-unauthorized | Running in OpenShift logger.go:42: 19:41:49 | examples-openshift-with-htpasswd/3-check-unauthorized | Using Jaeger basic authentication logger.go:42: 19:41:49 | examples-openshift-with-htpasswd/3-check-unauthorized | Try number 1/30 the https://with-htpasswd-kuttl-test-worthy-kangaroo.apps.rosa.ci-rosa-h-yd3i.24xa.s3.devshift.org/search logger.go:42: 19:41:50 | examples-openshift-with-htpasswd/3-check-unauthorized | curl response asserted properly logger.go:42: 19:41:50 | examples-openshift-with-htpasswd/3-check-unauthorized | test step completed 3-check-unauthorized logger.go:42: 19:41:50 | examples-openshift-with-htpasswd/4-check-authorized | starting test step 4-check-authorized logger.go:42: 19:41:50 | examples-openshift-with-htpasswd/4-check-authorized | running command: [./ensure-ingress-host.sh] logger.go:42: 19:41:50 | examples-openshift-with-htpasswd/4-check-authorized | Checking the Ingress host value was populated logger.go:42: 19:41:50 | examples-openshift-with-htpasswd/4-check-authorized | Try number 0 logger.go:42: 19:41:50 | examples-openshift-with-htpasswd/4-check-authorized | Hostname is with-htpasswd-kuttl-test-worthy-kangaroo.apps.rosa.ci-rosa-h-yd3i.24xa.s3.devshift.org logger.go:42: 19:41:50 | examples-openshift-with-htpasswd/4-check-authorized | running command: [sh -c JAEGER_USERNAME=awesomeuser JAEGER_PASSWORD=awesomepassword ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 200 true $NAMESPACE with-htpasswd] logger.go:42: 19:41:50 | examples-openshift-with-htpasswd/4-check-authorized | Checking an expected HTTP response logger.go:42: 19:41:50 | examples-openshift-with-htpasswd/4-check-authorized | Running in OpenShift logger.go:42: 19:41:50 | examples-openshift-with-htpasswd/4-check-authorized | Using Jaeger basic authentication logger.go:42: 19:41:50 | examples-openshift-with-htpasswd/4-check-authorized | Try number 1/30 the https://with-htpasswd-kuttl-test-worthy-kangaroo.apps.rosa.ci-rosa-h-yd3i.24xa.s3.devshift.org/search logger.go:42: 19:41:51 | examples-openshift-with-htpasswd/4-check-authorized | curl response asserted properly logger.go:42: 19:41:51 | examples-openshift-with-htpasswd/4-check-authorized | test step completed 4-check-authorized logger.go:42: 19:41:51 | examples-openshift-with-htpasswd | examples-openshift-with-htpasswd events from ns kuttl-test-worthy-kangaroo: logger.go:42: 19:41:51 | examples-openshift-with-htpasswd | 2024-08-29 19:41:36 +0000 UTC Normal Pod with-htpasswd-cfbdcdb76-4dg5s Binding Scheduled Successfully assigned kuttl-test-worthy-kangaroo/with-htpasswd-cfbdcdb76-4dg5s to ip-10-0-51-237.us-west-2.compute.internal default-scheduler logger.go:42: 19:41:51 | examples-openshift-with-htpasswd | 2024-08-29 19:41:36 +0000 UTC Normal Pod with-htpasswd-cfbdcdb76-4dg5s AddedInterface Add eth0 [10.130.0.43/23] from ovn-kubernetes multus logger.go:42: 19:41:51 | examples-openshift-with-htpasswd | 2024-08-29 19:41:36 +0000 UTC Normal Pod with-htpasswd-cfbdcdb76-4dg5s.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:345d3b6c507df9f8a18948a594e7a03698085689c6965faf79b8aadc74a0ceaa" already present on machine kubelet logger.go:42: 19:41:51 | examples-openshift-with-htpasswd | 2024-08-29 19:41:36 +0000 UTC Normal Pod with-htpasswd-cfbdcdb76-4dg5s.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 19:41:51 | examples-openshift-with-htpasswd | 2024-08-29 19:41:36 +0000 UTC Normal Pod with-htpasswd-cfbdcdb76-4dg5s.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 19:41:51 | examples-openshift-with-htpasswd | 2024-08-29 19:41:36 +0000 UTC Normal Pod with-htpasswd-cfbdcdb76-4dg5s.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1613032062a68faa414cd5754a666b7b78ad0e52cf328eaaed355719423a55dd" already present on machine kubelet logger.go:42: 19:41:51 | examples-openshift-with-htpasswd | 2024-08-29 19:41:36 +0000 UTC Normal ReplicaSet.apps with-htpasswd-cfbdcdb76 SuccessfulCreate Created pod: with-htpasswd-cfbdcdb76-4dg5s replicaset-controller logger.go:42: 19:41:51 | examples-openshift-with-htpasswd | 2024-08-29 19:41:36 +0000 UTC Normal Deployment.apps with-htpasswd ScalingReplicaSet Scaled up replica set with-htpasswd-cfbdcdb76 to 1 deployment-controller logger.go:42: 19:41:51 | examples-openshift-with-htpasswd | 2024-08-29 19:41:37 +0000 UTC Normal Pod with-htpasswd-cfbdcdb76-4dg5s.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 19:41:51 | examples-openshift-with-htpasswd | 2024-08-29 19:41:37 +0000 UTC Normal Pod with-htpasswd-cfbdcdb76-4dg5s.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 19:41:51 | examples-openshift-with-htpasswd | Deleting namespace: kuttl-test-worthy-kangaroo === CONT kuttl/harness/examples-all-in-one-with-options logger.go:42: 19:41:57 | examples-all-in-one-with-options | Creating namespace: kuttl-test-superb-seasnail logger.go:42: 19:41:57 | examples-all-in-one-with-options/0-install | starting test step 0-install logger.go:42: 19:41:58 | examples-all-in-one-with-options/0-install | Jaeger:kuttl-test-superb-seasnail/my-jaeger created logger.go:42: 19:42:00 | examples-all-in-one-with-options/0-install | test step completed 0-install logger.go:42: 19:42:00 | examples-all-in-one-with-options/1-smoke-test | starting test step 1-smoke-test logger.go:42: 19:42:00 | examples-all-in-one-with-options/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 19:42:02 | examples-all-in-one-with-options/1-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 19:42:09 | examples-all-in-one-with-options/1-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443/jaeger MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 19:42:10 | examples-all-in-one-with-options/1-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 19:42:10 | examples-all-in-one-with-options/1-smoke-test | job.batch/report-span created logger.go:42: 19:42:10 | examples-all-in-one-with-options/1-smoke-test | job.batch/check-span created logger.go:42: 19:42:22 | examples-all-in-one-with-options/1-smoke-test | test step completed 1-smoke-test logger.go:42: 19:42:22 | examples-all-in-one-with-options | examples-all-in-one-with-options events from ns kuttl-test-superb-seasnail: logger.go:42: 19:42:22 | examples-all-in-one-with-options | 2024-08-29 19:41:58 +0000 UTC Normal Pod my-jaeger-6789c75bc8-zw4pm Binding Scheduled Successfully assigned kuttl-test-superb-seasnail/my-jaeger-6789c75bc8-zw4pm to ip-10-0-51-237.us-west-2.compute.internal default-scheduler logger.go:42: 19:42:22 | examples-all-in-one-with-options | 2024-08-29 19:41:58 +0000 UTC Normal Pod my-jaeger-6789c75bc8-zw4pm AddedInterface Add eth0 [10.130.0.44/23] from ovn-kubernetes multus logger.go:42: 19:42:22 | examples-all-in-one-with-options | 2024-08-29 19:41:58 +0000 UTC Normal Pod my-jaeger-6789c75bc8-zw4pm.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:345d3b6c507df9f8a18948a594e7a03698085689c6965faf79b8aadc74a0ceaa" already present on machine kubelet logger.go:42: 19:42:22 | examples-all-in-one-with-options | 2024-08-29 19:41:58 +0000 UTC Normal Pod my-jaeger-6789c75bc8-zw4pm.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 19:42:22 | examples-all-in-one-with-options | 2024-08-29 19:41:58 +0000 UTC Normal Pod my-jaeger-6789c75bc8-zw4pm.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 19:42:22 | examples-all-in-one-with-options | 2024-08-29 19:41:58 +0000 UTC Normal Pod my-jaeger-6789c75bc8-zw4pm.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1613032062a68faa414cd5754a666b7b78ad0e52cf328eaaed355719423a55dd" already present on machine kubelet logger.go:42: 19:42:22 | examples-all-in-one-with-options | 2024-08-29 19:41:58 +0000 UTC Normal ReplicaSet.apps my-jaeger-6789c75bc8 SuccessfulCreate Created pod: my-jaeger-6789c75bc8-zw4pm replicaset-controller logger.go:42: 19:42:22 | examples-all-in-one-with-options | 2024-08-29 19:41:58 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-6789c75bc8 to 1 deployment-controller logger.go:42: 19:42:22 | examples-all-in-one-with-options | 2024-08-29 19:41:59 +0000 UTC Normal Pod my-jaeger-6789c75bc8-zw4pm.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 19:42:22 | examples-all-in-one-with-options | 2024-08-29 19:41:59 +0000 UTC Normal Pod my-jaeger-6789c75bc8-zw4pm.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 19:42:22 | examples-all-in-one-with-options | 2024-08-29 19:42:02 +0000 UTC Normal ReplicaSet.apps my-jaeger-6789c75bc8 SuccessfulDelete Deleted pod: my-jaeger-6789c75bc8-zw4pm replicaset-controller logger.go:42: 19:42:22 | examples-all-in-one-with-options | 2024-08-29 19:42:02 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled down replica set my-jaeger-6789c75bc8 to 0 from 1 deployment-controller logger.go:42: 19:42:22 | examples-all-in-one-with-options | 2024-08-29 19:42:03 +0000 UTC Normal Pod my-jaeger-6789c75bc8-zw4pm.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 19:42:22 | examples-all-in-one-with-options | 2024-08-29 19:42:03 +0000 UTC Normal Pod my-jaeger-6789c75bc8-zw4pm.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 19:42:22 | examples-all-in-one-with-options | 2024-08-29 19:42:04 +0000 UTC Normal Pod my-jaeger-64d65d4c99-8g7h7 Binding Scheduled Successfully assigned kuttl-test-superb-seasnail/my-jaeger-64d65d4c99-8g7h7 to ip-10-0-51-237.us-west-2.compute.internal default-scheduler logger.go:42: 19:42:22 | examples-all-in-one-with-options | 2024-08-29 19:42:04 +0000 UTC Normal Pod my-jaeger-64d65d4c99-8g7h7 AddedInterface Add eth0 [10.130.0.45/23] from ovn-kubernetes multus logger.go:42: 19:42:22 | examples-all-in-one-with-options | 2024-08-29 19:42:04 +0000 UTC Normal Pod my-jaeger-64d65d4c99-8g7h7.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:345d3b6c507df9f8a18948a594e7a03698085689c6965faf79b8aadc74a0ceaa" already present on machine kubelet logger.go:42: 19:42:22 | examples-all-in-one-with-options | 2024-08-29 19:42:04 +0000 UTC Normal Pod my-jaeger-64d65d4c99-8g7h7.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 19:42:22 | examples-all-in-one-with-options | 2024-08-29 19:42:04 +0000 UTC Normal Pod my-jaeger-64d65d4c99-8g7h7.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 19:42:22 | examples-all-in-one-with-options | 2024-08-29 19:42:04 +0000 UTC Normal Pod my-jaeger-64d65d4c99-8g7h7.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1613032062a68faa414cd5754a666b7b78ad0e52cf328eaaed355719423a55dd" already present on machine kubelet logger.go:42: 19:42:22 | examples-all-in-one-with-options | 2024-08-29 19:42:04 +0000 UTC Normal Pod my-jaeger-64d65d4c99-8g7h7.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 19:42:22 | examples-all-in-one-with-options | 2024-08-29 19:42:04 +0000 UTC Normal Pod my-jaeger-64d65d4c99-8g7h7.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 19:42:22 | examples-all-in-one-with-options | 2024-08-29 19:42:04 +0000 UTC Normal ReplicaSet.apps my-jaeger-64d65d4c99 SuccessfulCreate Created pod: my-jaeger-64d65d4c99-8g7h7 replicaset-controller logger.go:42: 19:42:22 | examples-all-in-one-with-options | 2024-08-29 19:42:04 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-64d65d4c99 to 1 deployment-controller logger.go:42: 19:42:22 | examples-all-in-one-with-options | 2024-08-29 19:42:10 +0000 UTC Normal Pod check-span-lqgql Binding Scheduled Successfully assigned kuttl-test-superb-seasnail/check-span-lqgql to ip-10-0-51-237.us-west-2.compute.internal default-scheduler logger.go:42: 19:42:22 | examples-all-in-one-with-options | 2024-08-29 19:42:10 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-lqgql job-controller logger.go:42: 19:42:22 | examples-all-in-one-with-options | 2024-08-29 19:42:10 +0000 UTC Normal Pod report-span-qq6kq Binding Scheduled Successfully assigned kuttl-test-superb-seasnail/report-span-qq6kq to ip-10-0-51-237.us-west-2.compute.internal default-scheduler logger.go:42: 19:42:22 | examples-all-in-one-with-options | 2024-08-29 19:42:10 +0000 UTC Normal Pod report-span-qq6kq AddedInterface Add eth0 [10.130.0.46/23] from ovn-kubernetes multus logger.go:42: 19:42:22 | examples-all-in-one-with-options | 2024-08-29 19:42:10 +0000 UTC Normal Pod report-span-qq6kq.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 19:42:22 | examples-all-in-one-with-options | 2024-08-29 19:42:10 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-qq6kq job-controller logger.go:42: 19:42:22 | examples-all-in-one-with-options | 2024-08-29 19:42:11 +0000 UTC Normal Pod check-span-lqgql AddedInterface Add eth0 [10.130.0.47/23] from ovn-kubernetes multus logger.go:42: 19:42:22 | examples-all-in-one-with-options | 2024-08-29 19:42:11 +0000 UTC Normal Pod check-span-lqgql.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 19:42:22 | examples-all-in-one-with-options | 2024-08-29 19:42:11 +0000 UTC Normal Pod check-span-lqgql.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 819ms (819ms including waiting) kubelet logger.go:42: 19:42:22 | examples-all-in-one-with-options | 2024-08-29 19:42:11 +0000 UTC Normal Pod report-span-qq6kq.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 798ms (798ms including waiting) kubelet logger.go:42: 19:42:22 | examples-all-in-one-with-options | 2024-08-29 19:42:11 +0000 UTC Normal Pod report-span-qq6kq.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 19:42:22 | examples-all-in-one-with-options | 2024-08-29 19:42:11 +0000 UTC Normal Pod report-span-qq6kq.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 19:42:22 | examples-all-in-one-with-options | 2024-08-29 19:42:12 +0000 UTC Normal Pod check-span-lqgql.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 19:42:22 | examples-all-in-one-with-options | 2024-08-29 19:42:12 +0000 UTC Normal Pod check-span-lqgql.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 19:42:22 | examples-all-in-one-with-options | 2024-08-29 19:42:22 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 19:42:22 | examples-all-in-one-with-options | Deleting namespace: kuttl-test-superb-seasnail === CONT kuttl/harness/examples-agent-with-priority-class logger.go:42: 19:42:34 | examples-agent-with-priority-class | Creating namespace: kuttl-test-loyal-leech logger.go:42: 19:42:34 | examples-agent-with-priority-class/0-install | starting test step 0-install logger.go:42: 19:42:35 | examples-agent-with-priority-class/0-install | SecurityContextConstraints:/daemonset-with-hostport created logger.go:42: 19:42:35 | examples-agent-with-priority-class/0-install | ServiceAccount:kuttl-test-loyal-leech/jaeger-agent-daemonset created logger.go:42: 19:42:35 | examples-agent-with-priority-class/0-install | test step completed 0-install logger.go:42: 19:42:35 | examples-agent-with-priority-class/1-add-policy | starting test step 1-add-policy logger.go:42: 19:42:35 | examples-agent-with-priority-class/1-add-policy | running command: [sh -c oc adm policy --namespace $NAMESPACE add-scc-to-user daemonset-with-hostport -z jaeger-agent-daemonset] logger.go:42: 19:42:35 | examples-agent-with-priority-class/1-add-policy | clusterrole.rbac.authorization.k8s.io/system:openshift:scc:daemonset-with-hostport added: "jaeger-agent-daemonset" logger.go:42: 19:42:35 | examples-agent-with-priority-class/1-add-policy | running command: [sh -c sleep 5] logger.go:42: 19:42:41 | examples-agent-with-priority-class/1-add-policy | test step completed 1-add-policy logger.go:42: 19:42:41 | examples-agent-with-priority-class/2-install | starting test step 2-install logger.go:42: 19:42:41 | examples-agent-with-priority-class/2-install | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE agent-as-daemonset /dev/null] logger.go:42: 19:42:50 | examples-agent-with-priority-class/2-install | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://agent-as-daemonset-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://agent-as-daemonset-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 19:42:50 | examples-agent-with-priority-class/2-install | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 19:42:51 | examples-agent-with-priority-class/2-install | job.batch/report-span created logger.go:42: 19:42:51 | examples-agent-with-priority-class/2-install | job.batch/check-span created logger.go:42: 19:42:51 | examples-agent-with-priority-class/2-install | PriorityClass:/high-priority created logger.go:42: 19:42:52 | examples-agent-with-priority-class/2-install | Jaeger:kuttl-test-loyal-leech/agent-as-daemonset updated logger.go:42: 19:43:03 | examples-agent-with-priority-class/2-install | test step completed 2-install logger.go:42: 19:43:03 | examples-agent-with-priority-class | examples-agent-with-priority-class events from ns kuttl-test-loyal-leech: logger.go:42: 19:43:03 | examples-agent-with-priority-class | 2024-08-29 19:42:43 +0000 UTC Normal Pod agent-as-daemonset-6d9d4bc599-r554v Binding Scheduled Successfully assigned kuttl-test-loyal-leech/agent-as-daemonset-6d9d4bc599-r554v to ip-10-0-51-237.us-west-2.compute.internal default-scheduler logger.go:42: 19:43:03 | examples-agent-with-priority-class | 2024-08-29 19:42:43 +0000 UTC Normal Pod agent-as-daemonset-6d9d4bc599-r554v AddedInterface Add eth0 [10.130.0.48/23] from ovn-kubernetes multus logger.go:42: 19:43:03 | examples-agent-with-priority-class | 2024-08-29 19:42:43 +0000 UTC Normal ReplicaSet.apps agent-as-daemonset-6d9d4bc599 SuccessfulCreate Created pod: agent-as-daemonset-6d9d4bc599-r554v replicaset-controller logger.go:42: 19:43:03 | examples-agent-with-priority-class | 2024-08-29 19:42:43 +0000 UTC Normal Deployment.apps agent-as-daemonset ScalingReplicaSet Scaled up replica set agent-as-daemonset-6d9d4bc599 to 1 deployment-controller logger.go:42: 19:43:03 | examples-agent-with-priority-class | 2024-08-29 19:42:44 +0000 UTC Normal Pod agent-as-daemonset-6d9d4bc599-r554v.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:345d3b6c507df9f8a18948a594e7a03698085689c6965faf79b8aadc74a0ceaa" already present on machine kubelet logger.go:42: 19:43:03 | examples-agent-with-priority-class | 2024-08-29 19:42:44 +0000 UTC Normal Pod agent-as-daemonset-6d9d4bc599-r554v.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 19:43:03 | examples-agent-with-priority-class | 2024-08-29 19:42:44 +0000 UTC Normal Pod agent-as-daemonset-6d9d4bc599-r554v.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 19:43:03 | examples-agent-with-priority-class | 2024-08-29 19:42:44 +0000 UTC Normal Pod agent-as-daemonset-6d9d4bc599-r554v.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1613032062a68faa414cd5754a666b7b78ad0e52cf328eaaed355719423a55dd" already present on machine kubelet logger.go:42: 19:43:03 | examples-agent-with-priority-class | 2024-08-29 19:42:44 +0000 UTC Normal Pod agent-as-daemonset-6d9d4bc599-r554v.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 19:43:03 | examples-agent-with-priority-class | 2024-08-29 19:42:44 +0000 UTC Normal Pod agent-as-daemonset-6d9d4bc599-r554v.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 19:43:03 | examples-agent-with-priority-class | 2024-08-29 19:42:46 +0000 UTC Warning Pod agent-as-daemonset-6d9d4bc599-r554v FailedMount MountVolume.SetUp failed for volume "agent-as-daemonset-service-ca" : configmap references non-existent config key: service-ca.crt kubelet logger.go:42: 19:43:03 | examples-agent-with-priority-class | 2024-08-29 19:42:51 +0000 UTC Normal Pod check-span-d9qvf Binding Scheduled Successfully assigned kuttl-test-loyal-leech/check-span-d9qvf to ip-10-0-51-237.us-west-2.compute.internal default-scheduler logger.go:42: 19:43:03 | examples-agent-with-priority-class | 2024-08-29 19:42:51 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-d9qvf job-controller logger.go:42: 19:43:03 | examples-agent-with-priority-class | 2024-08-29 19:42:51 +0000 UTC Normal Pod report-span-m8qsx Binding Scheduled Successfully assigned kuttl-test-loyal-leech/report-span-m8qsx to ip-10-0-51-237.us-west-2.compute.internal default-scheduler logger.go:42: 19:43:03 | examples-agent-with-priority-class | 2024-08-29 19:42:51 +0000 UTC Normal Pod report-span-m8qsx AddedInterface Add eth0 [10.130.0.49/23] from ovn-kubernetes multus logger.go:42: 19:43:03 | examples-agent-with-priority-class | 2024-08-29 19:42:51 +0000 UTC Normal Pod report-span-m8qsx.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 19:43:03 | examples-agent-with-priority-class | 2024-08-29 19:42:51 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-m8qsx job-controller logger.go:42: 19:43:03 | examples-agent-with-priority-class | 2024-08-29 19:42:52 +0000 UTC Warning DaemonSet.apps agent-as-daemonset-agent-daemonset FailedCreate Error creating: pods "agent-as-daemonset-agent-daemonset-" is forbidden: unable to validate against any security context constraint: [provider "anyuid": Forbidden: not usable by user or serviceaccount, provider restricted-v2: .containers[0].containers[0].hostPort: Invalid value: 5775: Host ports are not allowed to be used, provider restricted-v2: .containers[0].containers[0].hostPort: Invalid value: 5778: Host ports are not allowed to be used, provider restricted-v2: .containers[0].containers[0].hostPort: Invalid value: 6831: Host ports are not allowed to be used, provider restricted-v2: .containers[0].containers[0].hostPort: Invalid value: 6832: Host ports are not allowed to be used, provider restricted-v2: .containers[0].containers[0].hostPort: Invalid value: 14271: Host ports are not allowed to be used, provider "restricted": Forbidden: not usable by user or serviceaccount, provider "nonroot-v2": Forbidden: not usable by user or serviceaccount, provider "nonroot": Forbidden: not usable by user or serviceaccount, provider "pcap-dedicated-admins": Forbidden: not usable by user or serviceaccount, provider "hostmount-anyuid": Forbidden: not usable by user or serviceaccount, provider "daemonset-with-hostport": Forbidden: not usable by user or serviceaccount, provider "hostnetwork-v2": Forbidden: not usable by user or serviceaccount, provider "hostnetwork": Forbidden: not usable by user or serviceaccount, provider "hostaccess": Forbidden: not usable by user or serviceaccount, provider "node-exporter": Forbidden: not usable by user or serviceaccount, provider "privileged": Forbidden: not usable by user or serviceaccount] daemonset-controller logger.go:42: 19:43:03 | examples-agent-with-priority-class | 2024-08-29 19:42:52 +0000 UTC Normal Pod check-span-d9qvf AddedInterface Add eth0 [10.130.0.50/23] from ovn-kubernetes multus logger.go:42: 19:43:03 | examples-agent-with-priority-class | 2024-08-29 19:42:52 +0000 UTC Normal Pod check-span-d9qvf.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 19:43:03 | examples-agent-with-priority-class | 2024-08-29 19:42:52 +0000 UTC Normal Pod report-span-m8qsx.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 922ms (922ms including waiting) kubelet logger.go:42: 19:43:03 | examples-agent-with-priority-class | 2024-08-29 19:42:52 +0000 UTC Normal Pod report-span-m8qsx.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 19:43:03 | examples-agent-with-priority-class | 2024-08-29 19:42:52 +0000 UTC Normal Pod report-span-m8qsx.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 19:43:03 | examples-agent-with-priority-class | 2024-08-29 19:42:53 +0000 UTC Normal Pod check-span-d9qvf.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 829ms (829ms including waiting) kubelet logger.go:42: 19:43:03 | examples-agent-with-priority-class | 2024-08-29 19:42:53 +0000 UTC Normal Pod check-span-d9qvf.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 19:43:03 | examples-agent-with-priority-class | 2024-08-29 19:42:53 +0000 UTC Normal Pod check-span-d9qvf.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 19:43:03 | examples-agent-with-priority-class | 2024-08-29 19:43:03 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 19:43:04 | examples-agent-with-priority-class | Deleting namespace: kuttl-test-loyal-leech === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- FAIL: kuttl (1599.45s) --- FAIL: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (6.12s) --- FAIL: kuttl/harness/examples-simple-prod (607.48s) --- PASS: kuttl/harness/examples-with-sampling (65.42s) --- PASS: kuttl/harness/examples-with-cassandra (57.49s) --- PASS: kuttl/harness/examples-with-badger (37.06s) --- PASS: kuttl/harness/examples-simplest (36.89s) --- FAIL: kuttl/harness/examples-simple-prod-with-volumes (607.15s) --- PASS: kuttl/harness/examples-business-application-injected-sidecar (33.62s) --- PASS: kuttl/harness/examples-service-types (51.69s) --- PASS: kuttl/harness/examples-openshift-with-htpasswd (22.34s) --- PASS: kuttl/harness/examples-all-in-one-with-options (37.30s) --- PASS: kuttl/harness/examples-agent-with-priority-class (36.29s) FAIL + exit_code=1 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name examples --report --output /logs/artifacts/examples.xml ./artifacts/kuttl-report.xml time="2024-08-29T19:43:12Z" level=debug msg="Setting a new name for the test suites" time="2024-08-29T19:43:12Z" level=debug msg="Removing 'artifacts' TestCase" time="2024-08-29T19:43:12Z" level=debug msg="normalizing test case names" time="2024-08-29T19:43:12Z" level=debug msg="examples/artifacts -> examples_artifacts" time="2024-08-29T19:43:12Z" level=debug msg="examples/examples-simple-prod -> examples_examples_simple_prod" time="2024-08-29T19:43:12Z" level=debug msg="examples/examples-with-sampling -> examples_examples_with_sampling" time="2024-08-29T19:43:12Z" level=debug msg="examples/examples-with-cassandra -> examples_examples_with_cassandra" time="2024-08-29T19:43:12Z" level=debug msg="examples/examples-with-badger -> examples_examples_with_badger" time="2024-08-29T19:43:12Z" level=debug msg="examples/examples-simplest -> examples_examples_simplest" time="2024-08-29T19:43:12Z" level=debug msg="examples/examples-simple-prod-with-volumes -> examples_examples_simple_prod_with_volumes" time="2024-08-29T19:43:12Z" level=debug msg="examples/examples-business-application-injected-sidecar -> examples_examples_business_application_injected_sidecar" time="2024-08-29T19:43:12Z" level=debug msg="examples/examples-service-types -> examples_examples_service_types" time="2024-08-29T19:43:12Z" level=debug msg="examples/examples-openshift-with-htpasswd -> examples_examples_openshift_with_htpasswd" time="2024-08-29T19:43:12Z" level=debug msg="examples/examples-all-in-one-with-options -> examples_examples_all_in_one_with_options" time="2024-08-29T19:43:12Z" level=debug msg="examples/examples-agent-with-priority-class -> examples_examples_agent_with_priority_class" +---------------------------------------------------------+--------+ | NAME | RESULT | +---------------------------------------------------------+--------+ | examples_artifacts | passed | | examples_examples_simple_prod | failed | | examples_examples_with_sampling | passed | | examples_examples_with_cassandra | passed | | examples_examples_with_badger | passed | | examples_examples_simplest | passed | | examples_examples_simple_prod_with_volumes | failed | | examples_examples_business_application_injected_sidecar | passed | | examples_examples_service_types | passed | | examples_examples_openshift_with_htpasswd | passed | | examples_examples_all_in_one_with_options | passed | | examples_examples_agent_with_priority_class | passed | +---------------------------------------------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh generate false true + '[' 3 -ne 3 ']' + test_suite_name=generate + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/generate.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-generate make[2]: Entering directory '/tmp/jaeger-tests' test -s /tmp/jaeger-tests/bin/operator-sdk || curl -sLo /tmp/jaeger-tests/bin/operator-sdk https://github.com/operator-framework/operator-sdk/releases/download/v1.32.0/operator-sdk_`go env GOOS`_`go env GOARCH` ./hack/install/install-golangci-lint.sh Installing golangci-lint golangci-lint 1.55.2 is installed already ./hack/install/install-goimports.sh Installing goimports Try 0... go install golang.org/x/tools/cmd/goimports@v0.1.12 >>>> Formatting code... ./.ci/format.sh >>>> Building... ./hack/install/install-dependencies.sh Installing go dependencies Try 0... go mod download GOOS= GOARCH= CGO_ENABLED=0 GO111MODULE=on go build -ldflags "-X "github.com/jaegertracing/jaeger-operator/pkg/version".version="1.60.1" -X "github.com/jaegertracing/jaeger-operator/pkg/version".buildDate=2024-08-29T19:43:18Z -X "github.com/jaegertracing/jaeger-operator/pkg/version".defaultJaeger="1.60.0"" -o "bin/jaeger-operator" main.go JAEGER_VERSION="1.60.0" ./tests/e2e/generate/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.16.8 True False 39m Cluster version is 4.16.8' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.16.8 True False 39m Cluster version is 4.16.8' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/generate/render.sh ++ export SUITE_DIR=./tests/e2e/generate ++ SUITE_DIR=./tests/e2e/generate ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/generate ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + '[' true = true ']' + skip_test generate 'This test was skipped until https://github.com/jaegertracing/jaeger-operator/issues/2145 is fixed' + '[' 2 -ne 2 ']' + test_name=generate + message='This test was skipped until https://github.com/jaegertracing/jaeger-operator/issues/2145 is fixed' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/generate/_build + '[' _build '!=' _build ']' + rm -rf generate + warning 'generate: This test was skipped until https://github.com/jaegertracing/jaeger-operator/issues/2145 is fixed' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: generate: This test was skipped until https://github.com/jaegertracing/jaeger-operator/issues/2145 is fixed\e[0m' WAR: generate: This test was skipped until https://github.com/jaegertracing/jaeger-operator/issues/2145 is fixed make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running generate E2E tests' Running generate E2E tests + cd tests/e2e/generate/_build + set +e + KUBECONFIG=/tmp/kubeconfig-2030369819 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-rosa-h-yd3i.24xa.s3.devshift.org:443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 1 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === CONT kuttl/harness/artifacts logger.go:42: 19:43:38 | artifacts | Creating namespace: kuttl-test-novel-stork logger.go:42: 19:43:39 | artifacts | artifacts events from ns kuttl-test-novel-stork: logger.go:42: 19:43:39 | artifacts | Deleting namespace: kuttl-test-novel-stork === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (7.91s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (6.31s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name generate --report --output /logs/artifacts/generate.xml ./artifacts/kuttl-report.xml time="2024-08-29T19:43:45Z" level=debug msg="Setting a new name for the test suites" time="2024-08-29T19:43:45Z" level=debug msg="Removing 'artifacts' TestCase" time="2024-08-29T19:43:45Z" level=debug msg="normalizing test case names" time="2024-08-29T19:43:45Z" level=debug msg="generate/artifacts -> generate_artifacts" +--------------------+--------+ | NAME | RESULT | +--------------------+--------+ | generate_artifacts | passed | +--------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh miscellaneous false true + '[' 3 -ne 3 ']' + test_suite_name=miscellaneous + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/miscellaneous.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-miscellaneous make[2]: Entering directory '/tmp/jaeger-tests' SKIP_ES_EXTERNAL=true ./tests/e2e/miscellaneous/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.16.8 True False 39m Cluster version is 4.16.8' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.16.8 True False 39m Cluster version is 4.16.8' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/miscellaneous/render.sh ++ export SUITE_DIR=./tests/e2e/miscellaneous ++ SUITE_DIR=./tests/e2e/miscellaneous ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/miscellaneous ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + start_test collector-autoscale + '[' 1 -ne 1 ']' + test_name=collector-autoscale + echo =========================================================================== =========================================================================== + info 'Rendering files for test collector-autoscale' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test collector-autoscale\e[0m' Rendering files for test collector-autoscale + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build + '[' _build '!=' _build ']' + mkdir -p collector-autoscale + cd collector-autoscale + jaeger_name=simple-prod + jaeger_deploy_mode=production + [[ true = true ]] + [[ true = true ]] + jaeger_deploy_mode=production_autoprovisioned + ELASTICSEARCH_NODECOUNT=1 + render_install_jaeger simple-prod production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=simple-prod + JAEGER_NAME=simple-prod + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.collector.resources.requests.memory="200m"' 01-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.collector.autoscale=true 01-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.collector.minReplicas=1 01-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.collector.maxReplicas=2 01-install.yaml + version_lt 1.30 1.23 ++ echo 1.30 1.23 ++ tr ' ' '\n' ++ sort -rV ++ head -n 1 + test 1.30 '!=' 1.30 + rm ./03-assert.yaml + generate_otlp_e2e_tests http + test_protocol=http + is_secured=false + '[' true = true ']' + is_secured=true + start_test collector-otlp-allinone-http + '[' 1 -ne 1 ']' + test_name=collector-otlp-allinone-http + echo =========================================================================== =========================================================================== + info 'Rendering files for test collector-otlp-allinone-http' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test collector-otlp-allinone-http\e[0m' Rendering files for test collector-otlp-allinone-http + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/collector-autoscale + '[' collector-autoscale '!=' _build ']' + cd .. + mkdir -p collector-otlp-allinone-http + cd collector-otlp-allinone-http + render_install_jaeger my-jaeger allInOne 00 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=allInOne + test_step=00 + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_otlp_smoke_test my-jaeger http true 01 + '[' 4 -ne 4 ']' + jaeger=my-jaeger + reporting_protocol=http + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template + '[' http = grpc ']' + reporting_port=:4318 + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 + OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + REPORTING_PROTOCOL=http + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset OTEL_EXPORTER_OTLP_ENDPOINT + start_test collector-otlp-production-http + '[' 1 -ne 1 ']' + test_name=collector-otlp-production-http + echo =========================================================================== =========================================================================== + info 'Rendering files for test collector-otlp-production-http' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test collector-otlp-production-http\e[0m' Rendering files for test collector-otlp-production-http + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/collector-otlp-allinone-http + '[' collector-otlp-allinone-http '!=' _build ']' + cd .. + mkdir -p collector-otlp-production-http + cd collector-otlp-production-http + jaeger_deploy_mode=production + [[ true = true ]] + [[ true = true ]] + jaeger_deploy_mode=production_autoprovisioned + render_install_jaeger my-jaeger production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + render_otlp_smoke_test my-jaeger http true 02 + '[' 4 -ne 4 ']' + jaeger=my-jaeger + reporting_protocol=http + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template + '[' http = grpc ']' + reporting_port=:4318 + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 + OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + REPORTING_PROTOCOL=http + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset OTEL_EXPORTER_OTLP_ENDPOINT + generate_otlp_e2e_tests grpc + test_protocol=grpc + is_secured=false + '[' true = true ']' + is_secured=true + start_test collector-otlp-allinone-grpc + '[' 1 -ne 1 ']' + test_name=collector-otlp-allinone-grpc + echo =========================================================================== =========================================================================== + info 'Rendering files for test collector-otlp-allinone-grpc' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test collector-otlp-allinone-grpc\e[0m' Rendering files for test collector-otlp-allinone-grpc + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/collector-otlp-production-http + '[' collector-otlp-production-http '!=' _build ']' + cd .. + mkdir -p collector-otlp-allinone-grpc + cd collector-otlp-allinone-grpc + render_install_jaeger my-jaeger allInOne 00 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=allInOne + test_step=00 + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_otlp_smoke_test my-jaeger grpc true 01 + '[' 4 -ne 4 ']' + jaeger=my-jaeger + reporting_protocol=grpc + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template + '[' grpc = grpc ']' + reporting_port=:4317 + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 + OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + REPORTING_PROTOCOL=grpc + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset OTEL_EXPORTER_OTLP_ENDPOINT + start_test collector-otlp-production-grpc + '[' 1 -ne 1 ']' + test_name=collector-otlp-production-grpc + echo =========================================================================== =========================================================================== + info 'Rendering files for test collector-otlp-production-grpc' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test collector-otlp-production-grpc\e[0m' Rendering files for test collector-otlp-production-grpc + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/collector-otlp-allinone-grpc + '[' collector-otlp-allinone-grpc '!=' _build ']' + cd .. + mkdir -p collector-otlp-production-grpc + cd collector-otlp-production-grpc + jaeger_deploy_mode=production + [[ true = true ]] + [[ true = true ]] + jaeger_deploy_mode=production_autoprovisioned + render_install_jaeger my-jaeger production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + render_otlp_smoke_test my-jaeger grpc true 02 + '[' 4 -ne 4 ']' + jaeger=my-jaeger + reporting_protocol=grpc + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template + '[' grpc = grpc ']' + reporting_port=:4317 + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 + OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + REPORTING_PROTOCOL=grpc + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset OTEL_EXPORTER_OTLP_ENDPOINT + '[' true = true ']' + skip_test istio 'Test not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=istio + message='Test not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/collector-otlp-production-grpc + '[' collector-otlp-production-grpc '!=' _build ']' + cd .. + rm -rf istio + warning 'istio: Test not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: istio: Test not supported in OpenShift\e[0m' WAR: istio: Test not supported in OpenShift + '[' true = true ']' + skip_test outside-cluster 'Test not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=outside-cluster + message='Test not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build + '[' _build '!=' _build ']' + rm -rf outside-cluster + warning 'outside-cluster: Test not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: outside-cluster: Test not supported in OpenShift\e[0m' WAR: outside-cluster: Test not supported in OpenShift + start_test set-custom-img + '[' 1 -ne 1 ']' + test_name=set-custom-img + echo =========================================================================== =========================================================================== + info 'Rendering files for test set-custom-img' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test set-custom-img\e[0m' Rendering files for test set-custom-img + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build + '[' _build '!=' _build ']' + mkdir -p set-custom-img + cd set-custom-img + jaeger_name=my-jaeger + jaeger_deploy_mode=production + [[ true = true ]] + [[ true = true ]] + jaeger_deploy_mode=production_autoprovisioned + render_install_jaeger my-jaeger production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + cp ./01-install.yaml ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.collector.image="test"' ./02-install.yaml + '[' true = true ']' + skip_test non-cluster-wide 'Test not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=non-cluster-wide + message='Test not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/set-custom-img + '[' set-custom-img '!=' _build ']' + cd .. + rm -rf non-cluster-wide + warning 'non-cluster-wide: Test not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: non-cluster-wide: Test not supported in OpenShift\e[0m' WAR: non-cluster-wide: Test not supported in OpenShift make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running miscellaneous E2E tests' Running miscellaneous E2E tests + cd tests/e2e/miscellaneous/_build + set +e + KUBECONFIG=/tmp/kubeconfig-2030369819 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-rosa-h-yd3i.24xa.s3.devshift.org:443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 8 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/cassandra-spark === PAUSE kuttl/harness/cassandra-spark === RUN kuttl/harness/collector-autoscale === PAUSE kuttl/harness/collector-autoscale === RUN kuttl/harness/collector-otlp-allinone-grpc === PAUSE kuttl/harness/collector-otlp-allinone-grpc === RUN kuttl/harness/collector-otlp-allinone-http === PAUSE kuttl/harness/collector-otlp-allinone-http === RUN kuttl/harness/collector-otlp-production-grpc === PAUSE kuttl/harness/collector-otlp-production-grpc === RUN kuttl/harness/collector-otlp-production-http === PAUSE kuttl/harness/collector-otlp-production-http === RUN kuttl/harness/set-custom-img === PAUSE kuttl/harness/set-custom-img === CONT kuttl/harness/artifacts logger.go:42: 19:43:57 | artifacts | Creating namespace: kuttl-test-integral-pug logger.go:42: 19:43:57 | artifacts | artifacts events from ns kuttl-test-integral-pug: logger.go:42: 19:43:57 | artifacts | Deleting namespace: kuttl-test-integral-pug === CONT kuttl/harness/collector-otlp-allinone-http logger.go:42: 19:44:03 | collector-otlp-allinone-http | Creating namespace: kuttl-test-gorgeous-earwig logger.go:42: 19:44:03 | collector-otlp-allinone-http/0-install | starting test step 0-install logger.go:42: 19:44:04 | collector-otlp-allinone-http/0-install | Jaeger:kuttl-test-gorgeous-earwig/my-jaeger created logger.go:42: 19:44:07 | collector-otlp-allinone-http/0-install | test step completed 0-install logger.go:42: 19:44:07 | collector-otlp-allinone-http/1-smoke-test | starting test step 1-smoke-test logger.go:42: 19:44:07 | collector-otlp-allinone-http/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 19:44:09 | collector-otlp-allinone-http/1-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 19:44:16 | collector-otlp-allinone-http/1-smoke-test | running command: [sh -c REPORTING_PROTOCOL=http ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/otlp-smoke-test.yaml.template -o otlp-smoke-test-job.yaml] logger.go:42: 19:44:17 | collector-otlp-allinone-http/1-smoke-test | running command: [sh -c kubectl create -f otlp-smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 19:44:17 | collector-otlp-allinone-http/1-smoke-test | job.batch/report-span created logger.go:42: 19:44:17 | collector-otlp-allinone-http/1-smoke-test | job.batch/check-span created logger.go:42: 19:44:30 | collector-otlp-allinone-http/1-smoke-test | test step completed 1-smoke-test logger.go:42: 19:44:30 | collector-otlp-allinone-http | collector-otlp-allinone-http events from ns kuttl-test-gorgeous-earwig: logger.go:42: 19:44:30 | collector-otlp-allinone-http | 2024-08-29 19:44:04 +0000 UTC Normal Pod my-jaeger-5c85d999b6-fvm65 Binding Scheduled Successfully assigned kuttl-test-gorgeous-earwig/my-jaeger-5c85d999b6-fvm65 to ip-10-0-51-237.us-west-2.compute.internal default-scheduler logger.go:42: 19:44:30 | collector-otlp-allinone-http | 2024-08-29 19:44:04 +0000 UTC Normal Pod my-jaeger-5c85d999b6-fvm65 AddedInterface Add eth0 [10.130.0.51/23] from ovn-kubernetes multus logger.go:42: 19:44:30 | collector-otlp-allinone-http | 2024-08-29 19:44:04 +0000 UTC Normal Pod my-jaeger-5c85d999b6-fvm65.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:345d3b6c507df9f8a18948a594e7a03698085689c6965faf79b8aadc74a0ceaa" already present on machine kubelet logger.go:42: 19:44:30 | collector-otlp-allinone-http | 2024-08-29 19:44:04 +0000 UTC Normal ReplicaSet.apps my-jaeger-5c85d999b6 SuccessfulCreate Created pod: my-jaeger-5c85d999b6-fvm65 replicaset-controller logger.go:42: 19:44:30 | collector-otlp-allinone-http | 2024-08-29 19:44:04 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-5c85d999b6 to 1 deployment-controller logger.go:42: 19:44:30 | collector-otlp-allinone-http | 2024-08-29 19:44:05 +0000 UTC Normal Pod my-jaeger-5c85d999b6-fvm65.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 19:44:30 | collector-otlp-allinone-http | 2024-08-29 19:44:05 +0000 UTC Normal Pod my-jaeger-5c85d999b6-fvm65.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 19:44:30 | collector-otlp-allinone-http | 2024-08-29 19:44:05 +0000 UTC Normal Pod my-jaeger-5c85d999b6-fvm65.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1613032062a68faa414cd5754a666b7b78ad0e52cf328eaaed355719423a55dd" already present on machine kubelet logger.go:42: 19:44:30 | collector-otlp-allinone-http | 2024-08-29 19:44:05 +0000 UTC Normal Pod my-jaeger-5c85d999b6-fvm65.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 19:44:30 | collector-otlp-allinone-http | 2024-08-29 19:44:05 +0000 UTC Normal Pod my-jaeger-5c85d999b6-fvm65.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 19:44:30 | collector-otlp-allinone-http | 2024-08-29 19:44:09 +0000 UTC Normal Pod my-jaeger-5c85d999b6-fvm65.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 19:44:30 | collector-otlp-allinone-http | 2024-08-29 19:44:09 +0000 UTC Normal Pod my-jaeger-5c85d999b6-fvm65.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 19:44:30 | collector-otlp-allinone-http | 2024-08-29 19:44:09 +0000 UTC Normal ReplicaSet.apps my-jaeger-5c85d999b6 SuccessfulDelete Deleted pod: my-jaeger-5c85d999b6-fvm65 replicaset-controller logger.go:42: 19:44:30 | collector-otlp-allinone-http | 2024-08-29 19:44:09 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled down replica set my-jaeger-5c85d999b6 to 0 from 1 deployment-controller logger.go:42: 19:44:30 | collector-otlp-allinone-http | 2024-08-29 19:44:10 +0000 UTC Normal Pod my-jaeger-769c6cfb66-xt6z9 Binding Scheduled Successfully assigned kuttl-test-gorgeous-earwig/my-jaeger-769c6cfb66-xt6z9 to ip-10-0-51-237.us-west-2.compute.internal default-scheduler logger.go:42: 19:44:30 | collector-otlp-allinone-http | 2024-08-29 19:44:10 +0000 UTC Normal ReplicaSet.apps my-jaeger-769c6cfb66 SuccessfulCreate Created pod: my-jaeger-769c6cfb66-xt6z9 replicaset-controller logger.go:42: 19:44:30 | collector-otlp-allinone-http | 2024-08-29 19:44:10 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-769c6cfb66 to 1 deployment-controller logger.go:42: 19:44:30 | collector-otlp-allinone-http | 2024-08-29 19:44:11 +0000 UTC Normal Pod my-jaeger-769c6cfb66-xt6z9 AddedInterface Add eth0 [10.130.0.52/23] from ovn-kubernetes multus logger.go:42: 19:44:30 | collector-otlp-allinone-http | 2024-08-29 19:44:11 +0000 UTC Normal Pod my-jaeger-769c6cfb66-xt6z9.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:345d3b6c507df9f8a18948a594e7a03698085689c6965faf79b8aadc74a0ceaa" already present on machine kubelet logger.go:42: 19:44:30 | collector-otlp-allinone-http | 2024-08-29 19:44:11 +0000 UTC Normal Pod my-jaeger-769c6cfb66-xt6z9.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 19:44:30 | collector-otlp-allinone-http | 2024-08-29 19:44:11 +0000 UTC Normal Pod my-jaeger-769c6cfb66-xt6z9.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 19:44:30 | collector-otlp-allinone-http | 2024-08-29 19:44:11 +0000 UTC Normal Pod my-jaeger-769c6cfb66-xt6z9.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1613032062a68faa414cd5754a666b7b78ad0e52cf328eaaed355719423a55dd" already present on machine kubelet logger.go:42: 19:44:30 | collector-otlp-allinone-http | 2024-08-29 19:44:11 +0000 UTC Normal Pod my-jaeger-769c6cfb66-xt6z9.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 19:44:30 | collector-otlp-allinone-http | 2024-08-29 19:44:11 +0000 UTC Normal Pod my-jaeger-769c6cfb66-xt6z9.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 19:44:30 | collector-otlp-allinone-http | 2024-08-29 19:44:17 +0000 UTC Normal Pod check-span-446qw Binding Scheduled Successfully assigned kuttl-test-gorgeous-earwig/check-span-446qw to ip-10-0-51-237.us-west-2.compute.internal default-scheduler logger.go:42: 19:44:30 | collector-otlp-allinone-http | 2024-08-29 19:44:17 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-446qw job-controller logger.go:42: 19:44:30 | collector-otlp-allinone-http | 2024-08-29 19:44:17 +0000 UTC Normal Pod report-span-rlzz7 Binding Scheduled Successfully assigned kuttl-test-gorgeous-earwig/report-span-rlzz7 to ip-10-0-51-237.us-west-2.compute.internal default-scheduler logger.go:42: 19:44:30 | collector-otlp-allinone-http | 2024-08-29 19:44:17 +0000 UTC Normal Pod report-span-rlzz7 AddedInterface Add eth0 [10.130.0.53/23] from ovn-kubernetes multus logger.go:42: 19:44:30 | collector-otlp-allinone-http | 2024-08-29 19:44:17 +0000 UTC Normal Pod report-span-rlzz7.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 19:44:30 | collector-otlp-allinone-http | 2024-08-29 19:44:17 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-rlzz7 job-controller logger.go:42: 19:44:30 | collector-otlp-allinone-http | 2024-08-29 19:44:18 +0000 UTC Normal Pod check-span-446qw AddedInterface Add eth0 [10.130.0.54/23] from ovn-kubernetes multus logger.go:42: 19:44:30 | collector-otlp-allinone-http | 2024-08-29 19:44:18 +0000 UTC Normal Pod check-span-446qw.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 19:44:30 | collector-otlp-allinone-http | 2024-08-29 19:44:18 +0000 UTC Normal Pod report-span-rlzz7.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 818ms (818ms including waiting) kubelet logger.go:42: 19:44:30 | collector-otlp-allinone-http | 2024-08-29 19:44:18 +0000 UTC Normal Pod report-span-rlzz7.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 19:44:30 | collector-otlp-allinone-http | 2024-08-29 19:44:18 +0000 UTC Normal Pod report-span-rlzz7.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 19:44:30 | collector-otlp-allinone-http | 2024-08-29 19:44:19 +0000 UTC Normal Pod check-span-446qw.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 832ms (832ms including waiting) kubelet logger.go:42: 19:44:30 | collector-otlp-allinone-http | 2024-08-29 19:44:19 +0000 UTC Normal Pod check-span-446qw.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 19:44:30 | collector-otlp-allinone-http | 2024-08-29 19:44:19 +0000 UTC Normal Pod check-span-446qw.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 19:44:30 | collector-otlp-allinone-http | 2024-08-29 19:44:29 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 19:44:30 | collector-otlp-allinone-http | Deleting namespace: kuttl-test-gorgeous-earwig === CONT kuttl/harness/set-custom-img logger.go:42: 19:44:42 | set-custom-img | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 19:44:42 | set-custom-img | Ignoring check-collector-img.sh as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 19:44:42 | set-custom-img | Creating namespace: kuttl-test-touching-ferret logger.go:42: 19:44:42 | set-custom-img/1-install | starting test step 1-install logger.go:42: 19:44:43 | set-custom-img/1-install | Jaeger:kuttl-test-touching-ferret/my-jaeger created logger.go:42: 19:54:43 | set-custom-img/1-install | test step failed 1-install case.go:364: failed in step 1-install case.go:366: --- Deployment:kuttl-test-touching-ferret/my-jaeger-collector +++ Deployment:kuttl-test-touching-ferret/my-jaeger-collector @@ -1,10 +1,412 @@ apiVersion: apps/v1 kind: Deployment metadata: + annotations: + linkerd.io/inject: disabled + prometheus.io/port: "14269" + prometheus.io/scrape: "true" + labels: + app: jaeger + app.kubernetes.io/component: collector + app.kubernetes.io/instance: my-jaeger + app.kubernetes.io/managed-by: jaeger-operator + app.kubernetes.io/name: my-jaeger-collector + app.kubernetes.io/part-of: jaeger + managedFields: + - apiVersion: apps/v1 + fieldsType: FieldsV1 + fieldsV1: + f:metadata: + f:annotations: + .: {} + f:linkerd.io/inject: {} + f:prometheus.io/port: {} + f:prometheus.io/scrape: {} + f:labels: + .: {} + f:app: {} + f:app.kubernetes.io/component: {} + f:app.kubernetes.io/instance: {} + f:app.kubernetes.io/managed-by: {} + f:app.kubernetes.io/name: {} + f:app.kubernetes.io/part-of: {} + f:ownerReferences: + .: {} + k:{"uid":"b0ad0d0f-9296-44b9-bc68-f215a395844c"}: {} + f:spec: + f:progressDeadlineSeconds: {} + f:replicas: {} + f:revisionHistoryLimit: {} + f:selector: {} + f:strategy: + f:type: {} + f:template: + f:metadata: + f:annotations: + .: {} + f:linkerd.io/inject: {} + f:prometheus.io/port: {} + f:prometheus.io/scrape: {} + f:sidecar.istio.io/inject: {} + f:labels: + .: {} + f:app: {} + f:app.kubernetes.io/component: {} + f:app.kubernetes.io/instance: {} + f:app.kubernetes.io/managed-by: {} + f:app.kubernetes.io/name: {} + f:app.kubernetes.io/part-of: {} + f:spec: + f:containers: + k:{"name":"jaeger-collector"}: + .: {} + f:args: {} + f:env: + .: {} + k:{"name":"COLLECTOR_OTLP_ENABLED"}: + .: {} + f:name: {} + f:value: {} + k:{"name":"COLLECTOR_ZIPKIN_HOST_PORT"}: + .: {} + f:name: {} + f:value: {} + k:{"name":"SPAN_STORAGE_TYPE"}: + .: {} + f:name: {} + f:value: {} + f:image: {} + f:imagePullPolicy: {} + f:livenessProbe: + .: {} + f:failureThreshold: {} + f:httpGet: + .: {} + f:path: {} + f:port: {} + f:scheme: {} + f:initialDelaySeconds: {} + f:periodSeconds: {} + f:successThreshold: {} + f:timeoutSeconds: {} + f:name: {} + f:ports: + .: {} + k:{"containerPort":4317,"protocol":"TCP"}: + .: {} + f:containerPort: {} + f:name: {} + f:protocol: {} + k:{"containerPort":4318,"protocol":"TCP"}: + .: {} + f:containerPort: {} + f:name: {} + f:protocol: {} + k:{"containerPort":9411,"protocol":"TCP"}: + .: {} + f:containerPort: {} + f:name: {} + f:protocol: {} + k:{"containerPort":14250,"protocol":"TCP"}: + .: {} + f:containerPort: {} + f:name: {} + f:protocol: {} + k:{"containerPort":14267,"protocol":"TCP"}: + .: {} + f:containerPort: {} + f:name: {} + f:protocol: {} + k:{"containerPort":14268,"protocol":"TCP"}: + .: {} + f:containerPort: {} + f:name: {} + f:protocol: {} + k:{"containerPort":14269,"protocol":"TCP"}: + .: {} + f:containerPort: {} + f:name: {} + f:protocol: {} + f:readinessProbe: + .: {} + f:failureThreshold: {} + f:httpGet: + .: {} + f:path: {} + f:port: {} + f:scheme: {} + f:initialDelaySeconds: {} + f:periodSeconds: {} + f:successThreshold: {} + f:timeoutSeconds: {} + f:resources: {} + f:terminationMessagePath: {} + f:terminationMessagePolicy: {} + f:volumeMounts: + .: {} + k:{"mountPath":"/certs"}: + .: {} + f:mountPath: {} + f:name: {} + f:readOnly: {} + k:{"mountPath":"/etc/jaeger/sampling"}: + .: {} + f:mountPath: {} + f:name: {} + f:readOnly: {} + k:{"mountPath":"/etc/pki/ca-trust/extracted/pem"}: + .: {} + f:mountPath: {} + f:name: {} + f:readOnly: {} + k:{"mountPath":"/etc/tls-config"}: + .: {} + f:mountPath: {} + f:name: {} + f:readOnly: {} + f:dnsPolicy: {} + f:enableServiceLinks: {} + f:restartPolicy: {} + f:schedulerName: {} + f:securityContext: {} + f:serviceAccount: {} + f:serviceAccountName: {} + f:terminationGracePeriodSeconds: {} + f:volumes: + .: {} + k:{"name":"certs"}: + .: {} + f:name: {} + f:secret: + .: {} + f:defaultMode: {} + f:secretName: {} + k:{"name":"my-jaeger-collector-tls-config-volume"}: + .: {} + f:name: {} + f:secret: + .: {} + f:defaultMode: {} + f:secretName: {} + k:{"name":"my-jaeger-sampling-configuration-volume"}: + .: {} + f:configMap: + .: {} + f:defaultMode: {} + f:items: {} + f:name: {} + f:name: {} + k:{"name":"my-jaeger-trusted-ca"}: + .: {} + f:configMap: + .: {} + f:defaultMode: {} + f:items: {} + f:name: {} + f:name: {} + manager: jaeger-operator + operation: Update + time: "2024-08-29T19:44:47Z" + - apiVersion: apps/v1 + fieldsType: FieldsV1 + fieldsV1: + f:metadata: + f:annotations: + f:deployment.kubernetes.io/revision: {} + f:status: + f:conditions: + .: {} + k:{"type":"Available"}: + .: {} + f:lastTransitionTime: {} + f:lastUpdateTime: {} + f:message: {} + f:reason: {} + f:status: {} + f:type: {} + k:{"type":"Progressing"}: + .: {} + f:lastTransitionTime: {} + f:lastUpdateTime: {} + f:message: {} + f:reason: {} + f:status: {} + f:type: {} + f:observedGeneration: {} + f:replicas: {} + f:unavailableReplicas: {} + f:updatedReplicas: {} + manager: kube-controller-manager + operation: Update + subresource: status + time: "2024-08-29T19:44:48Z" name: my-jaeger-collector namespace: kuttl-test-touching-ferret + ownerReferences: + - apiVersion: jaegertracing.io/v1 + controller: true + kind: Jaeger + name: my-jaeger + uid: b0ad0d0f-9296-44b9-bc68-f215a395844c spec: + progressDeadlineSeconds: 600 replicas: 1 + revisionHistoryLimit: 10 + selector: + matchLabels: + app: jaeger + app.kubernetes.io/component: collector + app.kubernetes.io/instance: my-jaeger + app.kubernetes.io/managed-by: jaeger-operator + app.kubernetes.io/name: my-jaeger-collector + app.kubernetes.io/part-of: jaeger + strategy: + type: Recreate + template: + metadata: + annotations: + linkerd.io/inject: disabled + prometheus.io/port: "14269" + prometheus.io/scrape: "true" + sidecar.istio.io/inject: "false" + creationTimestamp: null + labels: + app: jaeger + app.kubernetes.io/component: collector + app.kubernetes.io/instance: my-jaeger + app.kubernetes.io/managed-by: jaeger-operator + app.kubernetes.io/name: my-jaeger-collector + app.kubernetes.io/part-of: jaeger + spec: + containers: + - args: + - --collector.grpc.tls.cert=/etc/tls-config/tls.crt + - --collector.grpc.tls.enabled=true + - --collector.grpc.tls.key=/etc/tls-config/tls.key + - --sampling.strategies-file=/etc/jaeger/sampling/sampling.json + - --es.server-urls=https://elasticsearch.kuttl-test-touching-ferret.svc.cluster.local:9200 + - --es.tls.enabled=true + - --es.tls.ca=/certs/ca + - --es.tls.cert=/certs/cert + - --es.tls.key=/certs/key + - --es.timeout=15s + - --es.num-shards=1 + - --es.num-replicas=0 + env: + - name: SPAN_STORAGE_TYPE + value: elasticsearch + - name: COLLECTOR_ZIPKIN_HOST_PORT + value: :9411 + - name: COLLECTOR_OTLP_ENABLED + value: "true" + image: registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:b90e0d264bb0b051c2a4b9e6d499f407fed58f834ab8ebf30263ee1cf7c7ea3c + imagePullPolicy: IfNotPresent + livenessProbe: + failureThreshold: 5 + httpGet: + path: / + port: 14269 + scheme: HTTP + initialDelaySeconds: 5 + periodSeconds: 15 + successThreshold: 1 + timeoutSeconds: 1 + name: jaeger-collector + ports: + - containerPort: 9411 + name: zipkin + protocol: TCP + - containerPort: 14267 + name: c-tchan-trft + protocol: TCP + - containerPort: 14268 + name: c-binary-trft + protocol: TCP + - containerPort: 14269 + name: admin-http + protocol: TCP + - containerPort: 14250 + name: grpc + protocol: TCP + - containerPort: 4317 + name: grpc-otlp + protocol: TCP + - containerPort: 4318 + name: http-otlp + protocol: TCP + readinessProbe: + failureThreshold: 3 + httpGet: + path: / + port: 14269 + scheme: HTTP + initialDelaySeconds: 1 + periodSeconds: 10 + successThreshold: 1 + timeoutSeconds: 1 + resources: {} + terminationMessagePath: /dev/termination-log + terminationMessagePolicy: File + volumeMounts: + - mountPath: /etc/jaeger/sampling + name: my-jaeger-sampling-configuration-volume + readOnly: true + - mountPath: /etc/tls-config + name: my-jaeger-collector-tls-config-volume + readOnly: true + - mountPath: /etc/pki/ca-trust/extracted/pem + name: my-jaeger-trusted-ca + readOnly: true + - mountPath: /certs + name: certs + readOnly: true + dnsPolicy: ClusterFirst + enableServiceLinks: false + restartPolicy: Always + schedulerName: default-scheduler + securityContext: {} + serviceAccount: my-jaeger + serviceAccountName: my-jaeger + terminationGracePeriodSeconds: 30 + volumes: + - configMap: + defaultMode: 420 + items: + - key: sampling + path: sampling.json + name: my-jaeger-sampling-configuration + name: my-jaeger-sampling-configuration-volume + - name: my-jaeger-collector-tls-config-volume + secret: + defaultMode: 420 + secretName: my-jaeger-collector-headless-tls + - configMap: + defaultMode: 420 + items: + - key: ca-bundle.crt + path: tls-ca-bundle.pem + name: my-jaeger-trusted-ca + name: my-jaeger-trusted-ca + - name: certs + secret: + defaultMode: 420 + secretName: my-jaeger-jaeger-elasticsearch status: - readyReplicas: 1 + conditions: + - lastTransitionTime: "2024-08-29T19:44:47Z" + lastUpdateTime: "2024-08-29T19:44:47Z" + message: Deployment does not have minimum availability. + reason: MinimumReplicasUnavailable + status: "False" + type: Available + - lastTransitionTime: "2024-08-29T19:44:47Z" + lastUpdateTime: "2024-08-29T19:44:48Z" + message: ReplicaSet "my-jaeger-collector-58c4f44fd8" is progressing. + reason: ReplicaSetUpdated + status: "True" + type: Progressing + observedGeneration: 1 + replicas: 1 + unavailableReplicas: 1 + updatedReplicas: 1 case.go:366: resource Deployment:kuttl-test-touching-ferret/my-jaeger-collector: .status.readyReplicas: key is missing from map case.go:366: --- Deployment:kuttl-test-touching-ferret/my-jaeger-query +++ Deployment:kuttl-test-touching-ferret/my-jaeger-query @@ -1,10 +1,508 @@ apiVersion: apps/v1 kind: Deployment metadata: + annotations: + linkerd.io/inject: disabled + prometheus.io/port: "16687" + prometheus.io/scrape: "true" + sidecar.jaegertracing.io/inject: my-jaeger + sidecar.jaegertracing.io/revision: "1" + labels: + app: jaeger + app.kubernetes.io/component: query + app.kubernetes.io/instance: my-jaeger + app.kubernetes.io/managed-by: jaeger-operator + app.kubernetes.io/name: my-jaeger-query + app.kubernetes.io/part-of: jaeger + sidecar.jaegertracing.io/injected: my-jaeger + managedFields: + - apiVersion: apps/v1 + fieldsType: FieldsV1 + fieldsV1: + f:metadata: + f:annotations: + .: {} + f:linkerd.io/inject: {} + f:prometheus.io/port: {} + f:prometheus.io/scrape: {} + f:sidecar.jaegertracing.io/inject: {} + f:sidecar.jaegertracing.io/revision: {} + f:labels: + .: {} + f:app: {} + f:app.kubernetes.io/component: {} + f:app.kubernetes.io/instance: {} + f:app.kubernetes.io/managed-by: {} + f:app.kubernetes.io/name: {} + f:app.kubernetes.io/part-of: {} + f:ownerReferences: + .: {} + k:{"uid":"b0ad0d0f-9296-44b9-bc68-f215a395844c"}: {} + f:spec: + f:progressDeadlineSeconds: {} + f:replicas: {} + f:revisionHistoryLimit: {} + f:selector: {} + f:strategy: + f:type: {} + f:template: + f:metadata: + f:annotations: + .: {} + f:linkerd.io/inject: {} + f:prometheus.io/port: {} + f:prometheus.io/scrape: {} + f:sidecar.istio.io/inject: {} + f:sidecar.jaegertracing.io/inject: {} + f:labels: + .: {} + f:app: {} + f:app.kubernetes.io/component: {} + f:app.kubernetes.io/instance: {} + f:app.kubernetes.io/managed-by: {} + f:app.kubernetes.io/name: {} + f:app.kubernetes.io/part-of: {} + f:spec: + f:containers: + k:{"name":"jaeger-query"}: + .: {} + f:args: {} + f:env: + .: {} + k:{"name":"JAEGER_DISABLED"}: + .: {} + f:name: {} + f:value: {} + k:{"name":"METRICS_STORAGE_TYPE"}: + .: {} + f:name: {} + k:{"name":"SPAN_STORAGE_TYPE"}: + .: {} + f:name: {} + f:value: {} + f:image: {} + f:imagePullPolicy: {} + f:livenessProbe: + .: {} + f:failureThreshold: {} + f:httpGet: + .: {} + f:path: {} + f:port: {} + f:scheme: {} + f:initialDelaySeconds: {} + f:periodSeconds: {} + f:successThreshold: {} + f:timeoutSeconds: {} + f:name: {} + f:ports: + .: {} + k:{"containerPort":16685,"protocol":"TCP"}: + .: {} + f:containerPort: {} + f:name: {} + f:protocol: {} + k:{"containerPort":16686,"protocol":"TCP"}: + .: {} + f:containerPort: {} + f:name: {} + f:protocol: {} + k:{"containerPort":16687,"protocol":"TCP"}: + .: {} + f:containerPort: {} + f:name: {} + f:protocol: {} + f:readinessProbe: + .: {} + f:failureThreshold: {} + f:httpGet: + .: {} + f:path: {} + f:port: {} + f:scheme: {} + f:initialDelaySeconds: {} + f:periodSeconds: {} + f:successThreshold: {} + f:timeoutSeconds: {} + f:resources: {} + f:terminationMessagePath: {} + f:terminationMessagePolicy: {} + f:volumeMounts: + .: {} + k:{"mountPath":"/certs"}: + .: {} + f:mountPath: {} + f:name: {} + f:readOnly: {} + k:{"mountPath":"/etc/config"}: + .: {} + f:mountPath: {} + f:name: {} + f:readOnly: {} + k:{"mountPath":"/etc/pki/ca-trust/extracted/pem"}: + .: {} + f:mountPath: {} + f:name: {} + f:readOnly: {} + k:{"name":"oauth-proxy"}: + .: {} + f:args: {} + f:image: {} + f:imagePullPolicy: {} + f:name: {} + f:ports: + .: {} + k:{"containerPort":8443,"protocol":"TCP"}: + .: {} + f:containerPort: {} + f:name: {} + f:protocol: {} + f:resources: {} + f:terminationMessagePath: {} + f:terminationMessagePolicy: {} + f:volumeMounts: + .: {} + k:{"mountPath":"/etc/pki/ca-trust/extracted/pem"}: + .: {} + f:mountPath: {} + f:name: {} + f:readOnly: {} + k:{"mountPath":"/etc/tls/private"}: + .: {} + f:mountPath: {} + f:name: {} + f:dnsPolicy: {} + f:enableServiceLinks: {} + f:restartPolicy: {} + f:schedulerName: {} + f:securityContext: {} + f:serviceAccount: {} + f:serviceAccountName: {} + f:terminationGracePeriodSeconds: {} + f:volumes: + .: {} + k:{"name":"certs"}: + .: {} + f:name: {} + f:secret: + .: {} + f:defaultMode: {} + f:secretName: {} + k:{"name":"my-jaeger-trusted-ca"}: + .: {} + f:configMap: + .: {} + f:defaultMode: {} + f:items: {} + f:name: {} + f:name: {} + k:{"name":"my-jaeger-ui-configuration-volume"}: + .: {} + f:configMap: + .: {} + f:defaultMode: {} + f:items: {} + f:name: {} + f:name: {} + k:{"name":"my-jaeger-ui-oauth-proxy-tls"}: + .: {} + f:name: {} + f:secret: + .: {} + f:defaultMode: {} + f:secretName: {} + manager: jaeger-operator + operation: Update + time: "2024-08-29T19:49:48Z" + - apiVersion: apps/v1 + fieldsType: FieldsV1 + fieldsV1: + f:metadata: + f:annotations: + f:deployment.kubernetes.io/revision: {} + f:status: + f:conditions: + .: {} + k:{"type":"Available"}: + .: {} + f:lastTransitionTime: {} + f:lastUpdateTime: {} + f:message: {} + f:reason: {} + f:status: {} + f:type: {} + k:{"type":"Progressing"}: + .: {} + f:lastTransitionTime: {} + f:lastUpdateTime: {} + f:message: {} + f:reason: {} + f:status: {} + f:type: {} + f:observedGeneration: {} + f:replicas: {} + f:unavailableReplicas: {} + f:updatedReplicas: {} + manager: kube-controller-manager + operation: Update + subresource: status + time: "2024-08-29T19:49:48Z" name: my-jaeger-query namespace: kuttl-test-touching-ferret + ownerReferences: + - apiVersion: jaegertracing.io/v1 + controller: true + kind: Jaeger + name: my-jaeger + uid: b0ad0d0f-9296-44b9-bc68-f215a395844c spec: + progressDeadlineSeconds: 600 replicas: 1 + revisionHistoryLimit: 10 + selector: + matchLabels: + app: jaeger + app.kubernetes.io/component: query + app.kubernetes.io/instance: my-jaeger + app.kubernetes.io/managed-by: jaeger-operator + app.kubernetes.io/name: my-jaeger-query + app.kubernetes.io/part-of: jaeger + strategy: + type: Recreate + template: + metadata: + annotations: + linkerd.io/inject: disabled + prometheus.io/port: "16687" + prometheus.io/scrape: "true" + sidecar.istio.io/inject: "false" + sidecar.jaegertracing.io/inject: my-jaeger + creationTimestamp: null + labels: + app: jaeger + app.kubernetes.io/component: query + app.kubernetes.io/instance: my-jaeger + app.kubernetes.io/managed-by: jaeger-operator + app.kubernetes.io/name: my-jaeger-query + app.kubernetes.io/part-of: jaeger + spec: + containers: + - args: + - --query.ui-config=/etc/config/ui.json + - --es.server-urls=https://elasticsearch.kuttl-test-touching-ferret.svc.cluster.local:9200 + - --es.tls.enabled=true + - --es.tls.ca=/certs/ca + - --es.tls.cert=/certs/cert + - --es.tls.key=/certs/key + - --es.timeout=15s + - --es.num-shards=1 + - --es.num-replicas=0 + env: + - name: SPAN_STORAGE_TYPE + value: elasticsearch + - name: METRICS_STORAGE_TYPE + - name: JAEGER_DISABLED + value: "false" + - name: JAEGER_SERVICE_NAME + value: my-jaeger.kuttl-test-touching-ferret + - name: JAEGER_PROPAGATION + value: jaeger,b3,w3c + image: registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:831f83e8396d61822fb2dd576e711e0f10c1f8c7bd03bf0d4bec49dfb90cacca + imagePullPolicy: IfNotPresent + livenessProbe: + failureThreshold: 5 + httpGet: + path: / + port: 16687 + scheme: HTTP + initialDelaySeconds: 5 + periodSeconds: 15 + successThreshold: 1 + timeoutSeconds: 1 + name: jaeger-query + ports: + - containerPort: 16685 + name: grpc-query + protocol: TCP + - containerPort: 16686 + name: query + protocol: TCP + - containerPort: 16687 + name: admin-http + protocol: TCP + readinessProbe: + failureThreshold: 3 + httpGet: + path: / + port: 16687 + scheme: HTTP + initialDelaySeconds: 1 + periodSeconds: 10 + successThreshold: 1 + timeoutSeconds: 1 + resources: {} + terminationMessagePath: /dev/termination-log + terminationMessagePolicy: File + volumeMounts: + - mountPath: /etc/config + name: my-jaeger-ui-configuration-volume + readOnly: true + - mountPath: /etc/pki/ca-trust/extracted/pem + name: my-jaeger-trusted-ca + readOnly: true + - mountPath: /certs + name: certs + readOnly: true + - args: + - --cookie-secret=tnZGqNugDDg6ei3AbgNNiF + - --https-address=:8443 + - '--openshift-sar={"namespace": "kuttl-test-touching-ferret", "resource": + "pods", "verb": "get"}' + - --openshift-service-account=my-jaeger-ui-proxy + - --provider=openshift + - --tls-cert=/etc/tls/private/tls.crt + - --tls-key=/etc/tls/private/tls.key + - --upstream=http://localhost:16686 + env: + - name: JAEGER_SERVICE_NAME + value: my-jaeger.kuttl-test-touching-ferret + - name: JAEGER_PROPAGATION + value: jaeger,b3,w3c + image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1613032062a68faa414cd5754a666b7b78ad0e52cf328eaaed355719423a55dd + imagePullPolicy: IfNotPresent + name: oauth-proxy + ports: + - containerPort: 8443 + name: public + protocol: TCP + resources: {} + terminationMessagePath: /dev/termination-log + terminationMessagePolicy: File + volumeMounts: + - mountPath: /etc/tls/private + name: my-jaeger-ui-oauth-proxy-tls + - mountPath: /etc/pki/ca-trust/extracted/pem + name: my-jaeger-trusted-ca + readOnly: true + - args: + - --agent.tags=cluster=undefined,deployment.name=my-jaeger-query,host.ip=${HOST_IP:},pod.name=${POD_NAME:},pod.namespace=kuttl-test-touching-ferret + - --reporter.grpc.host-port=dns:///my-jaeger-collector-headless.kuttl-test-touching-ferret.svc:14250 + - --reporter.grpc.tls.ca=/etc/pki/ca-trust/source/service-ca/service-ca.crt + - --reporter.grpc.tls.enabled=true + env: + - name: POD_NAME + valueFrom: + fieldRef: + apiVersion: v1 + fieldPath: metadata.name + - name: HOST_IP + valueFrom: + fieldRef: + apiVersion: v1 + fieldPath: status.hostIP + image: registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:5fcc9d365678f54c0f52005318dbbc32b87f743cae7e1a84568e769653f3b128 + imagePullPolicy: IfNotPresent + livenessProbe: + failureThreshold: 5 + httpGet: + path: / + port: 14271 + scheme: HTTP + initialDelaySeconds: 5 + periodSeconds: 15 + successThreshold: 1 + timeoutSeconds: 1 + name: jaeger-agent + ports: + - containerPort: 5775 + name: zk-compact-trft + protocol: UDP + - containerPort: 5778 + name: config-rest + protocol: TCP + - containerPort: 6831 + name: jg-compact-trft + protocol: UDP + - containerPort: 6832 + name: jg-binary-trft + protocol: UDP + - containerPort: 14271 + name: admin-http + protocol: TCP + readinessProbe: + failureThreshold: 3 + httpGet: + path: / + port: 14271 + scheme: HTTP + initialDelaySeconds: 1 + periodSeconds: 10 + successThreshold: 1 + timeoutSeconds: 1 + resources: {} + terminationMessagePath: /dev/termination-log + terminationMessagePolicy: File + volumeMounts: + - mountPath: /etc/pki/ca-trust/extracted/pem + name: my-jaeger-trusted-ca + readOnly: true + - mountPath: /etc/pki/ca-trust/source/service-ca + name: my-jaeger-service-ca + readOnly: true + dnsPolicy: ClusterFirst + enableServiceLinks: false + restartPolicy: Always + schedulerName: default-scheduler + securityContext: {} + serviceAccount: my-jaeger-ui-proxy + serviceAccountName: my-jaeger-ui-proxy + terminationGracePeriodSeconds: 30 + volumes: + - configMap: + defaultMode: 420 + items: + - key: ui + path: ui.json + name: my-jaeger-ui-configuration + name: my-jaeger-ui-configuration-volume + - configMap: + defaultMode: 420 + items: + - key: ca-bundle.crt + path: tls-ca-bundle.pem + name: my-jaeger-trusted-ca + name: my-jaeger-trusted-ca + - name: my-jaeger-ui-oauth-proxy-tls + secret: + defaultMode: 420 + secretName: my-jaeger-ui-oauth-proxy-tls + - name: certs + secret: + defaultMode: 420 + secretName: my-jaeger-jaeger-elasticsearch + - configMap: + defaultMode: 420 + items: + - key: service-ca.crt + path: service-ca.crt + name: my-jaeger-service-ca + name: my-jaeger-service-ca status: - readyReplicas: 1 + conditions: + - lastTransitionTime: "2024-08-29T19:44:48Z" + lastUpdateTime: "2024-08-29T19:44:48Z" + message: Deployment does not have minimum availability. + reason: MinimumReplicasUnavailable + status: "False" + type: Available + - lastTransitionTime: "2024-08-29T19:44:47Z" + lastUpdateTime: "2024-08-29T19:44:48Z" + message: ReplicaSet "my-jaeger-query-7d4875d969" is progressing. + reason: ReplicaSetUpdated + status: "True" + type: Progressing + observedGeneration: 3 + replicas: 1 + unavailableReplicas: 1 + updatedReplicas: 1 case.go:366: resource Deployment:kuttl-test-touching-ferret/my-jaeger-query: .status.readyReplicas: key is missing from map logger.go:42: 19:54:43 | set-custom-img | set-custom-img events from ns kuttl-test-touching-ferret: logger.go:42: 19:54:43 | set-custom-img | 2024-08-29 19:44:47 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-58c4f44fd8 to 1 deployment-controller logger.go:42: 19:54:43 | set-custom-img | 2024-08-29 19:44:47 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-7d4875d969 to 1 deployment-controller logger.go:42: 19:54:43 | set-custom-img | 2024-08-29 19:44:48 +0000 UTC Normal Pod my-jaeger-collector-58c4f44fd8-vhfkq Binding Scheduled Successfully assigned kuttl-test-touching-ferret/my-jaeger-collector-58c4f44fd8-vhfkq to ip-10-0-51-237.us-west-2.compute.internal default-scheduler logger.go:42: 19:54:43 | set-custom-img | 2024-08-29 19:44:48 +0000 UTC Normal Pod my-jaeger-collector-58c4f44fd8-vhfkq AddedInterface Add eth0 [10.130.0.55/23] from ovn-kubernetes multus logger.go:42: 19:54:43 | set-custom-img | 2024-08-29 19:44:48 +0000 UTC Normal Pod my-jaeger-collector-58c4f44fd8-vhfkq.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:b90e0d264bb0b051c2a4b9e6d499f407fed58f834ab8ebf30263ee1cf7c7ea3c" already present on machine kubelet logger.go:42: 19:54:43 | set-custom-img | 2024-08-29 19:44:48 +0000 UTC Normal Pod my-jaeger-collector-58c4f44fd8-vhfkq.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 19:54:43 | set-custom-img | 2024-08-29 19:44:48 +0000 UTC Normal Pod my-jaeger-collector-58c4f44fd8-vhfkq.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 19:54:43 | set-custom-img | 2024-08-29 19:44:48 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-58c4f44fd8 SuccessfulCreate Created pod: my-jaeger-collector-58c4f44fd8-vhfkq replicaset-controller logger.go:42: 19:54:43 | set-custom-img | 2024-08-29 19:44:48 +0000 UTC Normal Pod my-jaeger-query-7d4875d969-4jzhz Binding Scheduled Successfully assigned kuttl-test-touching-ferret/my-jaeger-query-7d4875d969-4jzhz to ip-10-0-51-237.us-west-2.compute.internal default-scheduler logger.go:42: 19:54:43 | set-custom-img | 2024-08-29 19:44:48 +0000 UTC Normal Pod my-jaeger-query-7d4875d969-4jzhz AddedInterface Add eth0 [10.130.0.56/23] from ovn-kubernetes multus logger.go:42: 19:54:43 | set-custom-img | 2024-08-29 19:44:48 +0000 UTC Normal Pod my-jaeger-query-7d4875d969-4jzhz.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:831f83e8396d61822fb2dd576e711e0f10c1f8c7bd03bf0d4bec49dfb90cacca" already present on machine kubelet logger.go:42: 19:54:43 | set-custom-img | 2024-08-29 19:44:48 +0000 UTC Normal Pod my-jaeger-query-7d4875d969-4jzhz.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 19:54:43 | set-custom-img | 2024-08-29 19:44:48 +0000 UTC Normal Pod my-jaeger-query-7d4875d969-4jzhz.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 19:54:43 | set-custom-img | 2024-08-29 19:44:48 +0000 UTC Normal Pod my-jaeger-query-7d4875d969-4jzhz.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1613032062a68faa414cd5754a666b7b78ad0e52cf328eaaed355719423a55dd" already present on machine kubelet logger.go:42: 19:54:43 | set-custom-img | 2024-08-29 19:44:48 +0000 UTC Normal Pod my-jaeger-query-7d4875d969-4jzhz.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 19:54:43 | set-custom-img | 2024-08-29 19:44:48 +0000 UTC Normal Pod my-jaeger-query-7d4875d969-4jzhz.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 19:54:43 | set-custom-img | 2024-08-29 19:44:48 +0000 UTC Normal Pod my-jaeger-query-7d4875d969-4jzhz.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:5fcc9d365678f54c0f52005318dbbc32b87f743cae7e1a84568e769653f3b128" already present on machine kubelet logger.go:42: 19:54:43 | set-custom-img | 2024-08-29 19:44:48 +0000 UTC Normal Pod my-jaeger-query-7d4875d969-4jzhz.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 19:54:43 | set-custom-img | 2024-08-29 19:44:48 +0000 UTC Normal Pod my-jaeger-query-7d4875d969-4jzhz.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 19:54:43 | set-custom-img | 2024-08-29 19:44:48 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-7d4875d969 SuccessfulCreate Created pod: my-jaeger-query-7d4875d969-4jzhz replicaset-controller logger.go:42: 19:54:43 | set-custom-img | 2024-08-29 19:44:49 +0000 UTC Warning Pod my-jaeger-collector-58c4f44fd8-vhfkq.spec.containers{jaeger-collector} Unhealthy Readiness probe failed: HTTP probe failed with statuscode: 503 kubelet logger.go:42: 19:54:43 | set-custom-img | 2024-08-29 19:44:50 +0000 UTC Warning Pod my-jaeger-query-7d4875d969-4jzhz.spec.containers{jaeger-query} Unhealthy Readiness probe failed: HTTP probe failed with statuscode: 503 kubelet logger.go:42: 19:54:43 | set-custom-img | 2024-08-29 19:45:00 +0000 UTC Warning Pod my-jaeger-collector-58c4f44fd8-vhfkq.spec.containers{jaeger-collector} BackOff Back-off restarting failed container jaeger-collector in pod my-jaeger-collector-58c4f44fd8-vhfkq_kuttl-test-touching-ferret(5d0c5562-a010-4bf9-8fc6-8245e48e7b24) kubelet logger.go:42: 19:54:43 | set-custom-img | 2024-08-29 19:45:00 +0000 UTC Warning Pod my-jaeger-query-7d4875d969-4jzhz.spec.containers{jaeger-query} BackOff Back-off restarting failed container jaeger-query in pod my-jaeger-query-7d4875d969-4jzhz_kuttl-test-touching-ferret(417a4c2e-588c-41c3-ba29-d4b65681791a) kubelet logger.go:42: 19:54:43 | set-custom-img | 2024-08-29 19:45:03 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 19:54:43 | set-custom-img | 2024-08-29 19:45:03 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 19:54:43 | set-custom-img | 2024-08-29 19:45:03 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 19:54:43 | set-custom-img | Deleting namespace: kuttl-test-touching-ferret === CONT kuttl/harness/collector-otlp-production-http logger.go:42: 19:54:50 | collector-otlp-production-http | Creating namespace: kuttl-test-able-eagle logger.go:42: 19:54:50 | collector-otlp-production-http/1-install | starting test step 1-install logger.go:42: 19:54:50 | collector-otlp-production-http/1-install | Jaeger:kuttl-test-able-eagle/my-jaeger created logger.go:42: 20:04:50 | collector-otlp-production-http/1-install | test step failed 1-install case.go:364: failed in step 1-install case.go:366: --- Deployment:kuttl-test-able-eagle/my-jaeger-collector +++ Deployment:kuttl-test-able-eagle/my-jaeger-collector @@ -1,10 +1,412 @@ apiVersion: apps/v1 kind: Deployment metadata: + annotations: + linkerd.io/inject: disabled + prometheus.io/port: "14269" + prometheus.io/scrape: "true" + labels: + app: jaeger + app.kubernetes.io/component: collector + app.kubernetes.io/instance: my-jaeger + app.kubernetes.io/managed-by: jaeger-operator + app.kubernetes.io/name: my-jaeger-collector + app.kubernetes.io/part-of: jaeger + managedFields: + - apiVersion: apps/v1 + fieldsType: FieldsV1 + fieldsV1: + f:metadata: + f:annotations: + .: {} + f:linkerd.io/inject: {} + f:prometheus.io/port: {} + f:prometheus.io/scrape: {} + f:labels: + .: {} + f:app: {} + f:app.kubernetes.io/component: {} + f:app.kubernetes.io/instance: {} + f:app.kubernetes.io/managed-by: {} + f:app.kubernetes.io/name: {} + f:app.kubernetes.io/part-of: {} + f:ownerReferences: + .: {} + k:{"uid":"c2221b71-f809-4624-97f0-8875840796ad"}: {} + f:spec: + f:progressDeadlineSeconds: {} + f:replicas: {} + f:revisionHistoryLimit: {} + f:selector: {} + f:strategy: + f:type: {} + f:template: + f:metadata: + f:annotations: + .: {} + f:linkerd.io/inject: {} + f:prometheus.io/port: {} + f:prometheus.io/scrape: {} + f:sidecar.istio.io/inject: {} + f:labels: + .: {} + f:app: {} + f:app.kubernetes.io/component: {} + f:app.kubernetes.io/instance: {} + f:app.kubernetes.io/managed-by: {} + f:app.kubernetes.io/name: {} + f:app.kubernetes.io/part-of: {} + f:spec: + f:containers: + k:{"name":"jaeger-collector"}: + .: {} + f:args: {} + f:env: + .: {} + k:{"name":"COLLECTOR_OTLP_ENABLED"}: + .: {} + f:name: {} + f:value: {} + k:{"name":"COLLECTOR_ZIPKIN_HOST_PORT"}: + .: {} + f:name: {} + f:value: {} + k:{"name":"SPAN_STORAGE_TYPE"}: + .: {} + f:name: {} + f:value: {} + f:image: {} + f:imagePullPolicy: {} + f:livenessProbe: + .: {} + f:failureThreshold: {} + f:httpGet: + .: {} + f:path: {} + f:port: {} + f:scheme: {} + f:initialDelaySeconds: {} + f:periodSeconds: {} + f:successThreshold: {} + f:timeoutSeconds: {} + f:name: {} + f:ports: + .: {} + k:{"containerPort":4317,"protocol":"TCP"}: + .: {} + f:containerPort: {} + f:name: {} + f:protocol: {} + k:{"containerPort":4318,"protocol":"TCP"}: + .: {} + f:containerPort: {} + f:name: {} + f:protocol: {} + k:{"containerPort":9411,"protocol":"TCP"}: + .: {} + f:containerPort: {} + f:name: {} + f:protocol: {} + k:{"containerPort":14250,"protocol":"TCP"}: + .: {} + f:containerPort: {} + f:name: {} + f:protocol: {} + k:{"containerPort":14267,"protocol":"TCP"}: + .: {} + f:containerPort: {} + f:name: {} + f:protocol: {} + k:{"containerPort":14268,"protocol":"TCP"}: + .: {} + f:containerPort: {} + f:name: {} + f:protocol: {} + k:{"containerPort":14269,"protocol":"TCP"}: + .: {} + f:containerPort: {} + f:name: {} + f:protocol: {} + f:readinessProbe: + .: {} + f:failureThreshold: {} + f:httpGet: + .: {} + f:path: {} + f:port: {} + f:scheme: {} + f:initialDelaySeconds: {} + f:periodSeconds: {} + f:successThreshold: {} + f:timeoutSeconds: {} + f:resources: {} + f:terminationMessagePath: {} + f:terminationMessagePolicy: {} + f:volumeMounts: + .: {} + k:{"mountPath":"/certs"}: + .: {} + f:mountPath: {} + f:name: {} + f:readOnly: {} + k:{"mountPath":"/etc/jaeger/sampling"}: + .: {} + f:mountPath: {} + f:name: {} + f:readOnly: {} + k:{"mountPath":"/etc/pki/ca-trust/extracted/pem"}: + .: {} + f:mountPath: {} + f:name: {} + f:readOnly: {} + k:{"mountPath":"/etc/tls-config"}: + .: {} + f:mountPath: {} + f:name: {} + f:readOnly: {} + f:dnsPolicy: {} + f:enableServiceLinks: {} + f:restartPolicy: {} + f:schedulerName: {} + f:securityContext: {} + f:serviceAccount: {} + f:serviceAccountName: {} + f:terminationGracePeriodSeconds: {} + f:volumes: + .: {} + k:{"name":"certs"}: + .: {} + f:name: {} + f:secret: + .: {} + f:defaultMode: {} + f:secretName: {} + k:{"name":"my-jaeger-collector-tls-config-volume"}: + .: {} + f:name: {} + f:secret: + .: {} + f:defaultMode: {} + f:secretName: {} + k:{"name":"my-jaeger-sampling-configuration-volume"}: + .: {} + f:configMap: + .: {} + f:defaultMode: {} + f:items: {} + f:name: {} + f:name: {} + k:{"name":"my-jaeger-trusted-ca"}: + .: {} + f:configMap: + .: {} + f:defaultMode: {} + f:items: {} + f:name: {} + f:name: {} + manager: jaeger-operator + operation: Update + time: "2024-08-29T19:54:54Z" + - apiVersion: apps/v1 + fieldsType: FieldsV1 + fieldsV1: + f:metadata: + f:annotations: + f:deployment.kubernetes.io/revision: {} + f:status: + f:conditions: + .: {} + k:{"type":"Available"}: + .: {} + f:lastTransitionTime: {} + f:lastUpdateTime: {} + f:message: {} + f:reason: {} + f:status: {} + f:type: {} + k:{"type":"Progressing"}: + .: {} + f:lastTransitionTime: {} + f:lastUpdateTime: {} + f:message: {} + f:reason: {} + f:status: {} + f:type: {} + f:observedGeneration: {} + f:replicas: {} + f:unavailableReplicas: {} + f:updatedReplicas: {} + manager: kube-controller-manager + operation: Update + subresource: status + time: "2024-08-29T19:54:54Z" name: my-jaeger-collector namespace: kuttl-test-able-eagle + ownerReferences: + - apiVersion: jaegertracing.io/v1 + controller: true + kind: Jaeger + name: my-jaeger + uid: c2221b71-f809-4624-97f0-8875840796ad spec: + progressDeadlineSeconds: 600 replicas: 1 + revisionHistoryLimit: 10 + selector: + matchLabels: + app: jaeger + app.kubernetes.io/component: collector + app.kubernetes.io/instance: my-jaeger + app.kubernetes.io/managed-by: jaeger-operator + app.kubernetes.io/name: my-jaeger-collector + app.kubernetes.io/part-of: jaeger + strategy: + type: Recreate + template: + metadata: + annotations: + linkerd.io/inject: disabled + prometheus.io/port: "14269" + prometheus.io/scrape: "true" + sidecar.istio.io/inject: "false" + creationTimestamp: null + labels: + app: jaeger + app.kubernetes.io/component: collector + app.kubernetes.io/instance: my-jaeger + app.kubernetes.io/managed-by: jaeger-operator + app.kubernetes.io/name: my-jaeger-collector + app.kubernetes.io/part-of: jaeger + spec: + containers: + - args: + - --collector.grpc.tls.cert=/etc/tls-config/tls.crt + - --collector.grpc.tls.enabled=true + - --collector.grpc.tls.key=/etc/tls-config/tls.key + - --sampling.strategies-file=/etc/jaeger/sampling/sampling.json + - --es.server-urls=https://elasticsearch.kuttl-test-able-eagle.svc.cluster.local:9200 + - --es.tls.enabled=true + - --es.tls.ca=/certs/ca + - --es.tls.cert=/certs/cert + - --es.tls.key=/certs/key + - --es.timeout=15s + - --es.num-shards=1 + - --es.num-replicas=0 + env: + - name: SPAN_STORAGE_TYPE + value: elasticsearch + - name: COLLECTOR_ZIPKIN_HOST_PORT + value: :9411 + - name: COLLECTOR_OTLP_ENABLED + value: "true" + image: registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:b90e0d264bb0b051c2a4b9e6d499f407fed58f834ab8ebf30263ee1cf7c7ea3c + imagePullPolicy: IfNotPresent + livenessProbe: + failureThreshold: 5 + httpGet: + path: / + port: 14269 + scheme: HTTP + initialDelaySeconds: 5 + periodSeconds: 15 + successThreshold: 1 + timeoutSeconds: 1 + name: jaeger-collector + ports: + - containerPort: 9411 + name: zipkin + protocol: TCP + - containerPort: 14267 + name: c-tchan-trft + protocol: TCP + - containerPort: 14268 + name: c-binary-trft + protocol: TCP + - containerPort: 14269 + name: admin-http + protocol: TCP + - containerPort: 14250 + name: grpc + protocol: TCP + - containerPort: 4317 + name: grpc-otlp + protocol: TCP + - containerPort: 4318 + name: http-otlp + protocol: TCP + readinessProbe: + failureThreshold: 3 + httpGet: + path: / + port: 14269 + scheme: HTTP + initialDelaySeconds: 1 + periodSeconds: 10 + successThreshold: 1 + timeoutSeconds: 1 + resources: {} + terminationMessagePath: /dev/termination-log + terminationMessagePolicy: File + volumeMounts: + - mountPath: /etc/jaeger/sampling + name: my-jaeger-sampling-configuration-volume + readOnly: true + - mountPath: /etc/tls-config + name: my-jaeger-collector-tls-config-volume + readOnly: true + - mountPath: /etc/pki/ca-trust/extracted/pem + name: my-jaeger-trusted-ca + readOnly: true + - mountPath: /certs + name: certs + readOnly: true + dnsPolicy: ClusterFirst + enableServiceLinks: false + restartPolicy: Always + schedulerName: default-scheduler + securityContext: {} + serviceAccount: my-jaeger + serviceAccountName: my-jaeger + terminationGracePeriodSeconds: 30 + volumes: + - configMap: + defaultMode: 420 + items: + - key: sampling + path: sampling.json + name: my-jaeger-sampling-configuration + name: my-jaeger-sampling-configuration-volume + - name: my-jaeger-collector-tls-config-volume + secret: + defaultMode: 420 + secretName: my-jaeger-collector-headless-tls + - configMap: + defaultMode: 420 + items: + - key: ca-bundle.crt + path: tls-ca-bundle.pem + name: my-jaeger-trusted-ca + name: my-jaeger-trusted-ca + - name: certs + secret: + defaultMode: 420 + secretName: my-jaeger-jaeger-elasticsearch status: - readyReplicas: 1 + conditions: + - lastTransitionTime: "2024-08-29T19:54:54Z" + lastUpdateTime: "2024-08-29T19:54:54Z" + message: Deployment does not have minimum availability. + reason: MinimumReplicasUnavailable + status: "False" + type: Available + - lastTransitionTime: "2024-08-29T19:54:54Z" + lastUpdateTime: "2024-08-29T19:54:54Z" + message: ReplicaSet "my-jaeger-collector-5b46cd58bf" is progressing. + reason: ReplicaSetUpdated + status: "True" + type: Progressing + observedGeneration: 1 + replicas: 1 + unavailableReplicas: 1 + updatedReplicas: 1 case.go:366: resource Deployment:kuttl-test-able-eagle/my-jaeger-collector: .status.readyReplicas: key is missing from map case.go:366: --- Deployment:kuttl-test-able-eagle/my-jaeger-query +++ Deployment:kuttl-test-able-eagle/my-jaeger-query @@ -1,10 +1,508 @@ apiVersion: apps/v1 kind: Deployment metadata: + annotations: + linkerd.io/inject: disabled + prometheus.io/port: "16687" + prometheus.io/scrape: "true" + sidecar.jaegertracing.io/inject: my-jaeger + sidecar.jaegertracing.io/revision: "1" + labels: + app: jaeger + app.kubernetes.io/component: query + app.kubernetes.io/instance: my-jaeger + app.kubernetes.io/managed-by: jaeger-operator + app.kubernetes.io/name: my-jaeger-query + app.kubernetes.io/part-of: jaeger + sidecar.jaegertracing.io/injected: my-jaeger + managedFields: + - apiVersion: apps/v1 + fieldsType: FieldsV1 + fieldsV1: + f:metadata: + f:annotations: + .: {} + f:linkerd.io/inject: {} + f:prometheus.io/port: {} + f:prometheus.io/scrape: {} + f:sidecar.jaegertracing.io/inject: {} + f:sidecar.jaegertracing.io/revision: {} + f:labels: + .: {} + f:app: {} + f:app.kubernetes.io/component: {} + f:app.kubernetes.io/instance: {} + f:app.kubernetes.io/managed-by: {} + f:app.kubernetes.io/name: {} + f:app.kubernetes.io/part-of: {} + f:ownerReferences: + .: {} + k:{"uid":"c2221b71-f809-4624-97f0-8875840796ad"}: {} + f:spec: + f:progressDeadlineSeconds: {} + f:replicas: {} + f:revisionHistoryLimit: {} + f:selector: {} + f:strategy: + f:type: {} + f:template: + f:metadata: + f:annotations: + .: {} + f:linkerd.io/inject: {} + f:prometheus.io/port: {} + f:prometheus.io/scrape: {} + f:sidecar.istio.io/inject: {} + f:sidecar.jaegertracing.io/inject: {} + f:labels: + .: {} + f:app: {} + f:app.kubernetes.io/component: {} + f:app.kubernetes.io/instance: {} + f:app.kubernetes.io/managed-by: {} + f:app.kubernetes.io/name: {} + f:app.kubernetes.io/part-of: {} + f:spec: + f:containers: + k:{"name":"jaeger-query"}: + .: {} + f:args: {} + f:env: + .: {} + k:{"name":"JAEGER_DISABLED"}: + .: {} + f:name: {} + f:value: {} + k:{"name":"METRICS_STORAGE_TYPE"}: + .: {} + f:name: {} + k:{"name":"SPAN_STORAGE_TYPE"}: + .: {} + f:name: {} + f:value: {} + f:image: {} + f:imagePullPolicy: {} + f:livenessProbe: + .: {} + f:failureThreshold: {} + f:httpGet: + .: {} + f:path: {} + f:port: {} + f:scheme: {} + f:initialDelaySeconds: {} + f:periodSeconds: {} + f:successThreshold: {} + f:timeoutSeconds: {} + f:name: {} + f:ports: + .: {} + k:{"containerPort":16685,"protocol":"TCP"}: + .: {} + f:containerPort: {} + f:name: {} + f:protocol: {} + k:{"containerPort":16686,"protocol":"TCP"}: + .: {} + f:containerPort: {} + f:name: {} + f:protocol: {} + k:{"containerPort":16687,"protocol":"TCP"}: + .: {} + f:containerPort: {} + f:name: {} + f:protocol: {} + f:readinessProbe: + .: {} + f:failureThreshold: {} + f:httpGet: + .: {} + f:path: {} + f:port: {} + f:scheme: {} + f:initialDelaySeconds: {} + f:periodSeconds: {} + f:successThreshold: {} + f:timeoutSeconds: {} + f:resources: {} + f:terminationMessagePath: {} + f:terminationMessagePolicy: {} + f:volumeMounts: + .: {} + k:{"mountPath":"/certs"}: + .: {} + f:mountPath: {} + f:name: {} + f:readOnly: {} + k:{"mountPath":"/etc/config"}: + .: {} + f:mountPath: {} + f:name: {} + f:readOnly: {} + k:{"mountPath":"/etc/pki/ca-trust/extracted/pem"}: + .: {} + f:mountPath: {} + f:name: {} + f:readOnly: {} + k:{"name":"oauth-proxy"}: + .: {} + f:args: {} + f:image: {} + f:imagePullPolicy: {} + f:name: {} + f:ports: + .: {} + k:{"containerPort":8443,"protocol":"TCP"}: + .: {} + f:containerPort: {} + f:name: {} + f:protocol: {} + f:resources: {} + f:terminationMessagePath: {} + f:terminationMessagePolicy: {} + f:volumeMounts: + .: {} + k:{"mountPath":"/etc/pki/ca-trust/extracted/pem"}: + .: {} + f:mountPath: {} + f:name: {} + f:readOnly: {} + k:{"mountPath":"/etc/tls/private"}: + .: {} + f:mountPath: {} + f:name: {} + f:dnsPolicy: {} + f:enableServiceLinks: {} + f:restartPolicy: {} + f:schedulerName: {} + f:securityContext: {} + f:serviceAccount: {} + f:serviceAccountName: {} + f:terminationGracePeriodSeconds: {} + f:volumes: + .: {} + k:{"name":"certs"}: + .: {} + f:name: {} + f:secret: + .: {} + f:defaultMode: {} + f:secretName: {} + k:{"name":"my-jaeger-trusted-ca"}: + .: {} + f:configMap: + .: {} + f:defaultMode: {} + f:items: {} + f:name: {} + f:name: {} + k:{"name":"my-jaeger-ui-configuration-volume"}: + .: {} + f:configMap: + .: {} + f:defaultMode: {} + f:items: {} + f:name: {} + f:name: {} + k:{"name":"my-jaeger-ui-oauth-proxy-tls"}: + .: {} + f:name: {} + f:secret: + .: {} + f:defaultMode: {} + f:secretName: {} + manager: jaeger-operator + operation: Update + time: "2024-08-29T19:59:55Z" + - apiVersion: apps/v1 + fieldsType: FieldsV1 + fieldsV1: + f:metadata: + f:annotations: + f:deployment.kubernetes.io/revision: {} + f:status: + f:conditions: + .: {} + k:{"type":"Available"}: + .: {} + f:lastTransitionTime: {} + f:lastUpdateTime: {} + f:message: {} + f:reason: {} + f:status: {} + f:type: {} + k:{"type":"Progressing"}: + .: {} + f:lastTransitionTime: {} + f:lastUpdateTime: {} + f:message: {} + f:reason: {} + f:status: {} + f:type: {} + f:observedGeneration: {} + f:replicas: {} + f:unavailableReplicas: {} + f:updatedReplicas: {} + manager: kube-controller-manager + operation: Update + subresource: status + time: "2024-08-29T19:59:55Z" name: my-jaeger-query namespace: kuttl-test-able-eagle + ownerReferences: + - apiVersion: jaegertracing.io/v1 + controller: true + kind: Jaeger + name: my-jaeger + uid: c2221b71-f809-4624-97f0-8875840796ad spec: + progressDeadlineSeconds: 600 replicas: 1 + revisionHistoryLimit: 10 + selector: + matchLabels: + app: jaeger + app.kubernetes.io/component: query + app.kubernetes.io/instance: my-jaeger + app.kubernetes.io/managed-by: jaeger-operator + app.kubernetes.io/name: my-jaeger-query + app.kubernetes.io/part-of: jaeger + strategy: + type: Recreate + template: + metadata: + annotations: + linkerd.io/inject: disabled + prometheus.io/port: "16687" + prometheus.io/scrape: "true" + sidecar.istio.io/inject: "false" + sidecar.jaegertracing.io/inject: my-jaeger + creationTimestamp: null + labels: + app: jaeger + app.kubernetes.io/component: query + app.kubernetes.io/instance: my-jaeger + app.kubernetes.io/managed-by: jaeger-operator + app.kubernetes.io/name: my-jaeger-query + app.kubernetes.io/part-of: jaeger + spec: + containers: + - args: + - --query.ui-config=/etc/config/ui.json + - --es.server-urls=https://elasticsearch.kuttl-test-able-eagle.svc.cluster.local:9200 + - --es.tls.enabled=true + - --es.tls.ca=/certs/ca + - --es.tls.cert=/certs/cert + - --es.tls.key=/certs/key + - --es.timeout=15s + - --es.num-shards=1 + - --es.num-replicas=0 + env: + - name: SPAN_STORAGE_TYPE + value: elasticsearch + - name: METRICS_STORAGE_TYPE + - name: JAEGER_DISABLED + value: "false" + - name: JAEGER_SERVICE_NAME + value: my-jaeger.kuttl-test-able-eagle + - name: JAEGER_PROPAGATION + value: jaeger,b3,w3c + image: registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:831f83e8396d61822fb2dd576e711e0f10c1f8c7bd03bf0d4bec49dfb90cacca + imagePullPolicy: IfNotPresent + livenessProbe: + failureThreshold: 5 + httpGet: + path: / + port: 16687 + scheme: HTTP + initialDelaySeconds: 5 + periodSeconds: 15 + successThreshold: 1 + timeoutSeconds: 1 + name: jaeger-query + ports: + - containerPort: 16685 + name: grpc-query + protocol: TCP + - containerPort: 16686 + name: query + protocol: TCP + - containerPort: 16687 + name: admin-http + protocol: TCP + readinessProbe: + failureThreshold: 3 + httpGet: + path: / + port: 16687 + scheme: HTTP + initialDelaySeconds: 1 + periodSeconds: 10 + successThreshold: 1 + timeoutSeconds: 1 + resources: {} + terminationMessagePath: /dev/termination-log + terminationMessagePolicy: File + volumeMounts: + - mountPath: /etc/config + name: my-jaeger-ui-configuration-volume + readOnly: true + - mountPath: /etc/pki/ca-trust/extracted/pem + name: my-jaeger-trusted-ca + readOnly: true + - mountPath: /certs + name: certs + readOnly: true + - args: + - --cookie-secret=kOdmrAz76mti89TboyzDY9 + - --https-address=:8443 + - '--openshift-sar={"namespace": "kuttl-test-able-eagle", "resource": "pods", + "verb": "get"}' + - --openshift-service-account=my-jaeger-ui-proxy + - --provider=openshift + - --tls-cert=/etc/tls/private/tls.crt + - --tls-key=/etc/tls/private/tls.key + - --upstream=http://localhost:16686 + env: + - name: JAEGER_SERVICE_NAME + value: my-jaeger.kuttl-test-able-eagle + - name: JAEGER_PROPAGATION + value: jaeger,b3,w3c + image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1613032062a68faa414cd5754a666b7b78ad0e52cf328eaaed355719423a55dd + imagePullPolicy: IfNotPresent + name: oauth-proxy + ports: + - containerPort: 8443 + name: public + protocol: TCP + resources: {} + terminationMessagePath: /dev/termination-log + terminationMessagePolicy: File + volumeMounts: + - mountPath: /etc/tls/private + name: my-jaeger-ui-oauth-proxy-tls + - mountPath: /etc/pki/ca-trust/extracted/pem + name: my-jaeger-trusted-ca + readOnly: true + - args: + - --agent.tags=cluster=undefined,deployment.name=my-jaeger-query,host.ip=${HOST_IP:},pod.name=${POD_NAME:},pod.namespace=kuttl-test-able-eagle + - --reporter.grpc.host-port=dns:///my-jaeger-collector-headless.kuttl-test-able-eagle.svc:14250 + - --reporter.grpc.tls.ca=/etc/pki/ca-trust/source/service-ca/service-ca.crt + - --reporter.grpc.tls.enabled=true + env: + - name: POD_NAME + valueFrom: + fieldRef: + apiVersion: v1 + fieldPath: metadata.name + - name: HOST_IP + valueFrom: + fieldRef: + apiVersion: v1 + fieldPath: status.hostIP + image: registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:5fcc9d365678f54c0f52005318dbbc32b87f743cae7e1a84568e769653f3b128 + imagePullPolicy: IfNotPresent + livenessProbe: + failureThreshold: 5 + httpGet: + path: / + port: 14271 + scheme: HTTP + initialDelaySeconds: 5 + periodSeconds: 15 + successThreshold: 1 + timeoutSeconds: 1 + name: jaeger-agent + ports: + - containerPort: 5775 + name: zk-compact-trft + protocol: UDP + - containerPort: 5778 + name: config-rest + protocol: TCP + - containerPort: 6831 + name: jg-compact-trft + protocol: UDP + - containerPort: 6832 + name: jg-binary-trft + protocol: UDP + - containerPort: 14271 + name: admin-http + protocol: TCP + readinessProbe: + failureThreshold: 3 + httpGet: + path: / + port: 14271 + scheme: HTTP + initialDelaySeconds: 1 + periodSeconds: 10 + successThreshold: 1 + timeoutSeconds: 1 + resources: {} + terminationMessagePath: /dev/termination-log + terminationMessagePolicy: File + volumeMounts: + - mountPath: /etc/pki/ca-trust/extracted/pem + name: my-jaeger-trusted-ca + readOnly: true + - mountPath: /etc/pki/ca-trust/source/service-ca + name: my-jaeger-service-ca + readOnly: true + dnsPolicy: ClusterFirst + enableServiceLinks: false + restartPolicy: Always + schedulerName: default-scheduler + securityContext: {} + serviceAccount: my-jaeger-ui-proxy + serviceAccountName: my-jaeger-ui-proxy + terminationGracePeriodSeconds: 30 + volumes: + - configMap: + defaultMode: 420 + items: + - key: ui + path: ui.json + name: my-jaeger-ui-configuration + name: my-jaeger-ui-configuration-volume + - configMap: + defaultMode: 420 + items: + - key: ca-bundle.crt + path: tls-ca-bundle.pem + name: my-jaeger-trusted-ca + name: my-jaeger-trusted-ca + - name: my-jaeger-ui-oauth-proxy-tls + secret: + defaultMode: 420 + secretName: my-jaeger-ui-oauth-proxy-tls + - name: certs + secret: + defaultMode: 420 + secretName: my-jaeger-jaeger-elasticsearch + - configMap: + defaultMode: 420 + items: + - key: service-ca.crt + path: service-ca.crt + name: my-jaeger-service-ca + name: my-jaeger-service-ca status: - readyReplicas: 1 + conditions: + - lastTransitionTime: "2024-08-29T19:54:54Z" + lastUpdateTime: "2024-08-29T19:54:54Z" + message: Deployment does not have minimum availability. + reason: MinimumReplicasUnavailable + status: "False" + type: Available + - lastTransitionTime: "2024-08-29T19:54:54Z" + lastUpdateTime: "2024-08-29T19:54:54Z" + message: ReplicaSet "my-jaeger-query-67cd86b95b" is progressing. + reason: ReplicaSetUpdated + status: "True" + type: Progressing + observedGeneration: 3 + replicas: 1 + unavailableReplicas: 1 + updatedReplicas: 1 case.go:366: resource Deployment:kuttl-test-able-eagle/my-jaeger-query: .status.readyReplicas: key is missing from map logger.go:42: 20:04:50 | collector-otlp-production-http | collector-otlp-production-http events from ns kuttl-test-able-eagle: logger.go:42: 20:04:50 | collector-otlp-production-http | 2024-08-29 19:54:54 +0000 UTC Normal Pod my-jaeger-collector-5b46cd58bf-m26tf Binding Scheduled Successfully assigned kuttl-test-able-eagle/my-jaeger-collector-5b46cd58bf-m26tf to ip-10-0-51-237.us-west-2.compute.internal default-scheduler logger.go:42: 20:04:50 | collector-otlp-production-http | 2024-08-29 19:54:54 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-5b46cd58bf SuccessfulCreate Created pod: my-jaeger-collector-5b46cd58bf-m26tf replicaset-controller logger.go:42: 20:04:50 | collector-otlp-production-http | 2024-08-29 19:54:54 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-5b46cd58bf to 1 deployment-controller logger.go:42: 20:04:50 | collector-otlp-production-http | 2024-08-29 19:54:54 +0000 UTC Normal Pod my-jaeger-query-67cd86b95b-hbdpc Binding Scheduled Successfully assigned kuttl-test-able-eagle/my-jaeger-query-67cd86b95b-hbdpc to ip-10-0-51-237.us-west-2.compute.internal default-scheduler logger.go:42: 20:04:50 | collector-otlp-production-http | 2024-08-29 19:54:54 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-67cd86b95b SuccessfulCreate Created pod: my-jaeger-query-67cd86b95b-hbdpc replicaset-controller logger.go:42: 20:04:50 | collector-otlp-production-http | 2024-08-29 19:54:54 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-67cd86b95b to 1 deployment-controller logger.go:42: 20:04:50 | collector-otlp-production-http | 2024-08-29 19:54:55 +0000 UTC Normal Pod my-jaeger-collector-5b46cd58bf-m26tf AddedInterface Add eth0 [10.130.0.57/23] from ovn-kubernetes multus logger.go:42: 20:04:50 | collector-otlp-production-http | 2024-08-29 19:54:55 +0000 UTC Normal Pod my-jaeger-collector-5b46cd58bf-m26tf.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:b90e0d264bb0b051c2a4b9e6d499f407fed58f834ab8ebf30263ee1cf7c7ea3c" already present on machine kubelet logger.go:42: 20:04:50 | collector-otlp-production-http | 2024-08-29 19:54:55 +0000 UTC Normal Pod my-jaeger-collector-5b46cd58bf-m26tf.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 20:04:50 | collector-otlp-production-http | 2024-08-29 19:54:55 +0000 UTC Normal Pod my-jaeger-collector-5b46cd58bf-m26tf.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 20:04:50 | collector-otlp-production-http | 2024-08-29 19:54:55 +0000 UTC Normal Pod my-jaeger-query-67cd86b95b-hbdpc AddedInterface Add eth0 [10.130.0.58/23] from ovn-kubernetes multus logger.go:42: 20:04:50 | collector-otlp-production-http | 2024-08-29 19:54:55 +0000 UTC Normal Pod my-jaeger-query-67cd86b95b-hbdpc.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:831f83e8396d61822fb2dd576e711e0f10c1f8c7bd03bf0d4bec49dfb90cacca" already present on machine kubelet logger.go:42: 20:04:50 | collector-otlp-production-http | 2024-08-29 19:54:55 +0000 UTC Normal Pod my-jaeger-query-67cd86b95b-hbdpc.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 20:04:50 | collector-otlp-production-http | 2024-08-29 19:54:55 +0000 UTC Normal Pod my-jaeger-query-67cd86b95b-hbdpc.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 20:04:50 | collector-otlp-production-http | 2024-08-29 19:54:55 +0000 UTC Normal Pod my-jaeger-query-67cd86b95b-hbdpc.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1613032062a68faa414cd5754a666b7b78ad0e52cf328eaaed355719423a55dd" already present on machine kubelet logger.go:42: 20:04:50 | collector-otlp-production-http | 2024-08-29 19:54:55 +0000 UTC Normal Pod my-jaeger-query-67cd86b95b-hbdpc.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 20:04:50 | collector-otlp-production-http | 2024-08-29 19:54:55 +0000 UTC Normal Pod my-jaeger-query-67cd86b95b-hbdpc.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 20:04:50 | collector-otlp-production-http | 2024-08-29 19:54:55 +0000 UTC Normal Pod my-jaeger-query-67cd86b95b-hbdpc.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:5fcc9d365678f54c0f52005318dbbc32b87f743cae7e1a84568e769653f3b128" already present on machine kubelet logger.go:42: 20:04:50 | collector-otlp-production-http | 2024-08-29 19:54:55 +0000 UTC Normal Pod my-jaeger-query-67cd86b95b-hbdpc.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 20:04:50 | collector-otlp-production-http | 2024-08-29 19:54:55 +0000 UTC Normal Pod my-jaeger-query-67cd86b95b-hbdpc.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 20:04:50 | collector-otlp-production-http | 2024-08-29 19:54:56 +0000 UTC Warning Pod my-jaeger-query-67cd86b95b-hbdpc.spec.containers{jaeger-query} Unhealthy Readiness probe failed: HTTP probe failed with statuscode: 503 kubelet logger.go:42: 20:04:50 | collector-otlp-production-http | 2024-08-29 19:54:57 +0000 UTC Warning Pod my-jaeger-collector-5b46cd58bf-m26tf.spec.containers{jaeger-collector} Unhealthy Readiness probe failed: HTTP probe failed with statuscode: 503 kubelet logger.go:42: 20:04:50 | collector-otlp-production-http | 2024-08-29 19:55:07 +0000 UTC Warning Pod my-jaeger-collector-5b46cd58bf-m26tf.spec.containers{jaeger-collector} BackOff Back-off restarting failed container jaeger-collector in pod my-jaeger-collector-5b46cd58bf-m26tf_kuttl-test-able-eagle(c8096373-50a5-4616-95b6-d45ceb29514c) kubelet logger.go:42: 20:04:50 | collector-otlp-production-http | 2024-08-29 19:55:07 +0000 UTC Warning Pod my-jaeger-query-67cd86b95b-hbdpc.spec.containers{jaeger-query} BackOff Back-off restarting failed container jaeger-query in pod my-jaeger-query-67cd86b95b-hbdpc_kuttl-test-able-eagle(579c3bdb-4c42-435e-aeb8-5f9a4d22dc81) kubelet logger.go:42: 20:04:50 | collector-otlp-production-http | 2024-08-29 19:55:09 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 20:04:50 | collector-otlp-production-http | 2024-08-29 19:55:09 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 20:04:50 | collector-otlp-production-http | 2024-08-29 19:55:09 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 20:04:50 | collector-otlp-production-http | Deleting namespace: kuttl-test-able-eagle === CONT kuttl/harness/collector-otlp-production-grpc logger.go:42: 20:04:57 | collector-otlp-production-grpc | Creating namespace: kuttl-test-ethical-gibbon logger.go:42: 20:04:57 | collector-otlp-production-grpc/1-install | starting test step 1-install logger.go:42: 20:04:57 | collector-otlp-production-grpc/1-install | Jaeger:kuttl-test-ethical-gibbon/my-jaeger created logger.go:42: 20:14:58 | collector-otlp-production-grpc/1-install | test step failed 1-install case.go:364: failed in step 1-install case.go:366: --- Deployment:kuttl-test-ethical-gibbon/my-jaeger-collector +++ Deployment:kuttl-test-ethical-gibbon/my-jaeger-collector @@ -1,10 +1,412 @@ apiVersion: apps/v1 kind: Deployment metadata: + annotations: + linkerd.io/inject: disabled + prometheus.io/port: "14269" + prometheus.io/scrape: "true" + labels: + app: jaeger + app.kubernetes.io/component: collector + app.kubernetes.io/instance: my-jaeger + app.kubernetes.io/managed-by: jaeger-operator + app.kubernetes.io/name: my-jaeger-collector + app.kubernetes.io/part-of: jaeger + managedFields: + - apiVersion: apps/v1 + fieldsType: FieldsV1 + fieldsV1: + f:metadata: + f:annotations: + .: {} + f:linkerd.io/inject: {} + f:prometheus.io/port: {} + f:prometheus.io/scrape: {} + f:labels: + .: {} + f:app: {} + f:app.kubernetes.io/component: {} + f:app.kubernetes.io/instance: {} + f:app.kubernetes.io/managed-by: {} + f:app.kubernetes.io/name: {} + f:app.kubernetes.io/part-of: {} + f:ownerReferences: + .: {} + k:{"uid":"6a92244e-a496-4a93-9129-85917b6dd6db"}: {} + f:spec: + f:progressDeadlineSeconds: {} + f:replicas: {} + f:revisionHistoryLimit: {} + f:selector: {} + f:strategy: + f:type: {} + f:template: + f:metadata: + f:annotations: + .: {} + f:linkerd.io/inject: {} + f:prometheus.io/port: {} + f:prometheus.io/scrape: {} + f:sidecar.istio.io/inject: {} + f:labels: + .: {} + f:app: {} + f:app.kubernetes.io/component: {} + f:app.kubernetes.io/instance: {} + f:app.kubernetes.io/managed-by: {} + f:app.kubernetes.io/name: {} + f:app.kubernetes.io/part-of: {} + f:spec: + f:containers: + k:{"name":"jaeger-collector"}: + .: {} + f:args: {} + f:env: + .: {} + k:{"name":"COLLECTOR_OTLP_ENABLED"}: + .: {} + f:name: {} + f:value: {} + k:{"name":"COLLECTOR_ZIPKIN_HOST_PORT"}: + .: {} + f:name: {} + f:value: {} + k:{"name":"SPAN_STORAGE_TYPE"}: + .: {} + f:name: {} + f:value: {} + f:image: {} + f:imagePullPolicy: {} + f:livenessProbe: + .: {} + f:failureThreshold: {} + f:httpGet: + .: {} + f:path: {} + f:port: {} + f:scheme: {} + f:initialDelaySeconds: {} + f:periodSeconds: {} + f:successThreshold: {} + f:timeoutSeconds: {} + f:name: {} + f:ports: + .: {} + k:{"containerPort":4317,"protocol":"TCP"}: + .: {} + f:containerPort: {} + f:name: {} + f:protocol: {} + k:{"containerPort":4318,"protocol":"TCP"}: + .: {} + f:containerPort: {} + f:name: {} + f:protocol: {} + k:{"containerPort":9411,"protocol":"TCP"}: + .: {} + f:containerPort: {} + f:name: {} + f:protocol: {} + k:{"containerPort":14250,"protocol":"TCP"}: + .: {} + f:containerPort: {} + f:name: {} + f:protocol: {} + k:{"containerPort":14267,"protocol":"TCP"}: + .: {} + f:containerPort: {} + f:name: {} + f:protocol: {} + k:{"containerPort":14268,"protocol":"TCP"}: + .: {} + f:containerPort: {} + f:name: {} + f:protocol: {} + k:{"containerPort":14269,"protocol":"TCP"}: + .: {} + f:containerPort: {} + f:name: {} + f:protocol: {} + f:readinessProbe: + .: {} + f:failureThreshold: {} + f:httpGet: + .: {} + f:path: {} + f:port: {} + f:scheme: {} + f:initialDelaySeconds: {} + f:periodSeconds: {} + f:successThreshold: {} + f:timeoutSeconds: {} + f:resources: {} + f:terminationMessagePath: {} + f:terminationMessagePolicy: {} + f:volumeMounts: + .: {} + k:{"mountPath":"/certs"}: + .: {} + f:mountPath: {} + f:name: {} + f:readOnly: {} + k:{"mountPath":"/etc/jaeger/sampling"}: + .: {} + f:mountPath: {} + f:name: {} + f:readOnly: {} + k:{"mountPath":"/etc/pki/ca-trust/extracted/pem"}: + .: {} + f:mountPath: {} + f:name: {} + f:readOnly: {} + k:{"mountPath":"/etc/tls-config"}: + .: {} + f:mountPath: {} + f:name: {} + f:readOnly: {} + f:dnsPolicy: {} + f:enableServiceLinks: {} + f:restartPolicy: {} + f:schedulerName: {} + f:securityContext: {} + f:serviceAccount: {} + f:serviceAccountName: {} + f:terminationGracePeriodSeconds: {} + f:volumes: + .: {} + k:{"name":"certs"}: + .: {} + f:name: {} + f:secret: + .: {} + f:defaultMode: {} + f:secretName: {} + k:{"name":"my-jaeger-collector-tls-config-volume"}: + .: {} + f:name: {} + f:secret: + .: {} + f:defaultMode: {} + f:secretName: {} + k:{"name":"my-jaeger-sampling-configuration-volume"}: + .: {} + f:configMap: + .: {} + f:defaultMode: {} + f:items: {} + f:name: {} + f:name: {} + k:{"name":"my-jaeger-trusted-ca"}: + .: {} + f:configMap: + .: {} + f:defaultMode: {} + f:items: {} + f:name: {} + f:name: {} + manager: jaeger-operator + operation: Update + time: "2024-08-29T20:05:01Z" + - apiVersion: apps/v1 + fieldsType: FieldsV1 + fieldsV1: + f:metadata: + f:annotations: + f:deployment.kubernetes.io/revision: {} + f:status: + f:conditions: + .: {} + k:{"type":"Available"}: + .: {} + f:lastTransitionTime: {} + f:lastUpdateTime: {} + f:message: {} + f:reason: {} + f:status: {} + f:type: {} + k:{"type":"Progressing"}: + .: {} + f:lastTransitionTime: {} + f:lastUpdateTime: {} + f:message: {} + f:reason: {} + f:status: {} + f:type: {} + f:observedGeneration: {} + f:replicas: {} + f:unavailableReplicas: {} + f:updatedReplicas: {} + manager: kube-controller-manager + operation: Update + subresource: status + time: "2024-08-29T20:05:01Z" name: my-jaeger-collector namespace: kuttl-test-ethical-gibbon + ownerReferences: + - apiVersion: jaegertracing.io/v1 + controller: true + kind: Jaeger + name: my-jaeger + uid: 6a92244e-a496-4a93-9129-85917b6dd6db spec: + progressDeadlineSeconds: 600 replicas: 1 + revisionHistoryLimit: 10 + selector: + matchLabels: + app: jaeger + app.kubernetes.io/component: collector + app.kubernetes.io/instance: my-jaeger + app.kubernetes.io/managed-by: jaeger-operator + app.kubernetes.io/name: my-jaeger-collector + app.kubernetes.io/part-of: jaeger + strategy: + type: Recreate + template: + metadata: + annotations: + linkerd.io/inject: disabled + prometheus.io/port: "14269" + prometheus.io/scrape: "true" + sidecar.istio.io/inject: "false" + creationTimestamp: null + labels: + app: jaeger + app.kubernetes.io/component: collector + app.kubernetes.io/instance: my-jaeger + app.kubernetes.io/managed-by: jaeger-operator + app.kubernetes.io/name: my-jaeger-collector + app.kubernetes.io/part-of: jaeger + spec: + containers: + - args: + - --collector.grpc.tls.cert=/etc/tls-config/tls.crt + - --collector.grpc.tls.enabled=true + - --collector.grpc.tls.key=/etc/tls-config/tls.key + - --sampling.strategies-file=/etc/jaeger/sampling/sampling.json + - --es.server-urls=https://elasticsearch.kuttl-test-ethical-gibbon.svc.cluster.local:9200 + - --es.tls.enabled=true + - --es.tls.ca=/certs/ca + - --es.tls.cert=/certs/cert + - --es.tls.key=/certs/key + - --es.timeout=15s + - --es.num-shards=1 + - --es.num-replicas=0 + env: + - name: SPAN_STORAGE_TYPE + value: elasticsearch + - name: COLLECTOR_ZIPKIN_HOST_PORT + value: :9411 + - name: COLLECTOR_OTLP_ENABLED + value: "true" + image: registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:b90e0d264bb0b051c2a4b9e6d499f407fed58f834ab8ebf30263ee1cf7c7ea3c + imagePullPolicy: IfNotPresent + livenessProbe: + failureThreshold: 5 + httpGet: + path: / + port: 14269 + scheme: HTTP + initialDelaySeconds: 5 + periodSeconds: 15 + successThreshold: 1 + timeoutSeconds: 1 + name: jaeger-collector + ports: + - containerPort: 9411 + name: zipkin + protocol: TCP + - containerPort: 14267 + name: c-tchan-trft + protocol: TCP + - containerPort: 14268 + name: c-binary-trft + protocol: TCP + - containerPort: 14269 + name: admin-http + protocol: TCP + - containerPort: 14250 + name: grpc + protocol: TCP + - containerPort: 4317 + name: grpc-otlp + protocol: TCP + - containerPort: 4318 + name: http-otlp + protocol: TCP + readinessProbe: + failureThreshold: 3 + httpGet: + path: / + port: 14269 + scheme: HTTP + initialDelaySeconds: 1 + periodSeconds: 10 + successThreshold: 1 + timeoutSeconds: 1 + resources: {} + terminationMessagePath: /dev/termination-log + terminationMessagePolicy: File + volumeMounts: + - mountPath: /etc/jaeger/sampling + name: my-jaeger-sampling-configuration-volume + readOnly: true + - mountPath: /etc/tls-config + name: my-jaeger-collector-tls-config-volume + readOnly: true + - mountPath: /etc/pki/ca-trust/extracted/pem + name: my-jaeger-trusted-ca + readOnly: true + - mountPath: /certs + name: certs + readOnly: true + dnsPolicy: ClusterFirst + enableServiceLinks: false + restartPolicy: Always + schedulerName: default-scheduler + securityContext: {} + serviceAccount: my-jaeger + serviceAccountName: my-jaeger + terminationGracePeriodSeconds: 30 + volumes: + - configMap: + defaultMode: 420 + items: + - key: sampling + path: sampling.json + name: my-jaeger-sampling-configuration + name: my-jaeger-sampling-configuration-volume + - name: my-jaeger-collector-tls-config-volume + secret: + defaultMode: 420 + secretName: my-jaeger-collector-headless-tls + - configMap: + defaultMode: 420 + items: + - key: ca-bundle.crt + path: tls-ca-bundle.pem + name: my-jaeger-trusted-ca + name: my-jaeger-trusted-ca + - name: certs + secret: + defaultMode: 420 + secretName: my-jaeger-jaeger-elasticsearch status: - readyReplicas: 1 + conditions: + - lastTransitionTime: "2024-08-29T20:05:01Z" + lastUpdateTime: "2024-08-29T20:05:01Z" + message: Deployment does not have minimum availability. + reason: MinimumReplicasUnavailable + status: "False" + type: Available + - lastTransitionTime: "2024-08-29T20:05:01Z" + lastUpdateTime: "2024-08-29T20:05:01Z" + message: ReplicaSet "my-jaeger-collector-859bd6f798" is progressing. + reason: ReplicaSetUpdated + status: "True" + type: Progressing + observedGeneration: 1 + replicas: 1 + unavailableReplicas: 1 + updatedReplicas: 1 case.go:366: resource Deployment:kuttl-test-ethical-gibbon/my-jaeger-collector: .status.readyReplicas: key is missing from map case.go:366: --- Deployment:kuttl-test-ethical-gibbon/my-jaeger-query +++ Deployment:kuttl-test-ethical-gibbon/my-jaeger-query @@ -1,10 +1,508 @@ apiVersion: apps/v1 kind: Deployment metadata: + annotations: + linkerd.io/inject: disabled + prometheus.io/port: "16687" + prometheus.io/scrape: "true" + sidecar.jaegertracing.io/inject: my-jaeger + sidecar.jaegertracing.io/revision: "1" + labels: + app: jaeger + app.kubernetes.io/component: query + app.kubernetes.io/instance: my-jaeger + app.kubernetes.io/managed-by: jaeger-operator + app.kubernetes.io/name: my-jaeger-query + app.kubernetes.io/part-of: jaeger + sidecar.jaegertracing.io/injected: my-jaeger + managedFields: + - apiVersion: apps/v1 + fieldsType: FieldsV1 + fieldsV1: + f:metadata: + f:annotations: + .: {} + f:linkerd.io/inject: {} + f:prometheus.io/port: {} + f:prometheus.io/scrape: {} + f:sidecar.jaegertracing.io/inject: {} + f:sidecar.jaegertracing.io/revision: {} + f:labels: + .: {} + f:app: {} + f:app.kubernetes.io/component: {} + f:app.kubernetes.io/instance: {} + f:app.kubernetes.io/managed-by: {} + f:app.kubernetes.io/name: {} + f:app.kubernetes.io/part-of: {} + f:ownerReferences: + .: {} + k:{"uid":"6a92244e-a496-4a93-9129-85917b6dd6db"}: {} + f:spec: + f:progressDeadlineSeconds: {} + f:replicas: {} + f:revisionHistoryLimit: {} + f:selector: {} + f:strategy: + f:type: {} + f:template: + f:metadata: + f:annotations: + .: {} + f:linkerd.io/inject: {} + f:prometheus.io/port: {} + f:prometheus.io/scrape: {} + f:sidecar.istio.io/inject: {} + f:sidecar.jaegertracing.io/inject: {} + f:labels: + .: {} + f:app: {} + f:app.kubernetes.io/component: {} + f:app.kubernetes.io/instance: {} + f:app.kubernetes.io/managed-by: {} + f:app.kubernetes.io/name: {} + f:app.kubernetes.io/part-of: {} + f:spec: + f:containers: + k:{"name":"jaeger-query"}: + .: {} + f:args: {} + f:env: + .: {} + k:{"name":"JAEGER_DISABLED"}: + .: {} + f:name: {} + f:value: {} + k:{"name":"METRICS_STORAGE_TYPE"}: + .: {} + f:name: {} + k:{"name":"SPAN_STORAGE_TYPE"}: + .: {} + f:name: {} + f:value: {} + f:image: {} + f:imagePullPolicy: {} + f:livenessProbe: + .: {} + f:failureThreshold: {} + f:httpGet: + .: {} + f:path: {} + f:port: {} + f:scheme: {} + f:initialDelaySeconds: {} + f:periodSeconds: {} + f:successThreshold: {} + f:timeoutSeconds: {} + f:name: {} + f:ports: + .: {} + k:{"containerPort":16685,"protocol":"TCP"}: + .: {} + f:containerPort: {} + f:name: {} + f:protocol: {} + k:{"containerPort":16686,"protocol":"TCP"}: + .: {} + f:containerPort: {} + f:name: {} + f:protocol: {} + k:{"containerPort":16687,"protocol":"TCP"}: + .: {} + f:containerPort: {} + f:name: {} + f:protocol: {} + f:readinessProbe: + .: {} + f:failureThreshold: {} + f:httpGet: + .: {} + f:path: {} + f:port: {} + f:scheme: {} + f:initialDelaySeconds: {} + f:periodSeconds: {} + f:successThreshold: {} + f:timeoutSeconds: {} + f:resources: {} + f:terminationMessagePath: {} + f:terminationMessagePolicy: {} + f:volumeMounts: + .: {} + k:{"mountPath":"/certs"}: + .: {} + f:mountPath: {} + f:name: {} + f:readOnly: {} + k:{"mountPath":"/etc/config"}: + .: {} + f:mountPath: {} + f:name: {} + f:readOnly: {} + k:{"mountPath":"/etc/pki/ca-trust/extracted/pem"}: + .: {} + f:mountPath: {} + f:name: {} + f:readOnly: {} + k:{"name":"oauth-proxy"}: + .: {} + f:args: {} + f:image: {} + f:imagePullPolicy: {} + f:name: {} + f:ports: + .: {} + k:{"containerPort":8443,"protocol":"TCP"}: + .: {} + f:containerPort: {} + f:name: {} + f:protocol: {} + f:resources: {} + f:terminationMessagePath: {} + f:terminationMessagePolicy: {} + f:volumeMounts: + .: {} + k:{"mountPath":"/etc/pki/ca-trust/extracted/pem"}: + .: {} + f:mountPath: {} + f:name: {} + f:readOnly: {} + k:{"mountPath":"/etc/tls/private"}: + .: {} + f:mountPath: {} + f:name: {} + f:dnsPolicy: {} + f:enableServiceLinks: {} + f:restartPolicy: {} + f:schedulerName: {} + f:securityContext: {} + f:serviceAccount: {} + f:serviceAccountName: {} + f:terminationGracePeriodSeconds: {} + f:volumes: + .: {} + k:{"name":"certs"}: + .: {} + f:name: {} + f:secret: + .: {} + f:defaultMode: {} + f:secretName: {} + k:{"name":"my-jaeger-trusted-ca"}: + .: {} + f:configMap: + .: {} + f:defaultMode: {} + f:items: {} + f:name: {} + f:name: {} + k:{"name":"my-jaeger-ui-configuration-volume"}: + .: {} + f:configMap: + .: {} + f:defaultMode: {} + f:items: {} + f:name: {} + f:name: {} + k:{"name":"my-jaeger-ui-oauth-proxy-tls"}: + .: {} + f:name: {} + f:secret: + .: {} + f:defaultMode: {} + f:secretName: {} + manager: jaeger-operator + operation: Update + time: "2024-08-29T20:10:02Z" + - apiVersion: apps/v1 + fieldsType: FieldsV1 + fieldsV1: + f:metadata: + f:annotations: + f:deployment.kubernetes.io/revision: {} + f:status: + f:conditions: + .: {} + k:{"type":"Available"}: + .: {} + f:lastTransitionTime: {} + f:lastUpdateTime: {} + f:message: {} + f:reason: {} + f:status: {} + f:type: {} + k:{"type":"Progressing"}: + .: {} + f:lastTransitionTime: {} + f:lastUpdateTime: {} + f:message: {} + f:reason: {} + f:status: {} + f:type: {} + f:observedGeneration: {} + f:replicas: {} + f:unavailableReplicas: {} + f:updatedReplicas: {} + manager: kube-controller-manager + operation: Update + subresource: status + time: "2024-08-29T20:10:02Z" name: my-jaeger-query namespace: kuttl-test-ethical-gibbon + ownerReferences: + - apiVersion: jaegertracing.io/v1 + controller: true + kind: Jaeger + name: my-jaeger + uid: 6a92244e-a496-4a93-9129-85917b6dd6db spec: + progressDeadlineSeconds: 600 replicas: 1 + revisionHistoryLimit: 10 + selector: + matchLabels: + app: jaeger + app.kubernetes.io/component: query + app.kubernetes.io/instance: my-jaeger + app.kubernetes.io/managed-by: jaeger-operator + app.kubernetes.io/name: my-jaeger-query + app.kubernetes.io/part-of: jaeger + strategy: + type: Recreate + template: + metadata: + annotations: + linkerd.io/inject: disabled + prometheus.io/port: "16687" + prometheus.io/scrape: "true" + sidecar.istio.io/inject: "false" + sidecar.jaegertracing.io/inject: my-jaeger + creationTimestamp: null + labels: + app: jaeger + app.kubernetes.io/component: query + app.kubernetes.io/instance: my-jaeger + app.kubernetes.io/managed-by: jaeger-operator + app.kubernetes.io/name: my-jaeger-query + app.kubernetes.io/part-of: jaeger + spec: + containers: + - args: + - --query.ui-config=/etc/config/ui.json + - --es.server-urls=https://elasticsearch.kuttl-test-ethical-gibbon.svc.cluster.local:9200 + - --es.tls.enabled=true + - --es.tls.ca=/certs/ca + - --es.tls.cert=/certs/cert + - --es.tls.key=/certs/key + - --es.timeout=15s + - --es.num-shards=1 + - --es.num-replicas=0 + env: + - name: SPAN_STORAGE_TYPE + value: elasticsearch + - name: METRICS_STORAGE_TYPE + - name: JAEGER_DISABLED + value: "false" + - name: JAEGER_SERVICE_NAME + value: my-jaeger.kuttl-test-ethical-gibbon + - name: JAEGER_PROPAGATION + value: jaeger,b3,w3c + image: registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:831f83e8396d61822fb2dd576e711e0f10c1f8c7bd03bf0d4bec49dfb90cacca + imagePullPolicy: IfNotPresent + livenessProbe: + failureThreshold: 5 + httpGet: + path: / + port: 16687 + scheme: HTTP + initialDelaySeconds: 5 + periodSeconds: 15 + successThreshold: 1 + timeoutSeconds: 1 + name: jaeger-query + ports: + - containerPort: 16685 + name: grpc-query + protocol: TCP + - containerPort: 16686 + name: query + protocol: TCP + - containerPort: 16687 + name: admin-http + protocol: TCP + readinessProbe: + failureThreshold: 3 + httpGet: + path: / + port: 16687 + scheme: HTTP + initialDelaySeconds: 1 + periodSeconds: 10 + successThreshold: 1 + timeoutSeconds: 1 + resources: {} + terminationMessagePath: /dev/termination-log + terminationMessagePolicy: File + volumeMounts: + - mountPath: /etc/config + name: my-jaeger-ui-configuration-volume + readOnly: true + - mountPath: /etc/pki/ca-trust/extracted/pem + name: my-jaeger-trusted-ca + readOnly: true + - mountPath: /certs + name: certs + readOnly: true + - args: + - --cookie-secret=lszPu2N9Rj67QrdcJFwXS9 + - --https-address=:8443 + - '--openshift-sar={"namespace": "kuttl-test-ethical-gibbon", "resource": + "pods", "verb": "get"}' + - --openshift-service-account=my-jaeger-ui-proxy + - --provider=openshift + - --tls-cert=/etc/tls/private/tls.crt + - --tls-key=/etc/tls/private/tls.key + - --upstream=http://localhost:16686 + env: + - name: JAEGER_SERVICE_NAME + value: my-jaeger.kuttl-test-ethical-gibbon + - name: JAEGER_PROPAGATION + value: jaeger,b3,w3c + image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1613032062a68faa414cd5754a666b7b78ad0e52cf328eaaed355719423a55dd + imagePullPolicy: IfNotPresent + name: oauth-proxy + ports: + - containerPort: 8443 + name: public + protocol: TCP + resources: {} + terminationMessagePath: /dev/termination-log + terminationMessagePolicy: File + volumeMounts: + - mountPath: /etc/tls/private + name: my-jaeger-ui-oauth-proxy-tls + - mountPath: /etc/pki/ca-trust/extracted/pem + name: my-jaeger-trusted-ca + readOnly: true + - args: + - --agent.tags=cluster=undefined,deployment.name=my-jaeger-query,host.ip=${HOST_IP:},pod.name=${POD_NAME:},pod.namespace=kuttl-test-ethical-gibbon + - --reporter.grpc.host-port=dns:///my-jaeger-collector-headless.kuttl-test-ethical-gibbon.svc:14250 + - --reporter.grpc.tls.ca=/etc/pki/ca-trust/source/service-ca/service-ca.crt + - --reporter.grpc.tls.enabled=true + env: + - name: POD_NAME + valueFrom: + fieldRef: + apiVersion: v1 + fieldPath: metadata.name + - name: HOST_IP + valueFrom: + fieldRef: + apiVersion: v1 + fieldPath: status.hostIP + image: registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:5fcc9d365678f54c0f52005318dbbc32b87f743cae7e1a84568e769653f3b128 + imagePullPolicy: IfNotPresent + livenessProbe: + failureThreshold: 5 + httpGet: + path: / + port: 14271 + scheme: HTTP + initialDelaySeconds: 5 + periodSeconds: 15 + successThreshold: 1 + timeoutSeconds: 1 + name: jaeger-agent + ports: + - containerPort: 5775 + name: zk-compact-trft + protocol: UDP + - containerPort: 5778 + name: config-rest + protocol: TCP + - containerPort: 6831 + name: jg-compact-trft + protocol: UDP + - containerPort: 6832 + name: jg-binary-trft + protocol: UDP + - containerPort: 14271 + name: admin-http + protocol: TCP + readinessProbe: + failureThreshold: 3 + httpGet: + path: / + port: 14271 + scheme: HTTP + initialDelaySeconds: 1 + periodSeconds: 10 + successThreshold: 1 + timeoutSeconds: 1 + resources: {} + terminationMessagePath: /dev/termination-log + terminationMessagePolicy: File + volumeMounts: + - mountPath: /etc/pki/ca-trust/extracted/pem + name: my-jaeger-trusted-ca + readOnly: true + - mountPath: /etc/pki/ca-trust/source/service-ca + name: my-jaeger-service-ca + readOnly: true + dnsPolicy: ClusterFirst + enableServiceLinks: false + restartPolicy: Always + schedulerName: default-scheduler + securityContext: {} + serviceAccount: my-jaeger-ui-proxy + serviceAccountName: my-jaeger-ui-proxy + terminationGracePeriodSeconds: 30 + volumes: + - configMap: + defaultMode: 420 + items: + - key: ui + path: ui.json + name: my-jaeger-ui-configuration + name: my-jaeger-ui-configuration-volume + - configMap: + defaultMode: 420 + items: + - key: ca-bundle.crt + path: tls-ca-bundle.pem + name: my-jaeger-trusted-ca + name: my-jaeger-trusted-ca + - name: my-jaeger-ui-oauth-proxy-tls + secret: + defaultMode: 420 + secretName: my-jaeger-ui-oauth-proxy-tls + - name: certs + secret: + defaultMode: 420 + secretName: my-jaeger-jaeger-elasticsearch + - configMap: + defaultMode: 420 + items: + - key: service-ca.crt + path: service-ca.crt + name: my-jaeger-service-ca + name: my-jaeger-service-ca status: - readyReplicas: 1 + conditions: + - lastTransitionTime: "2024-08-29T20:05:01Z" + lastUpdateTime: "2024-08-29T20:05:01Z" + message: Deployment does not have minimum availability. + reason: MinimumReplicasUnavailable + status: "False" + type: Available + - lastTransitionTime: "2024-08-29T20:05:01Z" + lastUpdateTime: "2024-08-29T20:05:01Z" + message: ReplicaSet "my-jaeger-query-67c4fff457" is progressing. + reason: ReplicaSetUpdated + status: "True" + type: Progressing + observedGeneration: 3 + replicas: 1 + unavailableReplicas: 1 + updatedReplicas: 1 case.go:366: resource Deployment:kuttl-test-ethical-gibbon/my-jaeger-query: .status.readyReplicas: key is missing from map logger.go:42: 20:14:58 | collector-otlp-production-grpc | collector-otlp-production-grpc events from ns kuttl-test-ethical-gibbon: logger.go:42: 20:14:58 | collector-otlp-production-grpc | 2024-08-29 20:05:01 +0000 UTC Normal Pod my-jaeger-collector-859bd6f798-z92gk Binding Scheduled Successfully assigned kuttl-test-ethical-gibbon/my-jaeger-collector-859bd6f798-z92gk to ip-10-0-51-237.us-west-2.compute.internal default-scheduler logger.go:42: 20:14:58 | collector-otlp-production-grpc | 2024-08-29 20:05:01 +0000 UTC Normal Pod my-jaeger-collector-859bd6f798-z92gk AddedInterface Add eth0 [10.130.0.60/23] from ovn-kubernetes multus logger.go:42: 20:14:58 | collector-otlp-production-grpc | 2024-08-29 20:05:01 +0000 UTC Normal Pod my-jaeger-collector-859bd6f798-z92gk.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:b90e0d264bb0b051c2a4b9e6d499f407fed58f834ab8ebf30263ee1cf7c7ea3c" already present on machine kubelet logger.go:42: 20:14:58 | collector-otlp-production-grpc | 2024-08-29 20:05:01 +0000 UTC Normal Pod my-jaeger-collector-859bd6f798-z92gk.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 20:14:58 | collector-otlp-production-grpc | 2024-08-29 20:05:01 +0000 UTC Normal Pod my-jaeger-collector-859bd6f798-z92gk.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 20:14:58 | collector-otlp-production-grpc | 2024-08-29 20:05:01 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-859bd6f798 SuccessfulCreate Created pod: my-jaeger-collector-859bd6f798-z92gk replicaset-controller logger.go:42: 20:14:58 | collector-otlp-production-grpc | 2024-08-29 20:05:01 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-859bd6f798 to 1 deployment-controller logger.go:42: 20:14:58 | collector-otlp-production-grpc | 2024-08-29 20:05:01 +0000 UTC Normal Pod my-jaeger-query-67c4fff457-chzfq Binding Scheduled Successfully assigned kuttl-test-ethical-gibbon/my-jaeger-query-67c4fff457-chzfq to ip-10-0-51-237.us-west-2.compute.internal default-scheduler logger.go:42: 20:14:58 | collector-otlp-production-grpc | 2024-08-29 20:05:01 +0000 UTC Normal Pod my-jaeger-query-67c4fff457-chzfq AddedInterface Add eth0 [10.130.0.59/23] from ovn-kubernetes multus logger.go:42: 20:14:58 | collector-otlp-production-grpc | 2024-08-29 20:05:01 +0000 UTC Normal Pod my-jaeger-query-67c4fff457-chzfq.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:831f83e8396d61822fb2dd576e711e0f10c1f8c7bd03bf0d4bec49dfb90cacca" already present on machine kubelet logger.go:42: 20:14:58 | collector-otlp-production-grpc | 2024-08-29 20:05:01 +0000 UTC Normal Pod my-jaeger-query-67c4fff457-chzfq.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 20:14:58 | collector-otlp-production-grpc | 2024-08-29 20:05:01 +0000 UTC Normal Pod my-jaeger-query-67c4fff457-chzfq.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 20:14:58 | collector-otlp-production-grpc | 2024-08-29 20:05:01 +0000 UTC Normal Pod my-jaeger-query-67c4fff457-chzfq.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1613032062a68faa414cd5754a666b7b78ad0e52cf328eaaed355719423a55dd" already present on machine kubelet logger.go:42: 20:14:58 | collector-otlp-production-grpc | 2024-08-29 20:05:01 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-67c4fff457 SuccessfulCreate Created pod: my-jaeger-query-67c4fff457-chzfq replicaset-controller logger.go:42: 20:14:58 | collector-otlp-production-grpc | 2024-08-29 20:05:01 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-67c4fff457 to 1 deployment-controller logger.go:42: 20:14:58 | collector-otlp-production-grpc | 2024-08-29 20:05:02 +0000 UTC Normal Pod my-jaeger-query-67c4fff457-chzfq.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 20:14:58 | collector-otlp-production-grpc | 2024-08-29 20:05:02 +0000 UTC Normal Pod my-jaeger-query-67c4fff457-chzfq.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 20:14:58 | collector-otlp-production-grpc | 2024-08-29 20:05:02 +0000 UTC Normal Pod my-jaeger-query-67c4fff457-chzfq.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:5fcc9d365678f54c0f52005318dbbc32b87f743cae7e1a84568e769653f3b128" already present on machine kubelet logger.go:42: 20:14:58 | collector-otlp-production-grpc | 2024-08-29 20:05:02 +0000 UTC Normal Pod my-jaeger-query-67c4fff457-chzfq.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 20:14:58 | collector-otlp-production-grpc | 2024-08-29 20:05:02 +0000 UTC Normal Pod my-jaeger-query-67c4fff457-chzfq.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 20:14:58 | collector-otlp-production-grpc | 2024-08-29 20:05:03 +0000 UTC Warning Pod my-jaeger-collector-859bd6f798-z92gk.spec.containers{jaeger-collector} Unhealthy Readiness probe failed: HTTP probe failed with statuscode: 503 kubelet logger.go:42: 20:14:58 | collector-otlp-production-grpc | 2024-08-29 20:05:03 +0000 UTC Warning Pod my-jaeger-query-67c4fff457-chzfq.spec.containers{jaeger-query} Unhealthy Readiness probe failed: HTTP probe failed with statuscode: 503 kubelet logger.go:42: 20:14:58 | collector-otlp-production-grpc | 2024-08-29 20:05:13 +0000 UTC Warning Pod my-jaeger-collector-859bd6f798-z92gk.spec.containers{jaeger-collector} BackOff Back-off restarting failed container jaeger-collector in pod my-jaeger-collector-859bd6f798-z92gk_kuttl-test-ethical-gibbon(80a373bd-3254-4025-95f8-6936c4cfd3f5) kubelet logger.go:42: 20:14:58 | collector-otlp-production-grpc | 2024-08-29 20:05:13 +0000 UTC Warning Pod my-jaeger-query-67c4fff457-chzfq.spec.containers{jaeger-query} BackOff Back-off restarting failed container jaeger-query in pod my-jaeger-query-67c4fff457-chzfq_kuttl-test-ethical-gibbon(1c7130ca-3ffb-4e44-a38c-0ddcea327265) kubelet logger.go:42: 20:14:58 | collector-otlp-production-grpc | 2024-08-29 20:05:16 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 20:14:58 | collector-otlp-production-grpc | 2024-08-29 20:05:16 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 20:14:58 | collector-otlp-production-grpc | 2024-08-29 20:05:16 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 20:14:58 | collector-otlp-production-grpc | Deleting namespace: kuttl-test-ethical-gibbon === CONT kuttl/harness/collector-autoscale logger.go:42: 20:15:04 | collector-autoscale | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 20:15:04 | collector-autoscale | Creating namespace: kuttl-test-smooth-primate logger.go:42: 20:15:04 | collector-autoscale/1-install | starting test step 1-install logger.go:42: 20:15:05 | collector-autoscale/1-install | Jaeger:kuttl-test-smooth-primate/simple-prod created logger.go:42: 20:25:06 | collector-autoscale/1-install | test step failed 1-install case.go:364: failed in step 1-install case.go:366: --- Deployment:kuttl-test-smooth-primate/simple-prod-collector +++ Deployment:kuttl-test-smooth-primate/simple-prod-collector @@ -1,10 +1,418 @@ apiVersion: apps/v1 kind: Deployment metadata: + annotations: + linkerd.io/inject: disabled + prometheus.io/port: "14269" + prometheus.io/scrape: "true" + labels: + app: jaeger + app.kubernetes.io/component: collector + app.kubernetes.io/instance: simple-prod + app.kubernetes.io/managed-by: jaeger-operator + app.kubernetes.io/name: simple-prod-collector + app.kubernetes.io/part-of: jaeger + managedFields: + - apiVersion: apps/v1 + fieldsType: FieldsV1 + fieldsV1: + f:metadata: + f:annotations: + .: {} + f:linkerd.io/inject: {} + f:prometheus.io/port: {} + f:prometheus.io/scrape: {} + f:labels: + .: {} + f:app: {} + f:app.kubernetes.io/component: {} + f:app.kubernetes.io/instance: {} + f:app.kubernetes.io/managed-by: {} + f:app.kubernetes.io/name: {} + f:app.kubernetes.io/part-of: {} + f:ownerReferences: + .: {} + k:{"uid":"defbaf36-1f0e-4495-ba0e-38e00e35a9fd"}: {} + f:spec: + f:progressDeadlineSeconds: {} + f:replicas: {} + f:revisionHistoryLimit: {} + f:selector: {} + f:strategy: + f:type: {} + f:template: + f:metadata: + f:annotations: + .: {} + f:linkerd.io/inject: {} + f:prometheus.io/port: {} + f:prometheus.io/scrape: {} + f:sidecar.istio.io/inject: {} + f:labels: + .: {} + f:app: {} + f:app.kubernetes.io/component: {} + f:app.kubernetes.io/instance: {} + f:app.kubernetes.io/managed-by: {} + f:app.kubernetes.io/name: {} + f:app.kubernetes.io/part-of: {} + f:spec: + f:containers: + k:{"name":"jaeger-collector"}: + .: {} + f:args: {} + f:env: + .: {} + k:{"name":"COLLECTOR_OTLP_ENABLED"}: + .: {} + f:name: {} + f:value: {} + k:{"name":"COLLECTOR_ZIPKIN_HOST_PORT"}: + .: {} + f:name: {} + f:value: {} + k:{"name":"SPAN_STORAGE_TYPE"}: + .: {} + f:name: {} + f:value: {} + f:image: {} + f:imagePullPolicy: {} + f:livenessProbe: + .: {} + f:failureThreshold: {} + f:httpGet: + .: {} + f:path: {} + f:port: {} + f:scheme: {} + f:initialDelaySeconds: {} + f:periodSeconds: {} + f:successThreshold: {} + f:timeoutSeconds: {} + f:name: {} + f:ports: + .: {} + k:{"containerPort":4317,"protocol":"TCP"}: + .: {} + f:containerPort: {} + f:name: {} + f:protocol: {} + k:{"containerPort":4318,"protocol":"TCP"}: + .: {} + f:containerPort: {} + f:name: {} + f:protocol: {} + k:{"containerPort":9411,"protocol":"TCP"}: + .: {} + f:containerPort: {} + f:name: {} + f:protocol: {} + k:{"containerPort":14250,"protocol":"TCP"}: + .: {} + f:containerPort: {} + f:name: {} + f:protocol: {} + k:{"containerPort":14267,"protocol":"TCP"}: + .: {} + f:containerPort: {} + f:name: {} + f:protocol: {} + k:{"containerPort":14268,"protocol":"TCP"}: + .: {} + f:containerPort: {} + f:name: {} + f:protocol: {} + k:{"containerPort":14269,"protocol":"TCP"}: + .: {} + f:containerPort: {} + f:name: {} + f:protocol: {} + f:readinessProbe: + .: {} + f:failureThreshold: {} + f:httpGet: + .: {} + f:path: {} + f:port: {} + f:scheme: {} + f:initialDelaySeconds: {} + f:periodSeconds: {} + f:successThreshold: {} + f:timeoutSeconds: {} + f:resources: + .: {} + f:requests: + .: {} + f:memory: {} + f:terminationMessagePath: {} + f:terminationMessagePolicy: {} + f:volumeMounts: + .: {} + k:{"mountPath":"/certs"}: + .: {} + f:mountPath: {} + f:name: {} + f:readOnly: {} + k:{"mountPath":"/etc/jaeger/sampling"}: + .: {} + f:mountPath: {} + f:name: {} + f:readOnly: {} + k:{"mountPath":"/etc/pki/ca-trust/extracted/pem"}: + .: {} + f:mountPath: {} + f:name: {} + f:readOnly: {} + k:{"mountPath":"/etc/tls-config"}: + .: {} + f:mountPath: {} + f:name: {} + f:readOnly: {} + f:dnsPolicy: {} + f:enableServiceLinks: {} + f:restartPolicy: {} + f:schedulerName: {} + f:securityContext: {} + f:serviceAccount: {} + f:serviceAccountName: {} + f:terminationGracePeriodSeconds: {} + f:volumes: + .: {} + k:{"name":"certs"}: + .: {} + f:name: {} + f:secret: + .: {} + f:defaultMode: {} + f:secretName: {} + k:{"name":"simple-prod-collector-tls-config-volume"}: + .: {} + f:name: {} + f:secret: + .: {} + f:defaultMode: {} + f:secretName: {} + k:{"name":"simple-prod-sampling-configuration-volume"}: + .: {} + f:configMap: + .: {} + f:defaultMode: {} + f:items: {} + f:name: {} + f:name: {} + k:{"name":"simple-prod-trusted-ca"}: + .: {} + f:configMap: + .: {} + f:defaultMode: {} + f:items: {} + f:name: {} + f:name: {} + manager: jaeger-operator + operation: Update + time: "2024-08-29T20:15:10Z" + - apiVersion: apps/v1 + fieldsType: FieldsV1 + fieldsV1: + f:metadata: + f:annotations: + f:deployment.kubernetes.io/revision: {} + f:status: + f:conditions: + .: {} + k:{"type":"Available"}: + .: {} + f:lastTransitionTime: {} + f:lastUpdateTime: {} + f:message: {} + f:reason: {} + f:status: {} + f:type: {} + k:{"type":"Progressing"}: + .: {} + f:lastTransitionTime: {} + f:lastUpdateTime: {} + f:message: {} + f:reason: {} + f:status: {} + f:type: {} + f:observedGeneration: {} + f:replicas: {} + f:unavailableReplicas: {} + f:updatedReplicas: {} + manager: kube-controller-manager + operation: Update + subresource: status + time: "2024-08-29T20:15:10Z" name: simple-prod-collector namespace: kuttl-test-smooth-primate + ownerReferences: + - apiVersion: jaegertracing.io/v1 + controller: true + kind: Jaeger + name: simple-prod + uid: defbaf36-1f0e-4495-ba0e-38e00e35a9fd spec: + progressDeadlineSeconds: 600 replicas: 1 + revisionHistoryLimit: 10 + selector: + matchLabels: + app: jaeger + app.kubernetes.io/component: collector + app.kubernetes.io/instance: simple-prod + app.kubernetes.io/managed-by: jaeger-operator + app.kubernetes.io/name: simple-prod-collector + app.kubernetes.io/part-of: jaeger + strategy: + type: Recreate + template: + metadata: + annotations: + linkerd.io/inject: disabled + prometheus.io/port: "14269" + prometheus.io/scrape: "true" + sidecar.istio.io/inject: "false" + creationTimestamp: null + labels: + app: jaeger + app.kubernetes.io/component: collector + app.kubernetes.io/instance: simple-prod + app.kubernetes.io/managed-by: jaeger-operator + app.kubernetes.io/name: simple-prod-collector + app.kubernetes.io/part-of: jaeger + spec: + containers: + - args: + - --collector.grpc.tls.cert=/etc/tls-config/tls.crt + - --collector.grpc.tls.enabled=true + - --collector.grpc.tls.key=/etc/tls-config/tls.key + - --sampling.strategies-file=/etc/jaeger/sampling/sampling.json + - --es.server-urls=https://elasticsearch.kuttl-test-smooth-primate.svc.cluster.local:9200 + - --es.tls.enabled=true + - --es.tls.ca=/certs/ca + - --es.tls.cert=/certs/cert + - --es.tls.key=/certs/key + - --es.timeout=15s + - --es.num-shards=1 + - --es.num-replicas=0 + env: + - name: SPAN_STORAGE_TYPE + value: elasticsearch + - name: COLLECTOR_ZIPKIN_HOST_PORT + value: :9411 + - name: COLLECTOR_OTLP_ENABLED + value: "true" + image: registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:b90e0d264bb0b051c2a4b9e6d499f407fed58f834ab8ebf30263ee1cf7c7ea3c + imagePullPolicy: IfNotPresent + livenessProbe: + failureThreshold: 5 + httpGet: + path: / + port: 14269 + scheme: HTTP + initialDelaySeconds: 5 + periodSeconds: 15 + successThreshold: 1 + timeoutSeconds: 1 + name: jaeger-collector + ports: + - containerPort: 9411 + name: zipkin + protocol: TCP + - containerPort: 14267 + name: c-tchan-trft + protocol: TCP + - containerPort: 14268 + name: c-binary-trft + protocol: TCP + - containerPort: 14269 + name: admin-http + protocol: TCP + - containerPort: 14250 + name: grpc + protocol: TCP + - containerPort: 4317 + name: grpc-otlp + protocol: TCP + - containerPort: 4318 + name: http-otlp + protocol: TCP + readinessProbe: + failureThreshold: 3 + httpGet: + path: / + port: 14269 + scheme: HTTP + initialDelaySeconds: 1 + periodSeconds: 10 + successThreshold: 1 + timeoutSeconds: 1 + resources: + requests: + memory: 200m + terminationMessagePath: /dev/termination-log + terminationMessagePolicy: File + volumeMounts: + - mountPath: /etc/jaeger/sampling + name: simple-prod-sampling-configuration-volume + readOnly: true + - mountPath: /etc/tls-config + name: simple-prod-collector-tls-config-volume + readOnly: true + - mountPath: /etc/pki/ca-trust/extracted/pem + name: simple-prod-trusted-ca + readOnly: true + - mountPath: /certs + name: certs + readOnly: true + dnsPolicy: ClusterFirst + enableServiceLinks: false + restartPolicy: Always + schedulerName: default-scheduler + securityContext: {} + serviceAccount: simple-prod + serviceAccountName: simple-prod + terminationGracePeriodSeconds: 30 + volumes: + - configMap: + defaultMode: 420 + items: + - key: sampling + path: sampling.json + name: simple-prod-sampling-configuration + name: simple-prod-sampling-configuration-volume + - name: simple-prod-collector-tls-config-volume + secret: + defaultMode: 420 + secretName: simple-prod-collector-headless-tls + - configMap: + defaultMode: 420 + items: + - key: ca-bundle.crt + path: tls-ca-bundle.pem + name: simple-prod-trusted-ca + name: simple-prod-trusted-ca + - name: certs + secret: + defaultMode: 420 + secretName: simple-prod-jaeger-elasticsearch status: - readyReplicas: 1 + conditions: + - lastTransitionTime: "2024-08-29T20:15:10Z" + lastUpdateTime: "2024-08-29T20:15:10Z" + message: Deployment does not have minimum availability. + reason: MinimumReplicasUnavailable + status: "False" + type: Available + - lastTransitionTime: "2024-08-29T20:15:10Z" + lastUpdateTime: "2024-08-29T20:15:10Z" + message: ReplicaSet "simple-prod-collector-695f6875d7" is progressing. + reason: ReplicaSetUpdated + status: "True" + type: Progressing + observedGeneration: 1 + replicas: 1 + unavailableReplicas: 1 + updatedReplicas: 1 case.go:366: resource Deployment:kuttl-test-smooth-primate/simple-prod-collector: .status.readyReplicas: key is missing from map case.go:366: --- Deployment:kuttl-test-smooth-primate/simple-prod-query +++ Deployment:kuttl-test-smooth-primate/simple-prod-query @@ -1,10 +1,508 @@ apiVersion: apps/v1 kind: Deployment metadata: + annotations: + linkerd.io/inject: disabled + prometheus.io/port: "16687" + prometheus.io/scrape: "true" + sidecar.jaegertracing.io/inject: simple-prod + sidecar.jaegertracing.io/revision: "1" + labels: + app: jaeger + app.kubernetes.io/component: query + app.kubernetes.io/instance: simple-prod + app.kubernetes.io/managed-by: jaeger-operator + app.kubernetes.io/name: simple-prod-query + app.kubernetes.io/part-of: jaeger + sidecar.jaegertracing.io/injected: simple-prod + managedFields: + - apiVersion: apps/v1 + fieldsType: FieldsV1 + fieldsV1: + f:metadata: + f:annotations: + .: {} + f:linkerd.io/inject: {} + f:prometheus.io/port: {} + f:prometheus.io/scrape: {} + f:sidecar.jaegertracing.io/inject: {} + f:sidecar.jaegertracing.io/revision: {} + f:labels: + .: {} + f:app: {} + f:app.kubernetes.io/component: {} + f:app.kubernetes.io/instance: {} + f:app.kubernetes.io/managed-by: {} + f:app.kubernetes.io/name: {} + f:app.kubernetes.io/part-of: {} + f:ownerReferences: + .: {} + k:{"uid":"defbaf36-1f0e-4495-ba0e-38e00e35a9fd"}: {} + f:spec: + f:progressDeadlineSeconds: {} + f:replicas: {} + f:revisionHistoryLimit: {} + f:selector: {} + f:strategy: + f:type: {} + f:template: + f:metadata: + f:annotations: + .: {} + f:linkerd.io/inject: {} + f:prometheus.io/port: {} + f:prometheus.io/scrape: {} + f:sidecar.istio.io/inject: {} + f:sidecar.jaegertracing.io/inject: {} + f:labels: + .: {} + f:app: {} + f:app.kubernetes.io/component: {} + f:app.kubernetes.io/instance: {} + f:app.kubernetes.io/managed-by: {} + f:app.kubernetes.io/name: {} + f:app.kubernetes.io/part-of: {} + f:spec: + f:containers: + k:{"name":"jaeger-query"}: + .: {} + f:args: {} + f:env: + .: {} + k:{"name":"JAEGER_DISABLED"}: + .: {} + f:name: {} + f:value: {} + k:{"name":"METRICS_STORAGE_TYPE"}: + .: {} + f:name: {} + k:{"name":"SPAN_STORAGE_TYPE"}: + .: {} + f:name: {} + f:value: {} + f:image: {} + f:imagePullPolicy: {} + f:livenessProbe: + .: {} + f:failureThreshold: {} + f:httpGet: + .: {} + f:path: {} + f:port: {} + f:scheme: {} + f:initialDelaySeconds: {} + f:periodSeconds: {} + f:successThreshold: {} + f:timeoutSeconds: {} + f:name: {} + f:ports: + .: {} + k:{"containerPort":16685,"protocol":"TCP"}: + .: {} + f:containerPort: {} + f:name: {} + f:protocol: {} + k:{"containerPort":16686,"protocol":"TCP"}: + .: {} + f:containerPort: {} + f:name: {} + f:protocol: {} + k:{"containerPort":16687,"protocol":"TCP"}: + .: {} + f:containerPort: {} + f:name: {} + f:protocol: {} + f:readinessProbe: + .: {} + f:failureThreshold: {} + f:httpGet: + .: {} + f:path: {} + f:port: {} + f:scheme: {} + f:initialDelaySeconds: {} + f:periodSeconds: {} + f:successThreshold: {} + f:timeoutSeconds: {} + f:resources: {} + f:terminationMessagePath: {} + f:terminationMessagePolicy: {} + f:volumeMounts: + .: {} + k:{"mountPath":"/certs"}: + .: {} + f:mountPath: {} + f:name: {} + f:readOnly: {} + k:{"mountPath":"/etc/config"}: + .: {} + f:mountPath: {} + f:name: {} + f:readOnly: {} + k:{"mountPath":"/etc/pki/ca-trust/extracted/pem"}: + .: {} + f:mountPath: {} + f:name: {} + f:readOnly: {} + k:{"name":"oauth-proxy"}: + .: {} + f:args: {} + f:image: {} + f:imagePullPolicy: {} + f:name: {} + f:ports: + .: {} + k:{"containerPort":8443,"protocol":"TCP"}: + .: {} + f:containerPort: {} + f:name: {} + f:protocol: {} + f:resources: {} + f:terminationMessagePath: {} + f:terminationMessagePolicy: {} + f:volumeMounts: + .: {} + k:{"mountPath":"/etc/pki/ca-trust/extracted/pem"}: + .: {} + f:mountPath: {} + f:name: {} + f:readOnly: {} + k:{"mountPath":"/etc/tls/private"}: + .: {} + f:mountPath: {} + f:name: {} + f:dnsPolicy: {} + f:enableServiceLinks: {} + f:restartPolicy: {} + f:schedulerName: {} + f:securityContext: {} + f:serviceAccount: {} + f:serviceAccountName: {} + f:terminationGracePeriodSeconds: {} + f:volumes: + .: {} + k:{"name":"certs"}: + .: {} + f:name: {} + f:secret: + .: {} + f:defaultMode: {} + f:secretName: {} + k:{"name":"simple-prod-trusted-ca"}: + .: {} + f:configMap: + .: {} + f:defaultMode: {} + f:items: {} + f:name: {} + f:name: {} + k:{"name":"simple-prod-ui-configuration-volume"}: + .: {} + f:configMap: + .: {} + f:defaultMode: {} + f:items: {} + f:name: {} + f:name: {} + k:{"name":"simple-prod-ui-oauth-proxy-tls"}: + .: {} + f:name: {} + f:secret: + .: {} + f:defaultMode: {} + f:secretName: {} + manager: jaeger-operator + operation: Update + time: "2024-08-29T20:20:11Z" + - apiVersion: apps/v1 + fieldsType: FieldsV1 + fieldsV1: + f:metadata: + f:annotations: + f:deployment.kubernetes.io/revision: {} + f:status: + f:conditions: + .: {} + k:{"type":"Available"}: + .: {} + f:lastTransitionTime: {} + f:lastUpdateTime: {} + f:message: {} + f:reason: {} + f:status: {} + f:type: {} + k:{"type":"Progressing"}: + .: {} + f:lastTransitionTime: {} + f:lastUpdateTime: {} + f:message: {} + f:reason: {} + f:status: {} + f:type: {} + f:observedGeneration: {} + f:replicas: {} + f:unavailableReplicas: {} + f:updatedReplicas: {} + manager: kube-controller-manager + operation: Update + subresource: status + time: "2024-08-29T20:20:11Z" name: simple-prod-query namespace: kuttl-test-smooth-primate + ownerReferences: + - apiVersion: jaegertracing.io/v1 + controller: true + kind: Jaeger + name: simple-prod + uid: defbaf36-1f0e-4495-ba0e-38e00e35a9fd spec: + progressDeadlineSeconds: 600 replicas: 1 + revisionHistoryLimit: 10 + selector: + matchLabels: + app: jaeger + app.kubernetes.io/component: query + app.kubernetes.io/instance: simple-prod + app.kubernetes.io/managed-by: jaeger-operator + app.kubernetes.io/name: simple-prod-query + app.kubernetes.io/part-of: jaeger + strategy: + type: Recreate + template: + metadata: + annotations: + linkerd.io/inject: disabled + prometheus.io/port: "16687" + prometheus.io/scrape: "true" + sidecar.istio.io/inject: "false" + sidecar.jaegertracing.io/inject: simple-prod + creationTimestamp: null + labels: + app: jaeger + app.kubernetes.io/component: query + app.kubernetes.io/instance: simple-prod + app.kubernetes.io/managed-by: jaeger-operator + app.kubernetes.io/name: simple-prod-query + app.kubernetes.io/part-of: jaeger + spec: + containers: + - args: + - --query.ui-config=/etc/config/ui.json + - --es.server-urls=https://elasticsearch.kuttl-test-smooth-primate.svc.cluster.local:9200 + - --es.tls.enabled=true + - --es.tls.ca=/certs/ca + - --es.tls.cert=/certs/cert + - --es.tls.key=/certs/key + - --es.timeout=15s + - --es.num-shards=1 + - --es.num-replicas=0 + env: + - name: SPAN_STORAGE_TYPE + value: elasticsearch + - name: METRICS_STORAGE_TYPE + - name: JAEGER_DISABLED + value: "false" + - name: JAEGER_SERVICE_NAME + value: simple-prod.kuttl-test-smooth-primate + - name: JAEGER_PROPAGATION + value: jaeger,b3,w3c + image: registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:831f83e8396d61822fb2dd576e711e0f10c1f8c7bd03bf0d4bec49dfb90cacca + imagePullPolicy: IfNotPresent + livenessProbe: + failureThreshold: 5 + httpGet: + path: / + port: 16687 + scheme: HTTP + initialDelaySeconds: 5 + periodSeconds: 15 + successThreshold: 1 + timeoutSeconds: 1 + name: jaeger-query + ports: + - containerPort: 16685 + name: grpc-query + protocol: TCP + - containerPort: 16686 + name: query + protocol: TCP + - containerPort: 16687 + name: admin-http + protocol: TCP + readinessProbe: + failureThreshold: 3 + httpGet: + path: / + port: 16687 + scheme: HTTP + initialDelaySeconds: 1 + periodSeconds: 10 + successThreshold: 1 + timeoutSeconds: 1 + resources: {} + terminationMessagePath: /dev/termination-log + terminationMessagePolicy: File + volumeMounts: + - mountPath: /etc/config + name: simple-prod-ui-configuration-volume + readOnly: true + - mountPath: /etc/pki/ca-trust/extracted/pem + name: simple-prod-trusted-ca + readOnly: true + - mountPath: /certs + name: certs + readOnly: true + - args: + - --cookie-secret=ei6bn7bXB8vF4QFbsnKGux + - --https-address=:8443 + - '--openshift-sar={"namespace": "kuttl-test-smooth-primate", "resource": + "pods", "verb": "get"}' + - --openshift-service-account=simple-prod-ui-proxy + - --provider=openshift + - --tls-cert=/etc/tls/private/tls.crt + - --tls-key=/etc/tls/private/tls.key + - --upstream=http://localhost:16686 + env: + - name: JAEGER_SERVICE_NAME + value: simple-prod.kuttl-test-smooth-primate + - name: JAEGER_PROPAGATION + value: jaeger,b3,w3c + image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1613032062a68faa414cd5754a666b7b78ad0e52cf328eaaed355719423a55dd + imagePullPolicy: IfNotPresent + name: oauth-proxy + ports: + - containerPort: 8443 + name: public + protocol: TCP + resources: {} + terminationMessagePath: /dev/termination-log + terminationMessagePolicy: File + volumeMounts: + - mountPath: /etc/tls/private + name: simple-prod-ui-oauth-proxy-tls + - mountPath: /etc/pki/ca-trust/extracted/pem + name: simple-prod-trusted-ca + readOnly: true + - args: + - --agent.tags=cluster=undefined,deployment.name=simple-prod-query,host.ip=${HOST_IP:},pod.name=${POD_NAME:},pod.namespace=kuttl-test-smooth-primate + - --reporter.grpc.host-port=dns:///simple-prod-collector-headless.kuttl-test-smooth-primate.svc:14250 + - --reporter.grpc.tls.ca=/etc/pki/ca-trust/source/service-ca/service-ca.crt + - --reporter.grpc.tls.enabled=true + env: + - name: POD_NAME + valueFrom: + fieldRef: + apiVersion: v1 + fieldPath: metadata.name + - name: HOST_IP + valueFrom: + fieldRef: + apiVersion: v1 + fieldPath: status.hostIP + image: registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:5fcc9d365678f54c0f52005318dbbc32b87f743cae7e1a84568e769653f3b128 + imagePullPolicy: IfNotPresent + livenessProbe: + failureThreshold: 5 + httpGet: + path: / + port: 14271 + scheme: HTTP + initialDelaySeconds: 5 + periodSeconds: 15 + successThreshold: 1 + timeoutSeconds: 1 + name: jaeger-agent + ports: + - containerPort: 5775 + name: zk-compact-trft + protocol: UDP + - containerPort: 5778 + name: config-rest + protocol: TCP + - containerPort: 6831 + name: jg-compact-trft + protocol: UDP + - containerPort: 6832 + name: jg-binary-trft + protocol: UDP + - containerPort: 14271 + name: admin-http + protocol: TCP + readinessProbe: + failureThreshold: 3 + httpGet: + path: / + port: 14271 + scheme: HTTP + initialDelaySeconds: 1 + periodSeconds: 10 + successThreshold: 1 + timeoutSeconds: 1 + resources: {} + terminationMessagePath: /dev/termination-log + terminationMessagePolicy: File + volumeMounts: + - mountPath: /etc/pki/ca-trust/extracted/pem + name: simple-prod-trusted-ca + readOnly: true + - mountPath: /etc/pki/ca-trust/source/service-ca + name: simple-prod-service-ca + readOnly: true + dnsPolicy: ClusterFirst + enableServiceLinks: false + restartPolicy: Always + schedulerName: default-scheduler + securityContext: {} + serviceAccount: simple-prod-ui-proxy + serviceAccountName: simple-prod-ui-proxy + terminationGracePeriodSeconds: 30 + volumes: + - configMap: + defaultMode: 420 + items: + - key: ui + path: ui.json + name: simple-prod-ui-configuration + name: simple-prod-ui-configuration-volume + - configMap: + defaultMode: 420 + items: + - key: ca-bundle.crt + path: tls-ca-bundle.pem + name: simple-prod-trusted-ca + name: simple-prod-trusted-ca + - name: simple-prod-ui-oauth-proxy-tls + secret: + defaultMode: 420 + secretName: simple-prod-ui-oauth-proxy-tls + - name: certs + secret: + defaultMode: 420 + secretName: simple-prod-jaeger-elasticsearch + - configMap: + defaultMode: 420 + items: + - key: service-ca.crt + path: service-ca.crt + name: simple-prod-service-ca + name: simple-prod-service-ca status: - readyReplicas: 1 + conditions: + - lastTransitionTime: "2024-08-29T20:15:10Z" + lastUpdateTime: "2024-08-29T20:15:10Z" + message: Deployment does not have minimum availability. + reason: MinimumReplicasUnavailable + status: "False" + type: Available + - lastTransitionTime: "2024-08-29T20:15:10Z" + lastUpdateTime: "2024-08-29T20:15:10Z" + message: ReplicaSet "simple-prod-query-56b5b9cd8f" is progressing. + reason: ReplicaSetUpdated + status: "True" + type: Progressing + observedGeneration: 3 + replicas: 1 + unavailableReplicas: 1 + updatedReplicas: 1 case.go:366: resource Deployment:kuttl-test-smooth-primate/simple-prod-query: .status.readyReplicas: key is missing from map logger.go:42: 20:25:06 | collector-autoscale | collector-autoscale events from ns kuttl-test-smooth-primate: logger.go:42: 20:25:06 | collector-autoscale | 2024-08-29 20:15:10 +0000 UTC Normal Pod simple-prod-collector-695f6875d7-8p54z Binding Scheduled Successfully assigned kuttl-test-smooth-primate/simple-prod-collector-695f6875d7-8p54z to ip-10-0-51-237.us-west-2.compute.internal default-scheduler logger.go:42: 20:25:06 | collector-autoscale | 2024-08-29 20:15:10 +0000 UTC Normal Pod simple-prod-collector-695f6875d7-8p54z AddedInterface Add eth0 [10.130.0.61/23] from ovn-kubernetes multus logger.go:42: 20:25:06 | collector-autoscale | 2024-08-29 20:15:10 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-695f6875d7 SuccessfulCreate Created pod: simple-prod-collector-695f6875d7-8p54z replicaset-controller logger.go:42: 20:25:06 | collector-autoscale | 2024-08-29 20:15:10 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-695f6875d7 to 1 deployment-controller logger.go:42: 20:25:06 | collector-autoscale | 2024-08-29 20:15:10 +0000 UTC Normal Pod simple-prod-query-56b5b9cd8f-9ckkg Binding Scheduled Successfully assigned kuttl-test-smooth-primate/simple-prod-query-56b5b9cd8f-9ckkg to ip-10-0-51-237.us-west-2.compute.internal default-scheduler logger.go:42: 20:25:06 | collector-autoscale | 2024-08-29 20:15:10 +0000 UTC Normal ReplicaSet.apps simple-prod-query-56b5b9cd8f SuccessfulCreate Created pod: simple-prod-query-56b5b9cd8f-9ckkg replicaset-controller logger.go:42: 20:25:06 | collector-autoscale | 2024-08-29 20:15:10 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-56b5b9cd8f to 1 deployment-controller logger.go:42: 20:25:06 | collector-autoscale | 2024-08-29 20:15:11 +0000 UTC Normal Pod simple-prod-collector-695f6875d7-8p54z.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:b90e0d264bb0b051c2a4b9e6d499f407fed58f834ab8ebf30263ee1cf7c7ea3c" already present on machine kubelet logger.go:42: 20:25:06 | collector-autoscale | 2024-08-29 20:15:11 +0000 UTC Normal Pod simple-prod-collector-695f6875d7-8p54z.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 20:25:06 | collector-autoscale | 2024-08-29 20:15:11 +0000 UTC Normal Pod simple-prod-collector-695f6875d7-8p54z.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 20:25:06 | collector-autoscale | 2024-08-29 20:15:11 +0000 UTC Normal Pod simple-prod-query-56b5b9cd8f-9ckkg AddedInterface Add eth0 [10.130.0.62/23] from ovn-kubernetes multus logger.go:42: 20:25:06 | collector-autoscale | 2024-08-29 20:15:11 +0000 UTC Normal Pod simple-prod-query-56b5b9cd8f-9ckkg.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:831f83e8396d61822fb2dd576e711e0f10c1f8c7bd03bf0d4bec49dfb90cacca" already present on machine kubelet logger.go:42: 20:25:06 | collector-autoscale | 2024-08-29 20:15:11 +0000 UTC Normal Pod simple-prod-query-56b5b9cd8f-9ckkg.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 20:25:06 | collector-autoscale | 2024-08-29 20:15:11 +0000 UTC Normal Pod simple-prod-query-56b5b9cd8f-9ckkg.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 20:25:06 | collector-autoscale | 2024-08-29 20:15:11 +0000 UTC Normal Pod simple-prod-query-56b5b9cd8f-9ckkg.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1613032062a68faa414cd5754a666b7b78ad0e52cf328eaaed355719423a55dd" already present on machine kubelet logger.go:42: 20:25:06 | collector-autoscale | 2024-08-29 20:15:11 +0000 UTC Normal Pod simple-prod-query-56b5b9cd8f-9ckkg.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 20:25:06 | collector-autoscale | 2024-08-29 20:15:11 +0000 UTC Normal Pod simple-prod-query-56b5b9cd8f-9ckkg.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 20:25:06 | collector-autoscale | 2024-08-29 20:15:11 +0000 UTC Normal Pod simple-prod-query-56b5b9cd8f-9ckkg.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:5fcc9d365678f54c0f52005318dbbc32b87f743cae7e1a84568e769653f3b128" already present on machine kubelet logger.go:42: 20:25:06 | collector-autoscale | 2024-08-29 20:15:11 +0000 UTC Normal Pod simple-prod-query-56b5b9cd8f-9ckkg.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 20:25:06 | collector-autoscale | 2024-08-29 20:15:11 +0000 UTC Normal Pod simple-prod-query-56b5b9cd8f-9ckkg.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 20:25:06 | collector-autoscale | 2024-08-29 20:15:12 +0000 UTC Warning Pod simple-prod-collector-695f6875d7-8p54z.spec.containers{jaeger-collector} Unhealthy Readiness probe failed: HTTP probe failed with statuscode: 503 kubelet logger.go:42: 20:25:06 | collector-autoscale | 2024-08-29 20:15:12 +0000 UTC Warning Pod simple-prod-query-56b5b9cd8f-9ckkg.spec.containers{jaeger-query} Unhealthy Readiness probe failed: HTTP probe failed with statuscode: 503 kubelet logger.go:42: 20:25:06 | collector-autoscale | 2024-08-29 20:15:22 +0000 UTC Warning Pod simple-prod-collector-695f6875d7-8p54z.spec.containers{jaeger-collector} BackOff Back-off restarting failed container jaeger-collector in pod simple-prod-collector-695f6875d7-8p54z_kuttl-test-smooth-primate(4b55ec10-291a-414b-9a5b-7e546b7736af) kubelet logger.go:42: 20:25:06 | collector-autoscale | 2024-08-29 20:15:22 +0000 UTC Warning Pod simple-prod-query-56b5b9cd8f-9ckkg.spec.containers{jaeger-query} BackOff Back-off restarting failed container jaeger-query in pod simple-prod-query-56b5b9cd8f-9ckkg_kuttl-test-smooth-primate(7e61784d-5f3c-4cf5-a775-ce7626751931) kubelet logger.go:42: 20:25:06 | collector-autoscale | 2024-08-29 20:15:25 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 20:25:06 | collector-autoscale | 2024-08-29 20:15:25 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 20:25:06 | collector-autoscale | 2024-08-29 20:15:25 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 20:25:06 | collector-autoscale | Deleting namespace: kuttl-test-smooth-primate === CONT kuttl/harness/collector-otlp-allinone-grpc logger.go:42: 20:25:13 | collector-otlp-allinone-grpc | Creating namespace: kuttl-test-working-goblin logger.go:42: 20:25:13 | collector-otlp-allinone-grpc/0-install | starting test step 0-install logger.go:42: 20:25:13 | collector-otlp-allinone-grpc/0-install | Jaeger:kuttl-test-working-goblin/my-jaeger created logger.go:42: 20:25:15 | collector-otlp-allinone-grpc/0-install | test step completed 0-install logger.go:42: 20:25:15 | collector-otlp-allinone-grpc/1-smoke-test | starting test step 1-smoke-test logger.go:42: 20:25:15 | collector-otlp-allinone-grpc/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 20:25:17 | collector-otlp-allinone-grpc/1-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 20:25:25 | collector-otlp-allinone-grpc/1-smoke-test | running command: [sh -c REPORTING_PROTOCOL=grpc ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/otlp-smoke-test.yaml.template -o otlp-smoke-test-job.yaml] logger.go:42: 20:25:25 | collector-otlp-allinone-grpc/1-smoke-test | running command: [sh -c kubectl create -f otlp-smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 20:25:25 | collector-otlp-allinone-grpc/1-smoke-test | job.batch/report-span created logger.go:42: 20:25:26 | collector-otlp-allinone-grpc/1-smoke-test | job.batch/check-span created logger.go:42: 20:25:46 | collector-otlp-allinone-grpc/1-smoke-test | test step completed 1-smoke-test logger.go:42: 20:25:46 | collector-otlp-allinone-grpc | collector-otlp-allinone-grpc events from ns kuttl-test-working-goblin: logger.go:42: 20:25:46 | collector-otlp-allinone-grpc | 2024-08-29 20:25:14 +0000 UTC Normal Pod my-jaeger-5cdd898bf7-9dchx Binding Scheduled Successfully assigned kuttl-test-working-goblin/my-jaeger-5cdd898bf7-9dchx to ip-10-0-51-237.us-west-2.compute.internal default-scheduler logger.go:42: 20:25:46 | collector-otlp-allinone-grpc | 2024-08-29 20:25:14 +0000 UTC Normal Pod my-jaeger-5cdd898bf7-9dchx AddedInterface Add eth0 [10.130.0.63/23] from ovn-kubernetes multus logger.go:42: 20:25:46 | collector-otlp-allinone-grpc | 2024-08-29 20:25:14 +0000 UTC Normal Pod my-jaeger-5cdd898bf7-9dchx.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:345d3b6c507df9f8a18948a594e7a03698085689c6965faf79b8aadc74a0ceaa" already present on machine kubelet logger.go:42: 20:25:46 | collector-otlp-allinone-grpc | 2024-08-29 20:25:14 +0000 UTC Normal Pod my-jaeger-5cdd898bf7-9dchx.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 20:25:46 | collector-otlp-allinone-grpc | 2024-08-29 20:25:14 +0000 UTC Normal Pod my-jaeger-5cdd898bf7-9dchx.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 20:25:46 | collector-otlp-allinone-grpc | 2024-08-29 20:25:14 +0000 UTC Normal Pod my-jaeger-5cdd898bf7-9dchx.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1613032062a68faa414cd5754a666b7b78ad0e52cf328eaaed355719423a55dd" already present on machine kubelet logger.go:42: 20:25:46 | collector-otlp-allinone-grpc | 2024-08-29 20:25:14 +0000 UTC Normal Pod my-jaeger-5cdd898bf7-9dchx.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 20:25:46 | collector-otlp-allinone-grpc | 2024-08-29 20:25:14 +0000 UTC Normal Pod my-jaeger-5cdd898bf7-9dchx.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 20:25:46 | collector-otlp-allinone-grpc | 2024-08-29 20:25:14 +0000 UTC Normal ReplicaSet.apps my-jaeger-5cdd898bf7 SuccessfulCreate Created pod: my-jaeger-5cdd898bf7-9dchx replicaset-controller logger.go:42: 20:25:46 | collector-otlp-allinone-grpc | 2024-08-29 20:25:14 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-5cdd898bf7 to 1 deployment-controller logger.go:42: 20:25:46 | collector-otlp-allinone-grpc | 2024-08-29 20:25:18 +0000 UTC Normal Pod my-jaeger-5cdd898bf7-9dchx.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 20:25:46 | collector-otlp-allinone-grpc | 2024-08-29 20:25:18 +0000 UTC Normal Pod my-jaeger-5cdd898bf7-9dchx.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 20:25:46 | collector-otlp-allinone-grpc | 2024-08-29 20:25:18 +0000 UTC Normal ReplicaSet.apps my-jaeger-5cdd898bf7 SuccessfulDelete Deleted pod: my-jaeger-5cdd898bf7-9dchx replicaset-controller logger.go:42: 20:25:46 | collector-otlp-allinone-grpc | 2024-08-29 20:25:18 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled down replica set my-jaeger-5cdd898bf7 to 0 from 1 deployment-controller logger.go:42: 20:25:46 | collector-otlp-allinone-grpc | 2024-08-29 20:25:19 +0000 UTC Normal Pod my-jaeger-5585cd975c-5hqdn Binding Scheduled Successfully assigned kuttl-test-working-goblin/my-jaeger-5585cd975c-5hqdn to ip-10-0-51-237.us-west-2.compute.internal default-scheduler logger.go:42: 20:25:46 | collector-otlp-allinone-grpc | 2024-08-29 20:25:19 +0000 UTC Normal Pod my-jaeger-5585cd975c-5hqdn AddedInterface Add eth0 [10.130.0.64/23] from ovn-kubernetes multus logger.go:42: 20:25:46 | collector-otlp-allinone-grpc | 2024-08-29 20:25:19 +0000 UTC Normal Pod my-jaeger-5585cd975c-5hqdn.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:345d3b6c507df9f8a18948a594e7a03698085689c6965faf79b8aadc74a0ceaa" already present on machine kubelet logger.go:42: 20:25:46 | collector-otlp-allinone-grpc | 2024-08-29 20:25:19 +0000 UTC Normal Pod my-jaeger-5585cd975c-5hqdn.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 20:25:46 | collector-otlp-allinone-grpc | 2024-08-29 20:25:19 +0000 UTC Normal Pod my-jaeger-5585cd975c-5hqdn.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 20:25:46 | collector-otlp-allinone-grpc | 2024-08-29 20:25:19 +0000 UTC Normal Pod my-jaeger-5585cd975c-5hqdn.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1613032062a68faa414cd5754a666b7b78ad0e52cf328eaaed355719423a55dd" already present on machine kubelet logger.go:42: 20:25:46 | collector-otlp-allinone-grpc | 2024-08-29 20:25:19 +0000 UTC Normal Pod my-jaeger-5585cd975c-5hqdn.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 20:25:46 | collector-otlp-allinone-grpc | 2024-08-29 20:25:19 +0000 UTC Normal Pod my-jaeger-5585cd975c-5hqdn.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 20:25:46 | collector-otlp-allinone-grpc | 2024-08-29 20:25:19 +0000 UTC Normal ReplicaSet.apps my-jaeger-5585cd975c SuccessfulCreate Created pod: my-jaeger-5585cd975c-5hqdn replicaset-controller logger.go:42: 20:25:46 | collector-otlp-allinone-grpc | 2024-08-29 20:25:19 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-5585cd975c to 1 deployment-controller logger.go:42: 20:25:46 | collector-otlp-allinone-grpc | 2024-08-29 20:25:25 +0000 UTC Normal Pod report-span-rf2q8 Binding Scheduled Successfully assigned kuttl-test-working-goblin/report-span-rf2q8 to ip-10-0-51-237.us-west-2.compute.internal default-scheduler logger.go:42: 20:25:46 | collector-otlp-allinone-grpc | 2024-08-29 20:25:25 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-rf2q8 job-controller logger.go:42: 20:25:46 | collector-otlp-allinone-grpc | 2024-08-29 20:25:26 +0000 UTC Normal Pod check-span-zwf8f Binding Scheduled Successfully assigned kuttl-test-working-goblin/check-span-zwf8f to ip-10-0-51-237.us-west-2.compute.internal default-scheduler logger.go:42: 20:25:46 | collector-otlp-allinone-grpc | 2024-08-29 20:25:26 +0000 UTC Normal Pod check-span-zwf8f AddedInterface Add eth0 [10.130.0.66/23] from ovn-kubernetes multus logger.go:42: 20:25:46 | collector-otlp-allinone-grpc | 2024-08-29 20:25:26 +0000 UTC Normal Pod check-span-zwf8f.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 20:25:46 | collector-otlp-allinone-grpc | 2024-08-29 20:25:26 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-zwf8f job-controller logger.go:42: 20:25:46 | collector-otlp-allinone-grpc | 2024-08-29 20:25:26 +0000 UTC Normal Pod report-span-rf2q8 AddedInterface Add eth0 [10.130.0.65/23] from ovn-kubernetes multus logger.go:42: 20:25:46 | collector-otlp-allinone-grpc | 2024-08-29 20:25:26 +0000 UTC Normal Pod report-span-rf2q8.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 20:25:46 | collector-otlp-allinone-grpc | 2024-08-29 20:25:27 +0000 UTC Normal Pod check-span-zwf8f.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 810ms (810ms including waiting) kubelet logger.go:42: 20:25:46 | collector-otlp-allinone-grpc | 2024-08-29 20:25:27 +0000 UTC Normal Pod check-span-zwf8f.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 20:25:46 | collector-otlp-allinone-grpc | 2024-08-29 20:25:27 +0000 UTC Normal Pod check-span-zwf8f.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 20:25:46 | collector-otlp-allinone-grpc | 2024-08-29 20:25:27 +0000 UTC Normal Pod report-span-rf2q8.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 883ms (883ms including waiting) kubelet logger.go:42: 20:25:46 | collector-otlp-allinone-grpc | 2024-08-29 20:25:27 +0000 UTC Normal Pod report-span-rf2q8.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 20:25:46 | collector-otlp-allinone-grpc | 2024-08-29 20:25:27 +0000 UTC Normal Pod report-span-rf2q8.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 20:25:46 | collector-otlp-allinone-grpc | 2024-08-29 20:25:46 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 20:25:46 | collector-otlp-allinone-grpc | Deleting namespace: kuttl-test-working-goblin === CONT kuttl/harness/cassandra-spark logger.go:42: 20:25:58 | cassandra-spark | Ignoring 01-assert.yaml.template as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 20:25:58 | cassandra-spark | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 20:25:58 | cassandra-spark | Creating namespace: kuttl-test-equal-goshawk logger.go:42: 20:25:58 | cassandra-spark | cassandra-spark events from ns kuttl-test-equal-goshawk: logger.go:42: 20:25:58 | cassandra-spark | Deleting namespace: kuttl-test-equal-goshawk === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- FAIL: kuttl (2527.77s) --- FAIL: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (6.10s) --- PASS: kuttl/harness/collector-otlp-allinone-http (38.98s) --- FAIL: kuttl/harness/set-custom-img (607.29s) --- FAIL: kuttl/harness/collector-otlp-production-http (607.09s) --- FAIL: kuttl/harness/collector-otlp-production-grpc (607.72s) --- FAIL: kuttl/harness/collector-autoscale (608.21s) --- PASS: kuttl/harness/collector-otlp-allinone-grpc (45.62s) --- PASS: kuttl/harness/cassandra-spark (6.09s) FAIL + exit_code=1 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name miscellaneous --report --output /logs/artifacts/miscellaneous.xml ./artifacts/kuttl-report.xml time="2024-08-29T20:26:06Z" level=debug msg="Setting a new name for the test suites" time="2024-08-29T20:26:06Z" level=debug msg="Removing 'artifacts' TestCase" time="2024-08-29T20:26:06Z" level=debug msg="normalizing test case names" time="2024-08-29T20:26:06Z" level=debug msg="miscellaneous/artifacts -> miscellaneous_artifacts" time="2024-08-29T20:26:06Z" level=debug msg="miscellaneous/collector-otlp-allinone-http -> miscellaneous_collector_otlp_allinone_http" time="2024-08-29T20:26:06Z" level=debug msg="miscellaneous/set-custom-img -> miscellaneous_set_custom_img" time="2024-08-29T20:26:06Z" level=debug msg="miscellaneous/collector-otlp-production-http -> miscellaneous_collector_otlp_production_http" time="2024-08-29T20:26:06Z" level=debug msg="miscellaneous/collector-otlp-production-grpc -> miscellaneous_collector_otlp_production_grpc" time="2024-08-29T20:26:06Z" level=debug msg="miscellaneous/collector-autoscale -> miscellaneous_collector_autoscale" time="2024-08-29T20:26:06Z" level=debug msg="miscellaneous/collector-otlp-allinone-grpc -> miscellaneous_collector_otlp_allinone_grpc" time="2024-08-29T20:26:06Z" level=debug msg="miscellaneous/cassandra-spark -> miscellaneous_cassandra_spark" +----------------------------------------------+--------+ | NAME | RESULT | +----------------------------------------------+--------+ | miscellaneous_artifacts | passed | | miscellaneous_collector_otlp_allinone_http | passed | | miscellaneous_set_custom_img | failed | | miscellaneous_collector_otlp_production_http | failed | | miscellaneous_collector_otlp_production_grpc | failed | | miscellaneous_collector_autoscale | failed | | miscellaneous_collector_otlp_allinone_grpc | passed | | miscellaneous_cassandra_spark | passed | +----------------------------------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh sidecar false true + '[' 3 -ne 3 ']' + test_suite_name=sidecar + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/sidecar.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-sidecar make[2]: Entering directory '/tmp/jaeger-tests' ./tests/e2e/sidecar/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.16.8 True False 82m Cluster version is 4.16.8' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.16.8 True False 82m Cluster version is 4.16.8' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ head -n 1 +++ sort -V ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/sidecar/render.sh ++ export SUITE_DIR=./tests/e2e/sidecar ++ SUITE_DIR=./tests/e2e/sidecar ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/sidecar ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + jaeger_service_name=order + start_test sidecar-deployment + '[' 1 -ne 1 ']' + test_name=sidecar-deployment + echo =========================================================================== =========================================================================== + info 'Rendering files for test sidecar-deployment' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test sidecar-deployment\e[0m' Rendering files for test sidecar-deployment + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/sidecar/_build + '[' _build '!=' _build ']' + mkdir -p sidecar-deployment + cd sidecar-deployment + render_install_vertx 01 + '[' 1 -ne 1 ']' + test_step=01 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-assert.yaml.template -o ./01-assert.yaml + render_find_service agent-as-sidecar allInOne order 00 03 + '[' 5 -ne 5 ']' + jaeger=agent-as-sidecar + deployment_strategy=allInOne + service_name=order + job_number=00 + test_step=03 + export JAEGER_NAME=agent-as-sidecar + JAEGER_NAME=agent-as-sidecar + export JOB_NUMBER=00 + JOB_NUMBER=00 + export SERVICE_NAME=order + SERVICE_NAME=order + export JAEGER_QUERY_ENDPOINT + '[' true = true ']' + '[' allInOne '!=' allInOne ']' + template=/tmp/jaeger-tests/tests/templates/find-service.yaml.template + JAEGER_QUERY_ENDPOINT=http://agent-as-sidecar-query:16686 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/find-service.yaml.template -o ./03-find-service.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-find-service.yaml.template -o ./03-assert.yaml + unset JAEGER_NAME + unset SERVICE_NAME + unset JOB_NUMBER + unset JAEGER_COLLECTOR_ENDPOINT + render_find_service agent-as-sidecar2 allInOne order 01 06 + '[' 5 -ne 5 ']' + jaeger=agent-as-sidecar2 + deployment_strategy=allInOne + service_name=order + job_number=01 + test_step=06 + export JAEGER_NAME=agent-as-sidecar2 + JAEGER_NAME=agent-as-sidecar2 + export JOB_NUMBER=01 + JOB_NUMBER=01 + export SERVICE_NAME=order + SERVICE_NAME=order + export JAEGER_QUERY_ENDPOINT + '[' true = true ']' + '[' allInOne '!=' allInOne ']' + template=/tmp/jaeger-tests/tests/templates/find-service.yaml.template + JAEGER_QUERY_ENDPOINT=http://agent-as-sidecar2-query:16686 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/find-service.yaml.template -o ./06-find-service.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-find-service.yaml.template -o ./06-assert.yaml + unset JAEGER_NAME + unset SERVICE_NAME + unset JOB_NUMBER + unset JAEGER_COLLECTOR_ENDPOINT + start_test sidecar-namespace + '[' 1 -ne 1 ']' + test_name=sidecar-namespace + echo =========================================================================== =========================================================================== + info 'Rendering files for test sidecar-namespace' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test sidecar-namespace\e[0m' Rendering files for test sidecar-namespace + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/sidecar/_build/sidecar-deployment + '[' sidecar-deployment '!=' _build ']' + cd .. + mkdir -p sidecar-namespace + cd sidecar-namespace + render_install_vertx 01 + '[' 1 -ne 1 ']' + test_step=01 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-assert.yaml.template -o ./01-assert.yaml + render_find_service agent-as-sidecar allInOne order 00 03 + '[' 5 -ne 5 ']' + jaeger=agent-as-sidecar + deployment_strategy=allInOne + service_name=order + job_number=00 + test_step=03 + export JAEGER_NAME=agent-as-sidecar + JAEGER_NAME=agent-as-sidecar + export JOB_NUMBER=00 + JOB_NUMBER=00 + export SERVICE_NAME=order + SERVICE_NAME=order + export JAEGER_QUERY_ENDPOINT + '[' true = true ']' + '[' allInOne '!=' allInOne ']' + template=/tmp/jaeger-tests/tests/templates/find-service.yaml.template + JAEGER_QUERY_ENDPOINT=http://agent-as-sidecar-query:16686 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/find-service.yaml.template -o ./03-find-service.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-find-service.yaml.template -o ./03-assert.yaml + unset JAEGER_NAME + unset SERVICE_NAME + unset JOB_NUMBER + unset JAEGER_COLLECTOR_ENDPOINT + render_find_service agent-as-sidecar2 allInOne order 01 06 + '[' 5 -ne 5 ']' + jaeger=agent-as-sidecar2 + deployment_strategy=allInOne + service_name=order + job_number=01 + test_step=06 + export JAEGER_NAME=agent-as-sidecar2 + JAEGER_NAME=agent-as-sidecar2 + export JOB_NUMBER=01 + JOB_NUMBER=01 + export SERVICE_NAME=order + SERVICE_NAME=order + export JAEGER_QUERY_ENDPOINT + '[' true = true ']' + '[' allInOne '!=' allInOne ']' + template=/tmp/jaeger-tests/tests/templates/find-service.yaml.template + JAEGER_QUERY_ENDPOINT=http://agent-as-sidecar2-query:16686 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/find-service.yaml.template -o ./06-find-service.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-find-service.yaml.template -o ./06-assert.yaml + unset JAEGER_NAME + unset SERVICE_NAME + unset JOB_NUMBER + unset JAEGER_COLLECTOR_ENDPOINT + start_test sidecar-skip-webhook + '[' 1 -ne 1 ']' + test_name=sidecar-skip-webhook + echo =========================================================================== =========================================================================== + info 'Rendering files for test sidecar-skip-webhook' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test sidecar-skip-webhook\e[0m' Rendering files for test sidecar-skip-webhook + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/sidecar/_build/sidecar-namespace + '[' sidecar-namespace '!=' _build ']' + cd .. + mkdir -p sidecar-skip-webhook + cd sidecar-skip-webhook + render_install_vertx 01 + '[' 1 -ne 1 ']' + test_step=01 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-assert.yaml.template -o ./01-assert.yaml make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running sidecar E2E tests' Running sidecar E2E tests + cd tests/e2e/sidecar/_build + set +e + KUBECONFIG=/tmp/kubeconfig-2030369819 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-rosa-h-yd3i.24xa.s3.devshift.org:443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 4 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/sidecar-deployment === PAUSE kuttl/harness/sidecar-deployment === RUN kuttl/harness/sidecar-namespace === PAUSE kuttl/harness/sidecar-namespace === RUN kuttl/harness/sidecar-skip-webhook === PAUSE kuttl/harness/sidecar-skip-webhook === CONT kuttl/harness/artifacts logger.go:42: 20:26:15 | artifacts | Creating namespace: kuttl-test-sensible-mongoose logger.go:42: 20:26:15 | artifacts | artifacts events from ns kuttl-test-sensible-mongoose: logger.go:42: 20:26:15 | artifacts | Deleting namespace: kuttl-test-sensible-mongoose === CONT kuttl/harness/sidecar-namespace logger.go:42: 20:26:21 | sidecar-namespace | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 20:26:21 | sidecar-namespace | Creating namespace: kuttl-test-smiling-kid logger.go:42: 20:26:21 | sidecar-namespace/0-install | starting test step 0-install logger.go:42: 20:26:21 | sidecar-namespace/0-install | Jaeger:kuttl-test-smiling-kid/agent-as-sidecar created logger.go:42: 20:26:25 | sidecar-namespace/0-install | test step completed 0-install logger.go:42: 20:26:25 | sidecar-namespace/1-install | starting test step 1-install logger.go:42: 20:26:25 | sidecar-namespace/1-install | Deployment:kuttl-test-smiling-kid/vertx-create-span-sidecar created logger.go:42: 20:26:26 | sidecar-namespace/1-install | test step completed 1-install logger.go:42: 20:26:26 | sidecar-namespace/2-enable-injection | starting test step 2-enable-injection logger.go:42: 20:26:26 | sidecar-namespace/2-enable-injection | running command: [sh -c kubectl annotate --overwrite namespaces $NAMESPACE "sidecar.jaegertracing.io/inject"="true"] logger.go:42: 20:26:27 | sidecar-namespace/2-enable-injection | namespace/kuttl-test-smiling-kid annotated logger.go:42: 20:26:28 | sidecar-namespace/2-enable-injection | test step completed 2-enable-injection logger.go:42: 20:26:28 | sidecar-namespace/3-find-service | starting test step 3-find-service logger.go:42: 20:26:28 | sidecar-namespace/3-find-service | Job:kuttl-test-smiling-kid/00-find-service created logger.go:42: 20:26:40 | sidecar-namespace/3-find-service | test step completed 3-find-service logger.go:42: 20:26:40 | sidecar-namespace/4-other-instance | starting test step 4-other-instance logger.go:42: 20:26:40 | sidecar-namespace/4-other-instance | Jaeger:kuttl-test-smiling-kid/agent-as-sidecar2 created logger.go:42: 20:26:44 | sidecar-namespace/4-other-instance | test step completed 4-other-instance logger.go:42: 20:26:44 | sidecar-namespace/5-delete-first-instance | starting test step 5-delete-first-instance logger.go:42: 20:26:44 | sidecar-namespace/5-delete-first-instance | test step completed 5-delete-first-instance logger.go:42: 20:26:44 | sidecar-namespace/6-find-service | starting test step 6-find-service logger.go:42: 20:26:45 | sidecar-namespace/6-find-service | Job:kuttl-test-smiling-kid/01-find-service created logger.go:42: 20:26:57 | sidecar-namespace/6-find-service | test step completed 6-find-service logger.go:42: 20:26:57 | sidecar-namespace/7-disable-injection | starting test step 7-disable-injection logger.go:42: 20:26:57 | sidecar-namespace/7-disable-injection | running command: [sh -c kubectl annotate --overwrite namespaces $NAMESPACE "sidecar.jaegertracing.io/inject"="false"] logger.go:42: 20:26:58 | sidecar-namespace/7-disable-injection | namespace/kuttl-test-smiling-kid annotated logger.go:42: 20:26:59 | sidecar-namespace/7-disable-injection | test step completed 7-disable-injection logger.go:42: 20:26:59 | sidecar-namespace | sidecar-namespace events from ns kuttl-test-smiling-kid: logger.go:42: 20:26:59 | sidecar-namespace | 2024-08-29 20:26:22 +0000 UTC Normal Pod agent-as-sidecar-54c6697f5f-bz4mf Binding Scheduled Successfully assigned kuttl-test-smiling-kid/agent-as-sidecar-54c6697f5f-bz4mf to ip-10-0-51-237.us-west-2.compute.internal default-scheduler logger.go:42: 20:26:59 | sidecar-namespace | 2024-08-29 20:26:22 +0000 UTC Normal Pod agent-as-sidecar-54c6697f5f-bz4mf AddedInterface Add eth0 [10.130.0.67/23] from ovn-kubernetes multus logger.go:42: 20:26:59 | sidecar-namespace | 2024-08-29 20:26:22 +0000 UTC Normal Pod agent-as-sidecar-54c6697f5f-bz4mf.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:345d3b6c507df9f8a18948a594e7a03698085689c6965faf79b8aadc74a0ceaa" already present on machine kubelet logger.go:42: 20:26:59 | sidecar-namespace | 2024-08-29 20:26:22 +0000 UTC Normal Pod agent-as-sidecar-54c6697f5f-bz4mf.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 20:26:59 | sidecar-namespace | 2024-08-29 20:26:22 +0000 UTC Normal Pod agent-as-sidecar-54c6697f5f-bz4mf.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 20:26:59 | sidecar-namespace | 2024-08-29 20:26:22 +0000 UTC Normal ReplicaSet.apps agent-as-sidecar-54c6697f5f SuccessfulCreate Created pod: agent-as-sidecar-54c6697f5f-bz4mf replicaset-controller logger.go:42: 20:26:59 | sidecar-namespace | 2024-08-29 20:26:22 +0000 UTC Normal Deployment.apps agent-as-sidecar ScalingReplicaSet Scaled up replica set agent-as-sidecar-54c6697f5f to 1 deployment-controller logger.go:42: 20:26:59 | sidecar-namespace | 2024-08-29 20:26:25 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-wsxgf Binding Scheduled Successfully assigned kuttl-test-smiling-kid/vertx-create-span-sidecar-84d458b68c-wsxgf to ip-10-0-51-237.us-west-2.compute.internal default-scheduler logger.go:42: 20:26:59 | sidecar-namespace | 2024-08-29 20:26:25 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-wsxgf AddedInterface Add eth0 [10.130.0.68/23] from ovn-kubernetes multus logger.go:42: 20:26:59 | sidecar-namespace | 2024-08-29 20:26:25 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-wsxgf.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 20:26:59 | sidecar-namespace | 2024-08-29 20:26:25 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-84d458b68c SuccessfulCreate Created pod: vertx-create-span-sidecar-84d458b68c-wsxgf replicaset-controller logger.go:42: 20:26:59 | sidecar-namespace | 2024-08-29 20:26:25 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-84d458b68c to 1 deployment-controller logger.go:42: 20:26:59 | sidecar-namespace | 2024-08-29 20:26:26 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-wsxgf.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 20:26:59 | sidecar-namespace | 2024-08-29 20:26:26 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-wsxgf.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 20:26:59 | sidecar-namespace | 2024-08-29 20:26:27 +0000 UTC Normal Pod vertx-create-span-sidecar-6796bfbf45-t8lfp Binding Scheduled Successfully assigned kuttl-test-smiling-kid/vertx-create-span-sidecar-6796bfbf45-t8lfp to ip-10-0-51-237.us-west-2.compute.internal default-scheduler logger.go:42: 20:26:59 | sidecar-namespace | 2024-08-29 20:26:27 +0000 UTC Normal Pod vertx-create-span-sidecar-6796bfbf45-t8lfp AddedInterface Add eth0 [10.130.0.69/23] from ovn-kubernetes multus logger.go:42: 20:26:59 | sidecar-namespace | 2024-08-29 20:26:27 +0000 UTC Normal Pod vertx-create-span-sidecar-6796bfbf45-t8lfp.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 20:26:59 | sidecar-namespace | 2024-08-29 20:26:27 +0000 UTC Normal Pod vertx-create-span-sidecar-6796bfbf45-t8lfp.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 20:26:59 | sidecar-namespace | 2024-08-29 20:26:27 +0000 UTC Normal Pod vertx-create-span-sidecar-6796bfbf45-t8lfp.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 20:26:59 | sidecar-namespace | 2024-08-29 20:26:27 +0000 UTC Normal Pod vertx-create-span-sidecar-6796bfbf45-t8lfp.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:5fcc9d365678f54c0f52005318dbbc32b87f743cae7e1a84568e769653f3b128" already present on machine kubelet logger.go:42: 20:26:59 | sidecar-namespace | 2024-08-29 20:26:27 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-6796bfbf45 SuccessfulCreate Created pod: vertx-create-span-sidecar-6796bfbf45-t8lfp replicaset-controller logger.go:42: 20:26:59 | sidecar-namespace | 2024-08-29 20:26:27 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-6796bfbf45 to 1 deployment-controller logger.go:42: 20:26:59 | sidecar-namespace | 2024-08-29 20:26:28 +0000 UTC Normal Pod 00-find-service-tc87r Binding Scheduled Successfully assigned kuttl-test-smiling-kid/00-find-service-tc87r to ip-10-0-51-237.us-west-2.compute.internal default-scheduler logger.go:42: 20:26:59 | sidecar-namespace | 2024-08-29 20:26:28 +0000 UTC Normal Job.batch 00-find-service SuccessfulCreate Created pod: 00-find-service-tc87r job-controller logger.go:42: 20:26:59 | sidecar-namespace | 2024-08-29 20:26:28 +0000 UTC Normal Pod vertx-create-span-sidecar-6796bfbf45-t8lfp.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 20:26:59 | sidecar-namespace | 2024-08-29 20:26:28 +0000 UTC Normal Pod vertx-create-span-sidecar-6796bfbf45-t8lfp.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 20:26:59 | sidecar-namespace | 2024-08-29 20:26:29 +0000 UTC Normal Pod 00-find-service-tc87r AddedInterface Add eth0 [10.130.0.70/23] from ovn-kubernetes multus logger.go:42: 20:26:59 | sidecar-namespace | 2024-08-29 20:26:29 +0000 UTC Normal Pod 00-find-service-tc87r.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 20:26:59 | sidecar-namespace | 2024-08-29 20:26:30 +0000 UTC Normal Pod 00-find-service-tc87r.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 823ms (823ms including waiting) kubelet logger.go:42: 20:26:59 | sidecar-namespace | 2024-08-29 20:26:30 +0000 UTC Normal Pod 00-find-service-tc87r.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 20:26:59 | sidecar-namespace | 2024-08-29 20:26:30 +0000 UTC Normal Pod 00-find-service-tc87r.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 20:26:59 | sidecar-namespace | 2024-08-29 20:26:34 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-wsxgf.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.130.0.68:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 20:26:59 | sidecar-namespace | 2024-08-29 20:26:34 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-wsxgf.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.130.0.68:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 20:26:59 | sidecar-namespace | 2024-08-29 20:26:35 +0000 UTC Warning Pod vertx-create-span-sidecar-6796bfbf45-t8lfp.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.130.0.69:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 20:26:59 | sidecar-namespace | 2024-08-29 20:26:35 +0000 UTC Warning Pod vertx-create-span-sidecar-6796bfbf45-t8lfp.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.130.0.69:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 20:26:59 | sidecar-namespace | 2024-08-29 20:26:37 +0000 UTC Normal Pod vertx-create-span-sidecar-6796bfbf45-t8lfp.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 20:26:59 | sidecar-namespace | 2024-08-29 20:26:37 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-wsxgf.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 20:26:59 | sidecar-namespace | 2024-08-29 20:26:37 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-wsxgf.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.130.0.68:8080/": read tcp 10.130.0.2:54904->10.130.0.68:8080: read: connection reset by peer kubelet logger.go:42: 20:26:59 | sidecar-namespace | 2024-08-29 20:26:37 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-wsxgf.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.130.0.68:8080/": dial tcp 10.130.0.68:8080: connect: connection refused kubelet logger.go:42: 20:26:59 | sidecar-namespace | 2024-08-29 20:26:38 +0000 UTC Warning Pod vertx-create-span-sidecar-6796bfbf45-t8lfp.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.130.0.69:8080/": read tcp 10.130.0.2:33434->10.130.0.69:8080: read: connection reset by peer kubelet logger.go:42: 20:26:59 | sidecar-namespace | 2024-08-29 20:26:38 +0000 UTC Warning Pod vertx-create-span-sidecar-6796bfbf45-t8lfp.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.130.0.69:8080/": dial tcp 10.130.0.69:8080: connect: connection refused kubelet logger.go:42: 20:26:59 | sidecar-namespace | 2024-08-29 20:26:40 +0000 UTC Normal Job.batch 00-find-service Completed Job completed job-controller logger.go:42: 20:26:59 | sidecar-namespace | 2024-08-29 20:26:42 +0000 UTC Normal Pod agent-as-sidecar2-754bcb9bbf-qhn8l Binding Scheduled Successfully assigned kuttl-test-smiling-kid/agent-as-sidecar2-754bcb9bbf-qhn8l to ip-10-0-51-237.us-west-2.compute.internal default-scheduler logger.go:42: 20:26:59 | sidecar-namespace | 2024-08-29 20:26:42 +0000 UTC Normal Pod agent-as-sidecar2-754bcb9bbf-qhn8l AddedInterface Add eth0 [10.130.0.71/23] from ovn-kubernetes multus logger.go:42: 20:26:59 | sidecar-namespace | 2024-08-29 20:26:42 +0000 UTC Normal Pod agent-as-sidecar2-754bcb9bbf-qhn8l.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:345d3b6c507df9f8a18948a594e7a03698085689c6965faf79b8aadc74a0ceaa" already present on machine kubelet logger.go:42: 20:26:59 | sidecar-namespace | 2024-08-29 20:26:42 +0000 UTC Normal Pod agent-as-sidecar2-754bcb9bbf-qhn8l.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 20:26:59 | sidecar-namespace | 2024-08-29 20:26:42 +0000 UTC Normal Pod agent-as-sidecar2-754bcb9bbf-qhn8l.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 20:26:59 | sidecar-namespace | 2024-08-29 20:26:42 +0000 UTC Normal ReplicaSet.apps agent-as-sidecar2-754bcb9bbf SuccessfulCreate Created pod: agent-as-sidecar2-754bcb9bbf-qhn8l replicaset-controller logger.go:42: 20:26:59 | sidecar-namespace | 2024-08-29 20:26:42 +0000 UTC Normal Deployment.apps agent-as-sidecar2 ScalingReplicaSet Scaled up replica set agent-as-sidecar2-754bcb9bbf to 1 deployment-controller logger.go:42: 20:26:59 | sidecar-namespace | 2024-08-29 20:26:44 +0000 UTC Normal Pod agent-as-sidecar-54c6697f5f-bz4mf.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 20:26:59 | sidecar-namespace | 2024-08-29 20:26:44 +0000 UTC Normal Pod vertx-create-span-sidecar-7b9b8dcfdc-9lj6r Binding Scheduled Successfully assigned kuttl-test-smiling-kid/vertx-create-span-sidecar-7b9b8dcfdc-9lj6r to ip-10-0-51-237.us-west-2.compute.internal default-scheduler logger.go:42: 20:26:59 | sidecar-namespace | 2024-08-29 20:26:44 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-7b9b8dcfdc SuccessfulCreate Created pod: vertx-create-span-sidecar-7b9b8dcfdc-9lj6r replicaset-controller logger.go:42: 20:26:59 | sidecar-namespace | 2024-08-29 20:26:44 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-wsxgf.spec.containers{vertx-create-span-sidecar} Killing Stopping container vertx-create-span-sidecar kubelet logger.go:42: 20:26:59 | sidecar-namespace | 2024-08-29 20:26:44 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-84d458b68c SuccessfulDelete Deleted pod: vertx-create-span-sidecar-84d458b68c-wsxgf replicaset-controller logger.go:42: 20:26:59 | sidecar-namespace | 2024-08-29 20:26:44 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled down replica set vertx-create-span-sidecar-84d458b68c to 0 from 1 deployment-controller logger.go:42: 20:26:59 | sidecar-namespace | 2024-08-29 20:26:44 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-7b9b8dcfdc to 1 from 0 deployment-controller logger.go:42: 20:26:59 | sidecar-namespace | 2024-08-29 20:26:45 +0000 UTC Normal Pod 01-find-service-49ftq Binding Scheduled Successfully assigned kuttl-test-smiling-kid/01-find-service-49ftq to ip-10-0-57-246.us-west-2.compute.internal default-scheduler logger.go:42: 20:26:59 | sidecar-namespace | 2024-08-29 20:26:45 +0000 UTC Normal Pod 01-find-service-49ftq AddedInterface Add eth0 [10.129.0.33/23] from ovn-kubernetes multus logger.go:42: 20:26:59 | sidecar-namespace | 2024-08-29 20:26:45 +0000 UTC Normal Pod 01-find-service-49ftq.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 20:26:59 | sidecar-namespace | 2024-08-29 20:26:45 +0000 UTC Normal Job.batch 01-find-service SuccessfulCreate Created pod: 01-find-service-49ftq job-controller logger.go:42: 20:26:59 | sidecar-namespace | 2024-08-29 20:26:45 +0000 UTC Normal Pod vertx-create-span-sidecar-7b9b8dcfdc-9lj6r AddedInterface Add eth0 [10.130.0.72/23] from ovn-kubernetes multus logger.go:42: 20:26:59 | sidecar-namespace | 2024-08-29 20:26:45 +0000 UTC Normal Pod vertx-create-span-sidecar-7b9b8dcfdc-9lj6r.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 20:26:59 | sidecar-namespace | 2024-08-29 20:26:45 +0000 UTC Normal Pod vertx-create-span-sidecar-7b9b8dcfdc-9lj6r.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 20:26:59 | sidecar-namespace | 2024-08-29 20:26:45 +0000 UTC Normal Pod vertx-create-span-sidecar-7b9b8dcfdc-9lj6r.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 20:26:59 | sidecar-namespace | 2024-08-29 20:26:45 +0000 UTC Normal Pod vertx-create-span-sidecar-7b9b8dcfdc-9lj6r.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:5fcc9d365678f54c0f52005318dbbc32b87f743cae7e1a84568e769653f3b128" already present on machine kubelet logger.go:42: 20:26:59 | sidecar-namespace | 2024-08-29 20:26:45 +0000 UTC Normal Pod vertx-create-span-sidecar-7b9b8dcfdc-9lj6r.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 20:26:59 | sidecar-namespace | 2024-08-29 20:26:45 +0000 UTC Normal Pod vertx-create-span-sidecar-7b9b8dcfdc-9lj6r.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 20:26:59 | sidecar-namespace | 2024-08-29 20:26:45 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-wsxgf.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.130.0.68:8080/": read tcp 10.130.0.2:55732->10.130.0.68:8080: read: connection reset by peer kubelet logger.go:42: 20:26:59 | sidecar-namespace | 2024-08-29 20:26:46 +0000 UTC Normal Pod 01-find-service-49ftq.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 810ms (810ms including waiting) kubelet logger.go:42: 20:26:59 | sidecar-namespace | 2024-08-29 20:26:46 +0000 UTC Normal Pod 01-find-service-49ftq.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 20:26:59 | sidecar-namespace | 2024-08-29 20:26:46 +0000 UTC Normal Pod 01-find-service-49ftq.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 20:26:59 | sidecar-namespace | 2024-08-29 20:26:53 +0000 UTC Warning Pod vertx-create-span-sidecar-7b9b8dcfdc-9lj6r.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.130.0.72:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 20:26:59 | sidecar-namespace | 2024-08-29 20:26:53 +0000 UTC Warning Pod vertx-create-span-sidecar-7b9b8dcfdc-9lj6r.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.130.0.72:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 20:26:59 | sidecar-namespace | 2024-08-29 20:26:55 +0000 UTC Normal Pod vertx-create-span-sidecar-7b9b8dcfdc-9lj6r.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 20:26:59 | sidecar-namespace | 2024-08-29 20:26:55 +0000 UTC Warning Pod vertx-create-span-sidecar-7b9b8dcfdc-9lj6r.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.130.0.72:8080/": read tcp 10.130.0.2:47968->10.130.0.72:8080: read: connection reset by peer kubelet logger.go:42: 20:26:59 | sidecar-namespace | 2024-08-29 20:26:55 +0000 UTC Warning Pod vertx-create-span-sidecar-7b9b8dcfdc-9lj6r.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.130.0.72:8080/": dial tcp 10.130.0.72:8080: connect: connection refused kubelet logger.go:42: 20:26:59 | sidecar-namespace | 2024-08-29 20:26:56 +0000 UTC Normal Job.batch 01-find-service Completed Job completed job-controller logger.go:42: 20:26:59 | sidecar-namespace | 2024-08-29 20:26:58 +0000 UTC Normal Pod vertx-create-span-sidecar-574d94bb9d-v658t Binding Scheduled Successfully assigned kuttl-test-smiling-kid/vertx-create-span-sidecar-574d94bb9d-v658t to ip-10-0-51-237.us-west-2.compute.internal default-scheduler logger.go:42: 20:26:59 | sidecar-namespace | 2024-08-29 20:26:58 +0000 UTC Normal Pod vertx-create-span-sidecar-574d94bb9d-v658t AddedInterface Add eth0 [10.130.0.73/23] from ovn-kubernetes multus logger.go:42: 20:26:59 | sidecar-namespace | 2024-08-29 20:26:58 +0000 UTC Normal Pod vertx-create-span-sidecar-574d94bb9d-v658t.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 20:26:59 | sidecar-namespace | 2024-08-29 20:26:58 +0000 UTC Normal Pod vertx-create-span-sidecar-574d94bb9d-v658t.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 20:26:59 | sidecar-namespace | 2024-08-29 20:26:58 +0000 UTC Normal Pod vertx-create-span-sidecar-574d94bb9d-v658t.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 20:26:59 | sidecar-namespace | 2024-08-29 20:26:58 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-574d94bb9d SuccessfulCreate Created pod: vertx-create-span-sidecar-574d94bb9d-v658t replicaset-controller logger.go:42: 20:26:59 | sidecar-namespace | 2024-08-29 20:26:58 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-6796bfbf45 SuccessfulDelete Deleted pod: vertx-create-span-sidecar-6796bfbf45-t8lfp replicaset-controller logger.go:42: 20:26:59 | sidecar-namespace | 2024-08-29 20:26:58 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled down replica set vertx-create-span-sidecar-6796bfbf45 to 0 from 1 deployment-controller logger.go:42: 20:26:59 | sidecar-namespace | 2024-08-29 20:26:58 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-574d94bb9d to 1 from 0 deployment-controller logger.go:42: 20:26:59 | sidecar-namespace | Deleting namespace: kuttl-test-smiling-kid === CONT kuttl/harness/sidecar-skip-webhook logger.go:42: 20:27:06 | sidecar-skip-webhook | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 20:27:06 | sidecar-skip-webhook | Creating namespace: kuttl-test-generous-elephant logger.go:42: 20:27:06 | sidecar-skip-webhook/0-install | starting test step 0-install logger.go:42: 20:27:07 | sidecar-skip-webhook/0-install | Jaeger:kuttl-test-generous-elephant/agent-as-sidecar created logger.go:42: 20:27:10 | sidecar-skip-webhook/0-install | test step completed 0-install logger.go:42: 20:27:10 | sidecar-skip-webhook/1-install | starting test step 1-install logger.go:42: 20:27:10 | sidecar-skip-webhook/1-install | Deployment:kuttl-test-generous-elephant/vertx-create-span-sidecar created logger.go:42: 20:27:11 | sidecar-skip-webhook/1-install | test step completed 1-install logger.go:42: 20:27:11 | sidecar-skip-webhook/2-add-anotation-and-label | starting test step 2-add-anotation-and-label logger.go:42: 20:27:11 | sidecar-skip-webhook/2-add-anotation-and-label | running command: [kubectl label deployment vertx-create-span-sidecar app.kubernetes.io/name=jaeger-operator --namespace kuttl-test-generous-elephant] logger.go:42: 20:27:12 | sidecar-skip-webhook/2-add-anotation-and-label | deployment.apps/vertx-create-span-sidecar labeled logger.go:42: 20:27:12 | sidecar-skip-webhook/2-add-anotation-and-label | running command: [kubectl annotate --overwrite deployment vertx-create-span-sidecar sidecar.jaegertracing.io/inject=true --namespace kuttl-test-generous-elephant] logger.go:42: 20:27:12 | sidecar-skip-webhook/2-add-anotation-and-label | deployment.apps/vertx-create-span-sidecar annotated logger.go:42: 20:27:12 | sidecar-skip-webhook/2-add-anotation-and-label | test step completed 2-add-anotation-and-label logger.go:42: 20:27:12 | sidecar-skip-webhook/3-remove-label | starting test step 3-remove-label logger.go:42: 20:27:12 | sidecar-skip-webhook/3-remove-label | running command: [kubectl label deployment vertx-create-span-sidecar app.kubernetes.io/name- --namespace kuttl-test-generous-elephant] logger.go:42: 20:27:13 | sidecar-skip-webhook/3-remove-label | deployment.apps/vertx-create-span-sidecar unlabeled logger.go:42: 20:27:14 | sidecar-skip-webhook/3-remove-label | test step completed 3-remove-label logger.go:42: 20:27:14 | sidecar-skip-webhook | sidecar-skip-webhook events from ns kuttl-test-generous-elephant: logger.go:42: 20:27:14 | sidecar-skip-webhook | 2024-08-29 20:27:07 +0000 UTC Normal Pod agent-as-sidecar-666c6f5cbf-pkbkq Binding Scheduled Successfully assigned kuttl-test-generous-elephant/agent-as-sidecar-666c6f5cbf-pkbkq to ip-10-0-51-237.us-west-2.compute.internal default-scheduler logger.go:42: 20:27:14 | sidecar-skip-webhook | 2024-08-29 20:27:07 +0000 UTC Normal Pod agent-as-sidecar-666c6f5cbf-pkbkq AddedInterface Add eth0 [10.130.0.74/23] from ovn-kubernetes multus logger.go:42: 20:27:14 | sidecar-skip-webhook | 2024-08-29 20:27:07 +0000 UTC Normal Pod agent-as-sidecar-666c6f5cbf-pkbkq.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:345d3b6c507df9f8a18948a594e7a03698085689c6965faf79b8aadc74a0ceaa" already present on machine kubelet logger.go:42: 20:27:14 | sidecar-skip-webhook | 2024-08-29 20:27:07 +0000 UTC Normal ReplicaSet.apps agent-as-sidecar-666c6f5cbf SuccessfulCreate Created pod: agent-as-sidecar-666c6f5cbf-pkbkq replicaset-controller logger.go:42: 20:27:14 | sidecar-skip-webhook | 2024-08-29 20:27:07 +0000 UTC Normal Deployment.apps agent-as-sidecar ScalingReplicaSet Scaled up replica set agent-as-sidecar-666c6f5cbf to 1 deployment-controller logger.go:42: 20:27:14 | sidecar-skip-webhook | 2024-08-29 20:27:08 +0000 UTC Normal Pod agent-as-sidecar-666c6f5cbf-pkbkq.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 20:27:14 | sidecar-skip-webhook | 2024-08-29 20:27:08 +0000 UTC Normal Pod agent-as-sidecar-666c6f5cbf-pkbkq.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 20:27:14 | sidecar-skip-webhook | 2024-08-29 20:27:10 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-f4v2x Binding Scheduled Successfully assigned kuttl-test-generous-elephant/vertx-create-span-sidecar-84d458b68c-f4v2x to ip-10-0-51-237.us-west-2.compute.internal default-scheduler logger.go:42: 20:27:14 | sidecar-skip-webhook | 2024-08-29 20:27:10 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-84d458b68c SuccessfulCreate Created pod: vertx-create-span-sidecar-84d458b68c-f4v2x replicaset-controller logger.go:42: 20:27:14 | sidecar-skip-webhook | 2024-08-29 20:27:10 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-84d458b68c to 1 deployment-controller logger.go:42: 20:27:14 | sidecar-skip-webhook | 2024-08-29 20:27:11 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-f4v2x AddedInterface Add eth0 [10.130.0.75/23] from ovn-kubernetes multus logger.go:42: 20:27:14 | sidecar-skip-webhook | 2024-08-29 20:27:11 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-f4v2x.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 20:27:14 | sidecar-skip-webhook | 2024-08-29 20:27:11 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-f4v2x.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 20:27:14 | sidecar-skip-webhook | 2024-08-29 20:27:11 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-f4v2x.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 20:27:14 | sidecar-skip-webhook | 2024-08-29 20:27:12 +0000 UTC Normal Pod vertx-create-span-sidecar-5b4975c67-rv6nj Binding Scheduled Successfully assigned kuttl-test-generous-elephant/vertx-create-span-sidecar-5b4975c67-rv6nj to ip-10-0-51-237.us-west-2.compute.internal default-scheduler logger.go:42: 20:27:14 | sidecar-skip-webhook | 2024-08-29 20:27:12 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-5b4975c67 SuccessfulCreate Created pod: vertx-create-span-sidecar-5b4975c67-rv6nj replicaset-controller logger.go:42: 20:27:14 | sidecar-skip-webhook | 2024-08-29 20:27:12 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-5b4975c67 to 1 deployment-controller logger.go:42: 20:27:14 | sidecar-skip-webhook | 2024-08-29 20:27:13 +0000 UTC Normal Pod vertx-create-span-sidecar-5b4975c67-rv6nj AddedInterface Add eth0 [10.130.0.76/23] from ovn-kubernetes multus logger.go:42: 20:27:14 | sidecar-skip-webhook | 2024-08-29 20:27:13 +0000 UTC Normal Pod vertx-create-span-sidecar-5b4975c67-rv6nj.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 20:27:14 | sidecar-skip-webhook | 2024-08-29 20:27:13 +0000 UTC Normal Pod vertx-create-span-sidecar-5b4975c67-rv6nj.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 20:27:14 | sidecar-skip-webhook | 2024-08-29 20:27:13 +0000 UTC Normal Pod vertx-create-span-sidecar-5b4975c67-rv6nj.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 20:27:14 | sidecar-skip-webhook | 2024-08-29 20:27:13 +0000 UTC Normal Pod vertx-create-span-sidecar-5b4975c67-rv6nj.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:5fcc9d365678f54c0f52005318dbbc32b87f743cae7e1a84568e769653f3b128" already present on machine kubelet logger.go:42: 20:27:14 | sidecar-skip-webhook | 2024-08-29 20:27:13 +0000 UTC Normal Pod vertx-create-span-sidecar-5b4975c67-rv6nj.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 20:27:14 | sidecar-skip-webhook | 2024-08-29 20:27:13 +0000 UTC Normal Pod vertx-create-span-sidecar-5b4975c67-rv6nj.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 20:27:14 | sidecar-skip-webhook | Deleting namespace: kuttl-test-generous-elephant === CONT kuttl/harness/sidecar-deployment logger.go:42: 20:27:21 | sidecar-deployment | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 20:27:21 | sidecar-deployment | Creating namespace: kuttl-test-smart-burro logger.go:42: 20:27:21 | sidecar-deployment/0-install | starting test step 0-install logger.go:42: 20:27:21 | sidecar-deployment/0-install | Jaeger:kuttl-test-smart-burro/agent-as-sidecar created logger.go:42: 20:27:24 | sidecar-deployment/0-install | test step completed 0-install logger.go:42: 20:27:24 | sidecar-deployment/1-install | starting test step 1-install logger.go:42: 20:27:25 | sidecar-deployment/1-install | Deployment:kuttl-test-smart-burro/vertx-create-span-sidecar created logger.go:42: 20:27:27 | sidecar-deployment/1-install | test step completed 1-install logger.go:42: 20:27:27 | sidecar-deployment/2-enable-injection | starting test step 2-enable-injection logger.go:42: 20:27:27 | sidecar-deployment/2-enable-injection | running command: [kubectl annotate --overwrite deployment vertx-create-span-sidecar sidecar.jaegertracing.io/inject=true --namespace kuttl-test-smart-burro] logger.go:42: 20:27:27 | sidecar-deployment/2-enable-injection | deployment.apps/vertx-create-span-sidecar annotated logger.go:42: 20:27:30 | sidecar-deployment/2-enable-injection | test step completed 2-enable-injection logger.go:42: 20:27:30 | sidecar-deployment/3-find-service | starting test step 3-find-service logger.go:42: 20:27:30 | sidecar-deployment/3-find-service | Job:kuttl-test-smart-burro/00-find-service created logger.go:42: 20:27:42 | sidecar-deployment/3-find-service | test step completed 3-find-service logger.go:42: 20:27:42 | sidecar-deployment/4-other-instance | starting test step 4-other-instance logger.go:42: 20:27:42 | sidecar-deployment/4-other-instance | Jaeger:kuttl-test-smart-burro/agent-as-sidecar2 created logger.go:42: 20:27:46 | sidecar-deployment/4-other-instance | test step completed 4-other-instance logger.go:42: 20:27:46 | sidecar-deployment/5-delete-first-instance | starting test step 5-delete-first-instance logger.go:42: 20:27:46 | sidecar-deployment/5-delete-first-instance | test step completed 5-delete-first-instance logger.go:42: 20:27:46 | sidecar-deployment/6-find-service | starting test step 6-find-service logger.go:42: 20:27:47 | sidecar-deployment/6-find-service | Job:kuttl-test-smart-burro/01-find-service created logger.go:42: 20:28:07 | sidecar-deployment/6-find-service | test step completed 6-find-service logger.go:42: 20:28:07 | sidecar-deployment/7-disable-injection | starting test step 7-disable-injection logger.go:42: 20:28:07 | sidecar-deployment/7-disable-injection | running command: [kubectl annotate --overwrite deployment vertx-create-span-sidecar sidecar.jaegertracing.io/inject=false --namespace kuttl-test-smart-burro] logger.go:42: 20:28:07 | sidecar-deployment/7-disable-injection | deployment.apps/vertx-create-span-sidecar annotated logger.go:42: 20:28:14 | sidecar-deployment/7-disable-injection | test step completed 7-disable-injection logger.go:42: 20:28:15 | sidecar-deployment | sidecar-deployment events from ns kuttl-test-smart-burro: logger.go:42: 20:28:15 | sidecar-deployment | 2024-08-29 20:27:21 +0000 UTC Normal Deployment.apps agent-as-sidecar ScalingReplicaSet Scaled up replica set agent-as-sidecar-7766c49f9 to 1 deployment-controller logger.go:42: 20:28:15 | sidecar-deployment | 2024-08-29 20:27:22 +0000 UTC Normal Pod agent-as-sidecar-7766c49f9-zsr5g Binding Scheduled Successfully assigned kuttl-test-smart-burro/agent-as-sidecar-7766c49f9-zsr5g to ip-10-0-51-237.us-west-2.compute.internal default-scheduler logger.go:42: 20:28:15 | sidecar-deployment | 2024-08-29 20:27:22 +0000 UTC Normal Pod agent-as-sidecar-7766c49f9-zsr5g AddedInterface Add eth0 [10.130.0.77/23] from ovn-kubernetes multus logger.go:42: 20:28:15 | sidecar-deployment | 2024-08-29 20:27:22 +0000 UTC Normal Pod agent-as-sidecar-7766c49f9-zsr5g.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:345d3b6c507df9f8a18948a594e7a03698085689c6965faf79b8aadc74a0ceaa" already present on machine kubelet logger.go:42: 20:28:15 | sidecar-deployment | 2024-08-29 20:27:22 +0000 UTC Normal Pod agent-as-sidecar-7766c49f9-zsr5g.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 20:28:15 | sidecar-deployment | 2024-08-29 20:27:22 +0000 UTC Normal Pod agent-as-sidecar-7766c49f9-zsr5g.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 20:28:15 | sidecar-deployment | 2024-08-29 20:27:22 +0000 UTC Normal ReplicaSet.apps agent-as-sidecar-7766c49f9 SuccessfulCreate Created pod: agent-as-sidecar-7766c49f9-zsr5g replicaset-controller logger.go:42: 20:28:15 | sidecar-deployment | 2024-08-29 20:27:25 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-2ffq7 Binding Scheduled Successfully assigned kuttl-test-smart-burro/vertx-create-span-sidecar-84d458b68c-2ffq7 to ip-10-0-51-237.us-west-2.compute.internal default-scheduler logger.go:42: 20:28:15 | sidecar-deployment | 2024-08-29 20:27:25 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-2ffq7 AddedInterface Add eth0 [10.130.0.78/23] from ovn-kubernetes multus logger.go:42: 20:28:15 | sidecar-deployment | 2024-08-29 20:27:25 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-2ffq7.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 20:28:15 | sidecar-deployment | 2024-08-29 20:27:25 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-2ffq7.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 20:28:15 | sidecar-deployment | 2024-08-29 20:27:25 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-2ffq7.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 20:28:15 | sidecar-deployment | 2024-08-29 20:27:25 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-84d458b68c SuccessfulCreate Created pod: vertx-create-span-sidecar-84d458b68c-2ffq7 replicaset-controller logger.go:42: 20:28:15 | sidecar-deployment | 2024-08-29 20:27:25 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-84d458b68c to 1 deployment-controller logger.go:42: 20:28:15 | sidecar-deployment | 2024-08-29 20:27:27 +0000 UTC Normal Pod vertx-create-span-sidecar-5fcdc696b5-6pvqj Binding Scheduled Successfully assigned kuttl-test-smart-burro/vertx-create-span-sidecar-5fcdc696b5-6pvqj to ip-10-0-51-237.us-west-2.compute.internal default-scheduler logger.go:42: 20:28:15 | sidecar-deployment | 2024-08-29 20:27:27 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-5fcdc696b5 SuccessfulCreate Created pod: vertx-create-span-sidecar-5fcdc696b5-6pvqj replicaset-controller logger.go:42: 20:28:15 | sidecar-deployment | 2024-08-29 20:27:27 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-5fcdc696b5 to 1 deployment-controller logger.go:42: 20:28:15 | sidecar-deployment | 2024-08-29 20:27:28 +0000 UTC Normal Pod vertx-create-span-sidecar-5fcdc696b5-6pvqj AddedInterface Add eth0 [10.130.0.79/23] from ovn-kubernetes multus logger.go:42: 20:28:15 | sidecar-deployment | 2024-08-29 20:27:28 +0000 UTC Normal Pod vertx-create-span-sidecar-5fcdc696b5-6pvqj.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 20:28:15 | sidecar-deployment | 2024-08-29 20:27:28 +0000 UTC Normal Pod vertx-create-span-sidecar-5fcdc696b5-6pvqj.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 20:28:15 | sidecar-deployment | 2024-08-29 20:27:28 +0000 UTC Normal Pod vertx-create-span-sidecar-5fcdc696b5-6pvqj.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 20:28:15 | sidecar-deployment | 2024-08-29 20:27:28 +0000 UTC Normal Pod vertx-create-span-sidecar-5fcdc696b5-6pvqj.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:5fcc9d365678f54c0f52005318dbbc32b87f743cae7e1a84568e769653f3b128" already present on machine kubelet logger.go:42: 20:28:15 | sidecar-deployment | 2024-08-29 20:27:28 +0000 UTC Normal Pod vertx-create-span-sidecar-5fcdc696b5-6pvqj.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 20:28:15 | sidecar-deployment | 2024-08-29 20:27:28 +0000 UTC Normal Pod vertx-create-span-sidecar-5fcdc696b5-6pvqj.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 20:28:15 | sidecar-deployment | 2024-08-29 20:27:30 +0000 UTC Normal Pod 00-find-service-4cmcs Binding Scheduled Successfully assigned kuttl-test-smart-burro/00-find-service-4cmcs to ip-10-0-51-237.us-west-2.compute.internal default-scheduler logger.go:42: 20:28:15 | sidecar-deployment | 2024-08-29 20:27:30 +0000 UTC Normal Job.batch 00-find-service SuccessfulCreate Created pod: 00-find-service-4cmcs job-controller logger.go:42: 20:28:15 | sidecar-deployment | 2024-08-29 20:27:31 +0000 UTC Normal Pod 00-find-service-4cmcs AddedInterface Add eth0 [10.130.0.80/23] from ovn-kubernetes multus logger.go:42: 20:28:15 | sidecar-deployment | 2024-08-29 20:27:31 +0000 UTC Normal Pod 00-find-service-4cmcs.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 20:28:15 | sidecar-deployment | 2024-08-29 20:27:32 +0000 UTC Normal Pod 00-find-service-4cmcs.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 841ms (841ms including waiting) kubelet logger.go:42: 20:28:15 | sidecar-deployment | 2024-08-29 20:27:32 +0000 UTC Normal Pod 00-find-service-4cmcs.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 20:28:15 | sidecar-deployment | 2024-08-29 20:27:32 +0000 UTC Normal Pod 00-find-service-4cmcs.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 20:28:15 | sidecar-deployment | 2024-08-29 20:27:33 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-2ffq7.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.130.0.78:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 20:28:15 | sidecar-deployment | 2024-08-29 20:27:33 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-2ffq7.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.130.0.78:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 20:28:15 | sidecar-deployment | 2024-08-29 20:27:35 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-2ffq7.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 20:28:15 | sidecar-deployment | 2024-08-29 20:27:36 +0000 UTC Warning Pod vertx-create-span-sidecar-5fcdc696b5-6pvqj.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.130.0.79:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 20:28:15 | sidecar-deployment | 2024-08-29 20:27:36 +0000 UTC Warning Pod vertx-create-span-sidecar-5fcdc696b5-6pvqj.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.130.0.79:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 20:28:15 | sidecar-deployment | 2024-08-29 20:27:36 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-2ffq7.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.130.0.78:8080/": read tcp 10.130.0.2:52318->10.130.0.78:8080: read: connection reset by peer kubelet logger.go:42: 20:28:15 | sidecar-deployment | 2024-08-29 20:27:36 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-2ffq7.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.130.0.78:8080/": dial tcp 10.130.0.78:8080: connect: connection refused kubelet logger.go:42: 20:28:15 | sidecar-deployment | 2024-08-29 20:27:38 +0000 UTC Normal Pod vertx-create-span-sidecar-5fcdc696b5-6pvqj.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 20:28:15 | sidecar-deployment | 2024-08-29 20:27:38 +0000 UTC Warning Pod vertx-create-span-sidecar-5fcdc696b5-6pvqj.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.130.0.79:8080/": read tcp 10.130.0.2:56230->10.130.0.79:8080: read: connection reset by peer kubelet logger.go:42: 20:28:15 | sidecar-deployment | 2024-08-29 20:27:38 +0000 UTC Warning Pod vertx-create-span-sidecar-5fcdc696b5-6pvqj.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.130.0.79:8080/": dial tcp 10.130.0.79:8080: connect: connection refused kubelet logger.go:42: 20:28:15 | sidecar-deployment | 2024-08-29 20:27:42 +0000 UTC Normal Job.batch 00-find-service Completed Job completed job-controller logger.go:42: 20:28:15 | sidecar-deployment | 2024-08-29 20:27:43 +0000 UTC Normal Pod agent-as-sidecar2-747f49bf89-hsx4m Binding Scheduled Successfully assigned kuttl-test-smart-burro/agent-as-sidecar2-747f49bf89-hsx4m to ip-10-0-51-237.us-west-2.compute.internal default-scheduler logger.go:42: 20:28:15 | sidecar-deployment | 2024-08-29 20:27:43 +0000 UTC Normal Pod agent-as-sidecar2-747f49bf89-hsx4m AddedInterface Add eth0 [10.130.0.81/23] from ovn-kubernetes multus logger.go:42: 20:28:15 | sidecar-deployment | 2024-08-29 20:27:43 +0000 UTC Normal Pod agent-as-sidecar2-747f49bf89-hsx4m.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:345d3b6c507df9f8a18948a594e7a03698085689c6965faf79b8aadc74a0ceaa" already present on machine kubelet logger.go:42: 20:28:15 | sidecar-deployment | 2024-08-29 20:27:43 +0000 UTC Normal Pod agent-as-sidecar2-747f49bf89-hsx4m.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 20:28:15 | sidecar-deployment | 2024-08-29 20:27:43 +0000 UTC Normal Pod agent-as-sidecar2-747f49bf89-hsx4m.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 20:28:15 | sidecar-deployment | 2024-08-29 20:27:43 +0000 UTC Normal ReplicaSet.apps agent-as-sidecar2-747f49bf89 SuccessfulCreate Created pod: agent-as-sidecar2-747f49bf89-hsx4m replicaset-controller logger.go:42: 20:28:15 | sidecar-deployment | 2024-08-29 20:27:43 +0000 UTC Normal Deployment.apps agent-as-sidecar2 ScalingReplicaSet Scaled up replica set agent-as-sidecar2-747f49bf89 to 1 deployment-controller logger.go:42: 20:28:15 | sidecar-deployment | 2024-08-29 20:27:46 +0000 UTC Normal Pod agent-as-sidecar-7766c49f9-zsr5g.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 20:28:15 | sidecar-deployment | 2024-08-29 20:27:47 +0000 UTC Normal Pod 01-find-service-c4qrk Binding Scheduled Successfully assigned kuttl-test-smart-burro/01-find-service-c4qrk to ip-10-0-51-237.us-west-2.compute.internal default-scheduler logger.go:42: 20:28:15 | sidecar-deployment | 2024-08-29 20:27:47 +0000 UTC Normal Pod 01-find-service-c4qrk AddedInterface Add eth0 [10.130.0.82/23] from ovn-kubernetes multus logger.go:42: 20:28:15 | sidecar-deployment | 2024-08-29 20:27:47 +0000 UTC Normal Pod 01-find-service-c4qrk.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 20:28:15 | sidecar-deployment | 2024-08-29 20:27:47 +0000 UTC Normal Job.batch 01-find-service SuccessfulCreate Created pod: 01-find-service-c4qrk job-controller logger.go:42: 20:28:15 | sidecar-deployment | 2024-08-29 20:27:47 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-2ffq7.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.130.0.78:8080/": read tcp 10.130.0.2:50304->10.130.0.78:8080: read: connection reset by peer kubelet logger.go:42: 20:28:15 | sidecar-deployment | 2024-08-29 20:27:48 +0000 UTC Normal Pod 01-find-service-c4qrk.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 817ms (817ms including waiting) kubelet logger.go:42: 20:28:15 | sidecar-deployment | 2024-08-29 20:27:48 +0000 UTC Normal Pod 01-find-service-c4qrk.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 20:28:15 | sidecar-deployment | 2024-08-29 20:27:48 +0000 UTC Normal Pod 01-find-service-c4qrk.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 20:28:15 | sidecar-deployment | 2024-08-29 20:27:49 +0000 UTC Normal Pod vertx-create-span-sidecar-56b6b4ccb7-rckq5 Binding Scheduled Successfully assigned kuttl-test-smart-burro/vertx-create-span-sidecar-56b6b4ccb7-rckq5 to ip-10-0-51-237.us-west-2.compute.internal default-scheduler logger.go:42: 20:28:15 | sidecar-deployment | 2024-08-29 20:27:49 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-56b6b4ccb7 SuccessfulCreate Created pod: vertx-create-span-sidecar-56b6b4ccb7-rckq5 replicaset-controller logger.go:42: 20:28:15 | sidecar-deployment | 2024-08-29 20:27:49 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-84d458b68c SuccessfulDelete Deleted pod: vertx-create-span-sidecar-84d458b68c-2ffq7 replicaset-controller logger.go:42: 20:28:15 | sidecar-deployment | 2024-08-29 20:27:49 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled down replica set vertx-create-span-sidecar-84d458b68c to 0 from 1 deployment-controller logger.go:42: 20:28:15 | sidecar-deployment | 2024-08-29 20:27:49 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-56b6b4ccb7 to 1 from 0 deployment-controller logger.go:42: 20:28:15 | sidecar-deployment | 2024-08-29 20:27:50 +0000 UTC Normal Pod vertx-create-span-sidecar-56b6b4ccb7-rckq5 AddedInterface Add eth0 [10.130.0.83/23] from ovn-kubernetes multus logger.go:42: 20:28:15 | sidecar-deployment | 2024-08-29 20:27:50 +0000 UTC Normal Pod vertx-create-span-sidecar-56b6b4ccb7-rckq5.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 20:28:15 | sidecar-deployment | 2024-08-29 20:27:50 +0000 UTC Normal Pod vertx-create-span-sidecar-56b6b4ccb7-rckq5.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 20:28:15 | sidecar-deployment | 2024-08-29 20:27:50 +0000 UTC Normal Pod vertx-create-span-sidecar-56b6b4ccb7-rckq5.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 20:28:15 | sidecar-deployment | 2024-08-29 20:27:50 +0000 UTC Normal Pod vertx-create-span-sidecar-56b6b4ccb7-rckq5.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:5fcc9d365678f54c0f52005318dbbc32b87f743cae7e1a84568e769653f3b128" already present on machine kubelet logger.go:42: 20:28:15 | sidecar-deployment | 2024-08-29 20:27:50 +0000 UTC Normal Pod vertx-create-span-sidecar-56b6b4ccb7-rckq5.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 20:28:15 | sidecar-deployment | 2024-08-29 20:27:50 +0000 UTC Normal Pod vertx-create-span-sidecar-56b6b4ccb7-rckq5.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 20:28:15 | sidecar-deployment | 2024-08-29 20:27:58 +0000 UTC Warning Pod vertx-create-span-sidecar-56b6b4ccb7-rckq5.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.130.0.83:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 20:28:15 | sidecar-deployment | 2024-08-29 20:27:58 +0000 UTC Warning Pod vertx-create-span-sidecar-56b6b4ccb7-rckq5.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.130.0.83:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 20:28:15 | sidecar-deployment | 2024-08-29 20:28:00 +0000 UTC Normal Pod vertx-create-span-sidecar-56b6b4ccb7-rckq5.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 20:28:15 | sidecar-deployment | 2024-08-29 20:28:00 +0000 UTC Warning Pod vertx-create-span-sidecar-56b6b4ccb7-rckq5.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.130.0.83:8080/": read tcp 10.130.0.2:38124->10.130.0.83:8080: read: connection reset by peer kubelet logger.go:42: 20:28:15 | sidecar-deployment | 2024-08-29 20:28:00 +0000 UTC Warning Pod vertx-create-span-sidecar-56b6b4ccb7-rckq5.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.130.0.83:8080/": dial tcp 10.130.0.83:8080: connect: connection refused kubelet logger.go:42: 20:28:15 | sidecar-deployment | 2024-08-29 20:28:06 +0000 UTC Normal Job.batch 01-find-service Completed Job completed job-controller logger.go:42: 20:28:15 | sidecar-deployment | 2024-08-29 20:28:07 +0000 UTC Normal Pod vertx-create-span-sidecar-54767b7b46-dv4qj Binding Scheduled Successfully assigned kuttl-test-smart-burro/vertx-create-span-sidecar-54767b7b46-dv4qj to ip-10-0-57-246.us-west-2.compute.internal default-scheduler logger.go:42: 20:28:15 | sidecar-deployment | 2024-08-29 20:28:07 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-54767b7b46 SuccessfulCreate Created pod: vertx-create-span-sidecar-54767b7b46-dv4qj replicaset-controller logger.go:42: 20:28:15 | sidecar-deployment | 2024-08-29 20:28:07 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-5fcdc696b5 SuccessfulDelete Deleted pod: vertx-create-span-sidecar-5fcdc696b5-6pvqj replicaset-controller logger.go:42: 20:28:15 | sidecar-deployment | 2024-08-29 20:28:07 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled down replica set vertx-create-span-sidecar-5fcdc696b5 to 0 from 1 deployment-controller logger.go:42: 20:28:15 | sidecar-deployment | 2024-08-29 20:28:07 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-54767b7b46 to 1 from 0 deployment-controller logger.go:42: 20:28:15 | sidecar-deployment | 2024-08-29 20:28:08 +0000 UTC Normal Pod vertx-create-span-sidecar-54767b7b46-dv4qj AddedInterface Add eth0 [10.129.0.34/23] from ovn-kubernetes multus logger.go:42: 20:28:15 | sidecar-deployment | 2024-08-29 20:28:08 +0000 UTC Normal Pod vertx-create-span-sidecar-54767b7b46-dv4qj.spec.containers{vertx-create-span-sidecar} Pulling Pulling image "jaegertracing/vertx-create-span:operator-e2e-tests" kubelet logger.go:42: 20:28:15 | sidecar-deployment | 2024-08-29 20:28:13 +0000 UTC Normal Pod vertx-create-span-sidecar-54767b7b46-dv4qj.spec.containers{vertx-create-span-sidecar} Pulled Successfully pulled image "jaegertracing/vertx-create-span:operator-e2e-tests" in 5.481s (5.481s including waiting) kubelet logger.go:42: 20:28:15 | sidecar-deployment | 2024-08-29 20:28:13 +0000 UTC Normal Pod vertx-create-span-sidecar-54767b7b46-dv4qj.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 20:28:15 | sidecar-deployment | 2024-08-29 20:28:13 +0000 UTC Normal Pod vertx-create-span-sidecar-54767b7b46-dv4qj.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 20:28:15 | sidecar-deployment | Deleting namespace: kuttl-test-smart-burro === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (127.26s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (6.10s) --- PASS: kuttl/harness/sidecar-namespace (45.26s) --- PASS: kuttl/harness/sidecar-skip-webhook (14.48s) --- PASS: kuttl/harness/sidecar-deployment (60.76s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name sidecar --report --output /logs/artifacts/sidecar.xml ./artifacts/kuttl-report.xml time="2024-08-29T20:28:22Z" level=debug msg="Setting a new name for the test suites" time="2024-08-29T20:28:22Z" level=debug msg="Removing 'artifacts' TestCase" time="2024-08-29T20:28:22Z" level=debug msg="normalizing test case names" time="2024-08-29T20:28:22Z" level=debug msg="sidecar/artifacts -> sidecar_artifacts" time="2024-08-29T20:28:22Z" level=debug msg="sidecar/sidecar-namespace -> sidecar_sidecar_namespace" time="2024-08-29T20:28:22Z" level=debug msg="sidecar/sidecar-skip-webhook -> sidecar_sidecar_skip_webhook" time="2024-08-29T20:28:22Z" level=debug msg="sidecar/sidecar-deployment -> sidecar_sidecar_deployment" +------------------------------+--------+ | NAME | RESULT | +------------------------------+--------+ | sidecar_artifacts | passed | | sidecar_sidecar_namespace | passed | | sidecar_sidecar_skip_webhook | passed | | sidecar_sidecar_deployment | passed | +------------------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh streaming false true + '[' 3 -ne 3 ']' + test_suite_name=streaming + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/streaming.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-streaming make[2]: Entering directory '/tmp/jaeger-tests' >>>> Elasticsearch image not loaded because SKIP_ES_EXTERNAL is true KAFKA_VERSION=3.6.0 \ SKIP_KAFKA=false \ SKIP_ES_EXTERNAL=true \ ./tests/e2e/streaming/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.16.8 True False 84m Cluster version is 4.16.8' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.16.8 True False 84m Cluster version is 4.16.8' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/streaming/render.sh ++ export SUITE_DIR=./tests/e2e/streaming ++ SUITE_DIR=./tests/e2e/streaming ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/streaming ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + is_secured=false + '[' true = true ']' + is_secured=true + '[' false = true ']' + start_test streaming-simple + '[' 1 -ne 1 ']' + test_name=streaming-simple + echo =========================================================================== =========================================================================== + info 'Rendering files for test streaming-simple' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test streaming-simple\e[0m' Rendering files for test streaming-simple + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/streaming/_build + '[' _build '!=' _build ']' + mkdir -p streaming-simple + cd streaming-simple + render_install_kafka my-cluster 00 + '[' 2 -ne 2 ']' + cluster_name=my-cluster + test_step=00 + CLUSTER_NAME=my-cluster + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/kafka-install.yaml.template -o ./00-install.yaml + render_assert_kafka false my-cluster 00 + '[' 3 -ne 3 ']' + autoprovisioned=false + cluster_name=my-cluster + test_step=00 + '[' false = true ']' + '[' false = true ']' + '[' false = false ']' + replicas=1 + CLUSTER_NAME=my-cluster + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-zookeeper-cluster.yaml.template -o ./00-assert.yaml ++ expr 00 + 1 + CLUSTER_NAME=my-cluster + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-kafka-cluster.yaml.template -o ./01-assert.yaml ++ expr 00 + 2 + CLUSTER_NAME=my-cluster + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-entity-operator.yaml.template -o ./02-assert.yaml + render_install_elasticsearch upstream 03 + '[' 2 -ne 2 ']' + deploy_mode=upstream + test_step=03 + '[' upstream = upstream ']' + '[' true = true ']' + template=/tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template + /tmp/jaeger-tests/bin/yq eval -s '"elasticsearch_" + $index' /tmp/jaeger-tests/tests/elasticsearch.yml + /tmp/jaeger-tests/bin/yq eval -i '.spec.template.spec.serviceAccountName="deploy-elasticsearch"' ./elasticsearch_0.yml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template -o ./03-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/elasticsearch-assert.yaml.template -o ./03-assert.yaml + JAEGER_NAME=simple-streaming + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/streaming-jaeger-assert.yaml.template -o ./04-assert.yaml + render_smoke_test simple-streaming true 05 + '[' 3 -ne 3 ']' + jaeger=simple-streaming + is_secured=true + test_step=05 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simple-streaming-query:443 + JAEGER_QUERY_ENDPOINT=https://simple-streaming-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simple-streaming-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simple-streaming-collector-headless:14268 + export JAEGER_NAME=simple-streaming + JAEGER_NAME=simple-streaming + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./05-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./05-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + '[' false = true ']' + start_test streaming-with-tls + '[' 1 -ne 1 ']' + test_name=streaming-with-tls + echo =========================================================================== =========================================================================== + info 'Rendering files for test streaming-with-tls' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test streaming-with-tls\e[0m' Rendering files for test streaming-with-tls + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/streaming/_build/streaming-simple + '[' streaming-simple '!=' _build ']' + cd .. + mkdir -p streaming-with-tls + cd streaming-with-tls + render_install_kafka my-cluster 00 + '[' 2 -ne 2 ']' + cluster_name=my-cluster + test_step=00 + CLUSTER_NAME=my-cluster + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/kafka-install.yaml.template -o ./00-install.yaml + render_assert_kafka false my-cluster 00 + '[' 3 -ne 3 ']' + autoprovisioned=false + cluster_name=my-cluster + test_step=00 + '[' false = true ']' + '[' false = true ']' + '[' false = false ']' + replicas=1 + CLUSTER_NAME=my-cluster + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-zookeeper-cluster.yaml.template -o ./00-assert.yaml ++ expr 00 + 1 + CLUSTER_NAME=my-cluster + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-kafka-cluster.yaml.template -o ./01-assert.yaml ++ expr 00 + 2 + CLUSTER_NAME=my-cluster + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-entity-operator.yaml.template -o ./02-assert.yaml + render_install_elasticsearch upstream 03 + '[' 2 -ne 2 ']' + deploy_mode=upstream + test_step=03 + '[' upstream = upstream ']' + '[' true = true ']' + template=/tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template + /tmp/jaeger-tests/bin/yq eval -s '"elasticsearch_" + $index' /tmp/jaeger-tests/tests/elasticsearch.yml + /tmp/jaeger-tests/bin/yq eval -i '.spec.template.spec.serviceAccountName="deploy-elasticsearch"' ./elasticsearch_0.yml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template -o ./03-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/elasticsearch-assert.yaml.template -o ./03-assert.yaml + render_smoke_test tls-streaming true 05 + '[' 3 -ne 3 ']' + jaeger=tls-streaming + is_secured=true + test_step=05 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://tls-streaming-query:443 + JAEGER_QUERY_ENDPOINT=https://tls-streaming-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://tls-streaming-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://tls-streaming-collector-headless:14268 + export JAEGER_NAME=tls-streaming + JAEGER_NAME=tls-streaming + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./05-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./05-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + '[' false = true ']' + start_test streaming-with-autoprovisioning-autoscale + '[' 1 -ne 1 ']' + test_name=streaming-with-autoprovisioning-autoscale + echo =========================================================================== =========================================================================== + info 'Rendering files for test streaming-with-autoprovisioning-autoscale' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test streaming-with-autoprovisioning-autoscale\e[0m' Rendering files for test streaming-with-autoprovisioning-autoscale + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/streaming/_build/streaming-with-tls + '[' streaming-with-tls '!=' _build ']' + cd .. + mkdir -p streaming-with-autoprovisioning-autoscale + cd streaming-with-autoprovisioning-autoscale + '[' true = true ']' + rm ./00-install.yaml ./00-assert.yaml + render_install_elasticsearch upstream 01 + '[' 2 -ne 2 ']' + deploy_mode=upstream + test_step=01 + '[' upstream = upstream ']' + '[' true = true ']' + template=/tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template + /tmp/jaeger-tests/bin/yq eval -s '"elasticsearch_" + $index' /tmp/jaeger-tests/tests/elasticsearch.yml + /tmp/jaeger-tests/bin/yq eval -i '.spec.template.spec.serviceAccountName="deploy-elasticsearch"' ./elasticsearch_0.yml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/elasticsearch-assert.yaml.template -o ./01-assert.yaml + jaeger_name=auto-provisioned + /tmp/jaeger-tests/bin/yq e -i '.spec.ingester.resources.requests.memory="20Mi"' ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.ingester.resources.requests.memory="500m"' ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.ingester.autoscale=true ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.ingester.minReplicas=1 ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.ingester.maxReplicas=2 ./02-install.yaml + render_assert_kafka true auto-provisioned 03 + '[' 3 -ne 3 ']' + autoprovisioned=true + cluster_name=auto-provisioned + test_step=03 + '[' true = true ']' + is_kafka_minimal_enabled + namespaces=(observability openshift-operators openshift-distributed-tracing) + for i in "${namespaces[@]}" ++ kubectl get pods -n observability -l name=jaeger-operator -o yaml ++ /tmp/jaeger-tests/bin/yq e '.items[0].spec.containers[0].env[] | select(.name=="KAFKA-PROVISIONING-MINIMAL").value' + enabled= + '[' '' == true ']' + for i in "${namespaces[@]}" ++ kubectl get pods -n openshift-operators -l name=jaeger-operator -o yaml ++ /tmp/jaeger-tests/bin/yq e '.items[0].spec.containers[0].env[] | select(.name=="KAFKA-PROVISIONING-MINIMAL").value' + enabled= + '[' '' == true ']' + for i in "${namespaces[@]}" ++ kubectl get pods -n openshift-distributed-tracing -l name=jaeger-operator -o yaml ++ /tmp/jaeger-tests/bin/yq e '.items[0].spec.containers[0].env[] | select(.name=="KAFKA-PROVISIONING-MINIMAL").value' + enabled=true + '[' true == true ']' + return 0 + replicas=1 + CLUSTER_NAME=auto-provisioned + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-zookeeper-cluster.yaml.template -o ./03-assert.yaml ++ expr 03 + 1 + CLUSTER_NAME=auto-provisioned + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-kafka-cluster.yaml.template -o ./04-assert.yaml ++ expr 03 + 2 + CLUSTER_NAME=auto-provisioned + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-entity-operator.yaml.template -o ./05-assert.yaml + version_lt 1.30 1.23 ++ echo 1.30 1.23 ++ tr ' ' '\n' ++ sort -rV ++ head -n 1 + test 1.30 '!=' 1.30 + rm ./08-assert.yaml + skip_test streaming-with-tls 'This test is flaky in Prow CI' + '[' 2 -ne 2 ']' + test_name=streaming-with-tls + message='This test is flaky in Prow CI' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/streaming/_build/streaming-with-autoprovisioning-autoscale + '[' streaming-with-autoprovisioning-autoscale '!=' _build ']' + cd .. + rm -rf streaming-with-tls + warning 'streaming-with-tls: This test is flaky in Prow CI' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: streaming-with-tls: This test is flaky in Prow CI\e[0m' WAR: streaming-with-tls: This test is flaky in Prow CI + skip_test streaming-simple 'This test is flaky in Prow CI' + '[' 2 -ne 2 ']' + test_name=streaming-simple + message='This test is flaky in Prow CI' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/streaming/_build + '[' _build '!=' _build ']' + rm -rf streaming-simple + warning 'streaming-simple: This test is flaky in Prow CI' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: streaming-simple: This test is flaky in Prow CI\e[0m' WAR: streaming-simple: This test is flaky in Prow CI make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running streaming E2E tests' Running streaming E2E tests + cd tests/e2e/streaming/_build + set +e + KUBECONFIG=/tmp/kubeconfig-2030369819 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-rosa-h-yd3i.24xa.s3.devshift.org:443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 2 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/streaming-with-autoprovisioning-autoscale === PAUSE kuttl/harness/streaming-with-autoprovisioning-autoscale === CONT kuttl/harness/artifacts logger.go:42: 20:28:36 | artifacts | Creating namespace: kuttl-test-known-snipe logger.go:42: 20:28:36 | artifacts | artifacts events from ns kuttl-test-known-snipe: logger.go:42: 20:28:36 | artifacts | Deleting namespace: kuttl-test-known-snipe === CONT kuttl/harness/streaming-with-autoprovisioning-autoscale logger.go:42: 20:28:42 | streaming-with-autoprovisioning-autoscale | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 20:28:42 | streaming-with-autoprovisioning-autoscale | Ignoring elasticsearch_0.yml as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 20:28:42 | streaming-with-autoprovisioning-autoscale | Ignoring elasticsearch_1.yml as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 20:28:42 | streaming-with-autoprovisioning-autoscale | Creating namespace: kuttl-test-willing-panda logger.go:42: 20:28:42 | streaming-with-autoprovisioning-autoscale/1-install | starting test step 1-install logger.go:42: 20:28:42 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c oc create sa deploy-elasticsearch -n $NAMESPACE 2>&1 | grep -v "already exists" || true] logger.go:42: 20:28:42 | streaming-with-autoprovisioning-autoscale/1-install | serviceaccount/deploy-elasticsearch created logger.go:42: 20:28:42 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c oc adm policy add-scc-to-user privileged -z deploy-elasticsearch -n $NAMESPACE 2>&1 | grep -v "already exists" || true] logger.go:42: 20:28:43 | streaming-with-autoprovisioning-autoscale/1-install | clusterrole.rbac.authorization.k8s.io/system:openshift:scc:privileged added: "deploy-elasticsearch" logger.go:42: 20:28:43 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c sleep 6] logger.go:42: 20:28:49 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c kubectl apply -f elasticsearch_0.yml -n $NAMESPACE] logger.go:42: 20:28:50 | streaming-with-autoprovisioning-autoscale/1-install | statefulset.apps/elasticsearch created logger.go:42: 20:28:50 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c sleep 3] logger.go:42: 20:28:53 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c kubectl apply -f elasticsearch_1.yml -n $NAMESPACE] logger.go:42: 20:28:54 | streaming-with-autoprovisioning-autoscale/1-install | service/elasticsearch created logger.go:42: 20:29:11 | streaming-with-autoprovisioning-autoscale/1-install | test step completed 1-install logger.go:42: 20:29:11 | streaming-with-autoprovisioning-autoscale/2-install | starting test step 2-install logger.go:42: 20:29:11 | streaming-with-autoprovisioning-autoscale/2-install | Jaeger:kuttl-test-willing-panda/auto-provisioned created logger.go:42: 20:29:11 | streaming-with-autoprovisioning-autoscale/2-install | test step completed 2-install logger.go:42: 20:29:11 | streaming-with-autoprovisioning-autoscale/3- | starting test step 3- logger.go:42: 20:30:03 | streaming-with-autoprovisioning-autoscale/3- | test step completed 3- logger.go:42: 20:30:03 | streaming-with-autoprovisioning-autoscale/4- | starting test step 4- logger.go:42: 20:30:34 | streaming-with-autoprovisioning-autoscale/4- | test step completed 4- logger.go:42: 20:30:34 | streaming-with-autoprovisioning-autoscale/5- | starting test step 5- logger.go:42: 20:30:57 | streaming-with-autoprovisioning-autoscale/5- | test step completed 5- logger.go:42: 20:30:57 | streaming-with-autoprovisioning-autoscale/6- | starting test step 6- logger.go:42: 20:31:12 | streaming-with-autoprovisioning-autoscale/6- | test step completed 6- logger.go:42: 20:31:12 | streaming-with-autoprovisioning-autoscale/7- | starting test step 7- logger.go:42: 20:31:14 | streaming-with-autoprovisioning-autoscale/7- | test step completed 7- logger.go:42: 20:31:14 | streaming-with-autoprovisioning-autoscale | streaming-with-autoprovisioning-autoscale events from ns kuttl-test-willing-panda: logger.go:42: 20:31:14 | streaming-with-autoprovisioning-autoscale | 2024-08-29 20:28:50 +0000 UTC Normal Pod elasticsearch-0 Binding Scheduled Successfully assigned kuttl-test-willing-panda/elasticsearch-0 to ip-10-0-51-237.us-west-2.compute.internal default-scheduler logger.go:42: 20:31:14 | streaming-with-autoprovisioning-autoscale | 2024-08-29 20:28:50 +0000 UTC Normal Pod elasticsearch-0 AddedInterface Add eth0 [10.130.0.84/23] from ovn-kubernetes multus logger.go:42: 20:31:14 | streaming-with-autoprovisioning-autoscale | 2024-08-29 20:28:50 +0000 UTC Normal Pod elasticsearch-0.spec.containers{elasticsearch} Pulling Pulling image "docker.elastic.co/elasticsearch/elasticsearch-oss:6.8.6" kubelet logger.go:42: 20:31:14 | streaming-with-autoprovisioning-autoscale | 2024-08-29 20:28:50 +0000 UTC Normal StatefulSet.apps elasticsearch SuccessfulCreate create Pod elasticsearch-0 in StatefulSet elasticsearch successful statefulset-controller logger.go:42: 20:31:14 | streaming-with-autoprovisioning-autoscale | 2024-08-29 20:29:00 +0000 UTC Normal Pod elasticsearch-0.spec.containers{elasticsearch} Pulled Successfully pulled image "docker.elastic.co/elasticsearch/elasticsearch-oss:6.8.6" in 9.323s (9.324s including waiting) kubelet logger.go:42: 20:31:14 | streaming-with-autoprovisioning-autoscale | 2024-08-29 20:29:00 +0000 UTC Normal Pod elasticsearch-0.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 20:31:14 | streaming-with-autoprovisioning-autoscale | 2024-08-29 20:29:00 +0000 UTC Normal Pod elasticsearch-0.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 20:31:14 | streaming-with-autoprovisioning-autoscale | 2024-08-29 20:29:05 +0000 UTC Warning Pod elasticsearch-0.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Get "http://10.130.0.84:9200/": dial tcp 10.130.0.84:9200: connect: connection refused kubelet logger.go:42: 20:31:14 | streaming-with-autoprovisioning-autoscale | 2024-08-29 20:29:15 +0000 UTC Normal PersistentVolumeClaim data-auto-provisioned-zookeeper-0 WaitForFirstConsumer waiting for first consumer to be created before binding persistentvolume-controller logger.go:42: 20:31:14 | streaming-with-autoprovisioning-autoscale | 2024-08-29 20:29:16 +0000 UTC Normal PodDisruptionBudget.policy auto-provisioned-zookeeper NoPods No matching pods found controllermanager logger.go:42: 20:31:14 | streaming-with-autoprovisioning-autoscale | 2024-08-29 20:29:16 +0000 UTC Normal PersistentVolumeClaim data-auto-provisioned-zookeeper-0 Provisioning External provisioner is provisioning volume for claim "kuttl-test-willing-panda/data-auto-provisioned-zookeeper-0" ebs.csi.aws.com_aws-ebs-csi-driver-controller-76f7947f-92pz4_c7c967fd-b256-4b31-990e-45d75c55c671 logger.go:42: 20:31:14 | streaming-with-autoprovisioning-autoscale | 2024-08-29 20:29:16 +0000 UTC Normal PersistentVolumeClaim data-auto-provisioned-zookeeper-0 ExternalProvisioning Waiting for a volume to be created either by the external provisioner 'ebs.csi.aws.com' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered. persistentvolume-controller logger.go:42: 20:31:14 | streaming-with-autoprovisioning-autoscale | 2024-08-29 20:29:21 +0000 UTC Normal Pod auto-provisioned-zookeeper-0 Binding Scheduled Successfully assigned kuttl-test-willing-panda/auto-provisioned-zookeeper-0 to ip-10-0-51-237.us-west-2.compute.internal default-scheduler logger.go:42: 20:31:14 | streaming-with-autoprovisioning-autoscale | 2024-08-29 20:29:21 +0000 UTC Normal PersistentVolumeClaim data-auto-provisioned-zookeeper-0 ProvisioningSucceeded Successfully provisioned volume pvc-71516426-26ce-43c2-9b0f-9f94f0abdc61 ebs.csi.aws.com_aws-ebs-csi-driver-controller-76f7947f-92pz4_c7c967fd-b256-4b31-990e-45d75c55c671 logger.go:42: 20:31:14 | streaming-with-autoprovisioning-autoscale | 2024-08-29 20:29:24 +0000 UTC Normal Pod auto-provisioned-zookeeper-0 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-71516426-26ce-43c2-9b0f-9f94f0abdc61" attachdetach-controller logger.go:42: 20:31:14 | streaming-with-autoprovisioning-autoscale | 2024-08-29 20:29:32 +0000 UTC Normal Pod auto-provisioned-zookeeper-0 AddedInterface Add eth0 [10.130.0.85/23] from ovn-kubernetes multus logger.go:42: 20:31:14 | streaming-with-autoprovisioning-autoscale | 2024-08-29 20:29:32 +0000 UTC Normal Pod auto-provisioned-zookeeper-0.spec.containers{zookeeper} Pulling Pulling image "registry.redhat.io/amq-streams/kafka-37-rhel9@sha256:ea5bb9a3f970f1db7c4335b851fe8c791798641e7b6fef0423797f138f4f30d7" kubelet logger.go:42: 20:31:14 | streaming-with-autoprovisioning-autoscale | 2024-08-29 20:29:43 +0000 UTC Normal Pod auto-provisioned-zookeeper-0.spec.containers{zookeeper} Pulled Successfully pulled image "registry.redhat.io/amq-streams/kafka-37-rhel9@sha256:ea5bb9a3f970f1db7c4335b851fe8c791798641e7b6fef0423797f138f4f30d7" in 10.512s (10.512s including waiting) kubelet logger.go:42: 20:31:14 | streaming-with-autoprovisioning-autoscale | 2024-08-29 20:29:43 +0000 UTC Normal Pod auto-provisioned-zookeeper-0.spec.containers{zookeeper} Created Created container zookeeper kubelet logger.go:42: 20:31:14 | streaming-with-autoprovisioning-autoscale | 2024-08-29 20:29:43 +0000 UTC Normal Pod auto-provisioned-zookeeper-0.spec.containers{zookeeper} Started Started container zookeeper kubelet logger.go:42: 20:31:14 | streaming-with-autoprovisioning-autoscale | 2024-08-29 20:30:04 +0000 UTC Normal PersistentVolumeClaim data-0-auto-provisioned-kafka-0 WaitForFirstConsumer waiting for first consumer to be created before binding persistentvolume-controller logger.go:42: 20:31:14 | streaming-with-autoprovisioning-autoscale | 2024-08-29 20:30:05 +0000 UTC Normal PodDisruptionBudget.policy auto-provisioned-kafka NoPods No matching pods found controllermanager logger.go:42: 20:31:14 | streaming-with-autoprovisioning-autoscale | 2024-08-29 20:30:05 +0000 UTC Normal PersistentVolumeClaim data-0-auto-provisioned-kafka-0 Provisioning External provisioner is provisioning volume for claim "kuttl-test-willing-panda/data-0-auto-provisioned-kafka-0" ebs.csi.aws.com_aws-ebs-csi-driver-controller-76f7947f-92pz4_c7c967fd-b256-4b31-990e-45d75c55c671 logger.go:42: 20:31:14 | streaming-with-autoprovisioning-autoscale | 2024-08-29 20:30:05 +0000 UTC Normal PersistentVolumeClaim data-0-auto-provisioned-kafka-0 ExternalProvisioning Waiting for a volume to be created either by the external provisioner 'ebs.csi.aws.com' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered. persistentvolume-controller logger.go:42: 20:31:14 | streaming-with-autoprovisioning-autoscale | 2024-08-29 20:30:09 +0000 UTC Normal PersistentVolumeClaim data-0-auto-provisioned-kafka-0 ProvisioningSucceeded Successfully provisioned volume pvc-0e72f8da-68ca-4075-aa9a-fd3d119a503d ebs.csi.aws.com_aws-ebs-csi-driver-controller-76f7947f-92pz4_c7c967fd-b256-4b31-990e-45d75c55c671 logger.go:42: 20:31:14 | streaming-with-autoprovisioning-autoscale | 2024-08-29 20:30:10 +0000 UTC Normal Pod auto-provisioned-kafka-0 Binding Scheduled Successfully assigned kuttl-test-willing-panda/auto-provisioned-kafka-0 to ip-10-0-51-237.us-west-2.compute.internal default-scheduler logger.go:42: 20:31:14 | streaming-with-autoprovisioning-autoscale | 2024-08-29 20:30:12 +0000 UTC Normal Pod auto-provisioned-kafka-0 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-0e72f8da-68ca-4075-aa9a-fd3d119a503d" attachdetach-controller logger.go:42: 20:31:14 | streaming-with-autoprovisioning-autoscale | 2024-08-29 20:30:13 +0000 UTC Normal Pod auto-provisioned-kafka-0 AddedInterface Add eth0 [10.130.0.86/23] from ovn-kubernetes multus logger.go:42: 20:31:14 | streaming-with-autoprovisioning-autoscale | 2024-08-29 20:30:13 +0000 UTC Normal Pod auto-provisioned-kafka-0.spec.containers{kafka} Pulled Container image "registry.redhat.io/amq-streams/kafka-37-rhel9@sha256:ea5bb9a3f970f1db7c4335b851fe8c791798641e7b6fef0423797f138f4f30d7" already present on machine kubelet logger.go:42: 20:31:14 | streaming-with-autoprovisioning-autoscale | 2024-08-29 20:30:13 +0000 UTC Normal Pod auto-provisioned-kafka-0.spec.containers{kafka} Created Created container kafka kubelet logger.go:42: 20:31:14 | streaming-with-autoprovisioning-autoscale | 2024-08-29 20:30:13 +0000 UTC Normal Pod auto-provisioned-kafka-0.spec.containers{kafka} Started Started container kafka kubelet logger.go:42: 20:31:14 | streaming-with-autoprovisioning-autoscale | 2024-08-29 20:30:36 +0000 UTC Normal Pod auto-provisioned-entity-operator-5984b6c687-c7w76 Binding Scheduled Successfully assigned kuttl-test-willing-panda/auto-provisioned-entity-operator-5984b6c687-c7w76 to ip-10-0-51-237.us-west-2.compute.internal default-scheduler logger.go:42: 20:31:14 | streaming-with-autoprovisioning-autoscale | 2024-08-29 20:30:36 +0000 UTC Normal Pod auto-provisioned-entity-operator-5984b6c687-c7w76 AddedInterface Add eth0 [10.130.0.87/23] from ovn-kubernetes multus logger.go:42: 20:31:14 | streaming-with-autoprovisioning-autoscale | 2024-08-29 20:30:36 +0000 UTC Normal Pod auto-provisioned-entity-operator-5984b6c687-c7w76.spec.containers{topic-operator} Pulled Container image "registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:a12c5647a10a6fcbc6963547022ebde643b3ee8ec89f3b33a73851312f7fdb76" already present on machine kubelet logger.go:42: 20:31:14 | streaming-with-autoprovisioning-autoscale | 2024-08-29 20:30:36 +0000 UTC Normal Pod auto-provisioned-entity-operator-5984b6c687-c7w76.spec.containers{topic-operator} Created Created container topic-operator kubelet logger.go:42: 20:31:14 | streaming-with-autoprovisioning-autoscale | 2024-08-29 20:30:36 +0000 UTC Normal Pod auto-provisioned-entity-operator-5984b6c687-c7w76.spec.containers{topic-operator} Started Started container topic-operator kubelet logger.go:42: 20:31:14 | streaming-with-autoprovisioning-autoscale | 2024-08-29 20:30:36 +0000 UTC Normal Pod auto-provisioned-entity-operator-5984b6c687-c7w76.spec.containers{user-operator} Pulled Container image "registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:a12c5647a10a6fcbc6963547022ebde643b3ee8ec89f3b33a73851312f7fdb76" already present on machine kubelet logger.go:42: 20:31:14 | streaming-with-autoprovisioning-autoscale | 2024-08-29 20:30:36 +0000 UTC Normal Pod auto-provisioned-entity-operator-5984b6c687-c7w76.spec.containers{user-operator} Created Created container user-operator kubelet logger.go:42: 20:31:14 | streaming-with-autoprovisioning-autoscale | 2024-08-29 20:30:36 +0000 UTC Normal Pod auto-provisioned-entity-operator-5984b6c687-c7w76.spec.containers{user-operator} Started Started container user-operator kubelet logger.go:42: 20:31:14 | streaming-with-autoprovisioning-autoscale | 2024-08-29 20:30:36 +0000 UTC Normal ReplicaSet.apps auto-provisioned-entity-operator-5984b6c687 SuccessfulCreate Created pod: auto-provisioned-entity-operator-5984b6c687-c7w76 replicaset-controller logger.go:42: 20:31:14 | streaming-with-autoprovisioning-autoscale | 2024-08-29 20:30:36 +0000 UTC Normal Deployment.apps auto-provisioned-entity-operator ScalingReplicaSet Scaled up replica set auto-provisioned-entity-operator-5984b6c687 to 1 deployment-controller logger.go:42: 20:31:14 | streaming-with-autoprovisioning-autoscale | 2024-08-29 20:31:00 +0000 UTC Normal Pod auto-provisioned-collector-64bf8889db-bfqzn Binding Scheduled Successfully assigned kuttl-test-willing-panda/auto-provisioned-collector-64bf8889db-bfqzn to ip-10-0-51-237.us-west-2.compute.internal default-scheduler logger.go:42: 20:31:14 | streaming-with-autoprovisioning-autoscale | 2024-08-29 20:31:00 +0000 UTC Normal Pod auto-provisioned-collector-64bf8889db-bfqzn AddedInterface Add eth0 [10.130.0.88/23] from ovn-kubernetes multus logger.go:42: 20:31:14 | streaming-with-autoprovisioning-autoscale | 2024-08-29 20:31:00 +0000 UTC Normal Pod auto-provisioned-collector-64bf8889db-bfqzn.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:b90e0d264bb0b051c2a4b9e6d499f407fed58f834ab8ebf30263ee1cf7c7ea3c" already present on machine kubelet logger.go:42: 20:31:14 | streaming-with-autoprovisioning-autoscale | 2024-08-29 20:31:00 +0000 UTC Normal Pod auto-provisioned-collector-64bf8889db-bfqzn.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 20:31:14 | streaming-with-autoprovisioning-autoscale | 2024-08-29 20:31:00 +0000 UTC Normal Pod auto-provisioned-collector-64bf8889db-bfqzn.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 20:31:14 | streaming-with-autoprovisioning-autoscale | 2024-08-29 20:31:00 +0000 UTC Normal ReplicaSet.apps auto-provisioned-collector-64bf8889db SuccessfulCreate Created pod: auto-provisioned-collector-64bf8889db-bfqzn replicaset-controller logger.go:42: 20:31:14 | streaming-with-autoprovisioning-autoscale | 2024-08-29 20:31:00 +0000 UTC Normal Deployment.apps auto-provisioned-collector ScalingReplicaSet Scaled up replica set auto-provisioned-collector-64bf8889db to 1 deployment-controller logger.go:42: 20:31:14 | streaming-with-autoprovisioning-autoscale | 2024-08-29 20:31:00 +0000 UTC Normal Pod auto-provisioned-ingester-7d578bf955-kpxng Binding Scheduled Successfully assigned kuttl-test-willing-panda/auto-provisioned-ingester-7d578bf955-kpxng to ip-10-0-57-246.us-west-2.compute.internal default-scheduler logger.go:42: 20:31:14 | streaming-with-autoprovisioning-autoscale | 2024-08-29 20:31:00 +0000 UTC Normal Pod auto-provisioned-ingester-7d578bf955-kpxng AddedInterface Add eth0 [10.129.0.35/23] from ovn-kubernetes multus logger.go:42: 20:31:14 | streaming-with-autoprovisioning-autoscale | 2024-08-29 20:31:00 +0000 UTC Normal ReplicaSet.apps auto-provisioned-ingester-7d578bf955 SuccessfulCreate Created pod: auto-provisioned-ingester-7d578bf955-kpxng replicaset-controller logger.go:42: 20:31:14 | streaming-with-autoprovisioning-autoscale | 2024-08-29 20:31:00 +0000 UTC Normal Deployment.apps auto-provisioned-ingester ScalingReplicaSet Scaled up replica set auto-provisioned-ingester-7d578bf955 to 1 deployment-controller logger.go:42: 20:31:14 | streaming-with-autoprovisioning-autoscale | 2024-08-29 20:31:00 +0000 UTC Normal Pod auto-provisioned-query-84465ff784-zzhj4 Binding Scheduled Successfully assigned kuttl-test-willing-panda/auto-provisioned-query-84465ff784-zzhj4 to ip-10-0-51-237.us-west-2.compute.internal default-scheduler logger.go:42: 20:31:14 | streaming-with-autoprovisioning-autoscale | 2024-08-29 20:31:00 +0000 UTC Warning Pod auto-provisioned-query-84465ff784-zzhj4 FailedMount MountVolume.SetUp failed for volume "auto-provisioned-ui-oauth-proxy-tls" : secret "auto-provisioned-ui-oauth-proxy-tls" not found kubelet logger.go:42: 20:31:14 | streaming-with-autoprovisioning-autoscale | 2024-08-29 20:31:00 +0000 UTC Normal ReplicaSet.apps auto-provisioned-query-84465ff784 SuccessfulCreate Created pod: auto-provisioned-query-84465ff784-zzhj4 replicaset-controller logger.go:42: 20:31:14 | streaming-with-autoprovisioning-autoscale | 2024-08-29 20:31:00 +0000 UTC Normal Deployment.apps auto-provisioned-query ScalingReplicaSet Scaled up replica set auto-provisioned-query-84465ff784 to 1 deployment-controller logger.go:42: 20:31:14 | streaming-with-autoprovisioning-autoscale | 2024-08-29 20:31:01 +0000 UTC Normal Pod auto-provisioned-ingester-7d578bf955-kpxng.spec.containers{jaeger-ingester} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-ingester-rhel8@sha256:fe18ffcf04f725dd7e5c1ba0c17ffab97dd4b6c51b20d40846e9a1479f233b6c" kubelet logger.go:42: 20:31:14 | streaming-with-autoprovisioning-autoscale | 2024-08-29 20:31:01 +0000 UTC Normal Pod auto-provisioned-query-84465ff784-zzhj4 AddedInterface Add eth0 [10.130.0.89/23] from ovn-kubernetes multus logger.go:42: 20:31:14 | streaming-with-autoprovisioning-autoscale | 2024-08-29 20:31:01 +0000 UTC Normal Pod auto-provisioned-query-84465ff784-zzhj4.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:831f83e8396d61822fb2dd576e711e0f10c1f8c7bd03bf0d4bec49dfb90cacca" already present on machine kubelet logger.go:42: 20:31:14 | streaming-with-autoprovisioning-autoscale | 2024-08-29 20:31:01 +0000 UTC Normal Pod auto-provisioned-query-84465ff784-zzhj4.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 20:31:14 | streaming-with-autoprovisioning-autoscale | 2024-08-29 20:31:01 +0000 UTC Normal Pod auto-provisioned-query-84465ff784-zzhj4.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 20:31:14 | streaming-with-autoprovisioning-autoscale | 2024-08-29 20:31:01 +0000 UTC Normal Pod auto-provisioned-query-84465ff784-zzhj4.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1613032062a68faa414cd5754a666b7b78ad0e52cf328eaaed355719423a55dd" already present on machine kubelet logger.go:42: 20:31:14 | streaming-with-autoprovisioning-autoscale | 2024-08-29 20:31:01 +0000 UTC Normal Pod auto-provisioned-query-84465ff784-zzhj4.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 20:31:14 | streaming-with-autoprovisioning-autoscale | 2024-08-29 20:31:01 +0000 UTC Normal Pod auto-provisioned-query-84465ff784-zzhj4.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 20:31:14 | streaming-with-autoprovisioning-autoscale | 2024-08-29 20:31:01 +0000 UTC Normal Pod auto-provisioned-query-84465ff784-zzhj4.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:5fcc9d365678f54c0f52005318dbbc32b87f743cae7e1a84568e769653f3b128" already present on machine kubelet logger.go:42: 20:31:14 | streaming-with-autoprovisioning-autoscale | 2024-08-29 20:31:01 +0000 UTC Normal Pod auto-provisioned-query-84465ff784-zzhj4.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 20:31:14 | streaming-with-autoprovisioning-autoscale | 2024-08-29 20:31:01 +0000 UTC Normal Pod auto-provisioned-query-84465ff784-zzhj4.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 20:31:14 | streaming-with-autoprovisioning-autoscale | 2024-08-29 20:31:10 +0000 UTC Normal Pod auto-provisioned-ingester-7d578bf955-kpxng.spec.containers{jaeger-ingester} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-ingester-rhel8@sha256:fe18ffcf04f725dd7e5c1ba0c17ffab97dd4b6c51b20d40846e9a1479f233b6c" in 9.939s (9.939s including waiting) kubelet logger.go:42: 20:31:14 | streaming-with-autoprovisioning-autoscale | 2024-08-29 20:31:11 +0000 UTC Normal Pod auto-provisioned-ingester-7d578bf955-kpxng.spec.containers{jaeger-ingester} Created Created container jaeger-ingester kubelet logger.go:42: 20:31:14 | streaming-with-autoprovisioning-autoscale | 2024-08-29 20:31:11 +0000 UTC Normal Pod auto-provisioned-ingester-7d578bf955-kpxng.spec.containers{jaeger-ingester} Started Started container jaeger-ingester kubelet logger.go:42: 20:31:14 | streaming-with-autoprovisioning-autoscale | Deleting namespace: kuttl-test-willing-panda === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (193.85s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (6.19s) --- PASS: kuttl/harness/streaming-with-autoprovisioning-autoscale (186.95s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name streaming --report --output /logs/artifacts/streaming.xml ./artifacts/kuttl-report.xml time="2024-08-29T20:31:49Z" level=debug msg="Setting a new name for the test suites" time="2024-08-29T20:31:49Z" level=debug msg="Removing 'artifacts' TestCase" time="2024-08-29T20:31:49Z" level=debug msg="normalizing test case names" time="2024-08-29T20:31:49Z" level=debug msg="streaming/artifacts -> streaming_artifacts" time="2024-08-29T20:31:49Z" level=debug msg="streaming/streaming-with-autoprovisioning-autoscale -> streaming_streaming_with_autoprovisioning_autoscale" +-----------------------------------------------------+--------+ | NAME | RESULT | +-----------------------------------------------------+--------+ | streaming_artifacts | passed | | streaming_streaming_with_autoprovisioning_autoscale | passed | +-----------------------------------------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh ui false true + '[' 3 -ne 3 ']' + test_suite_name=ui + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/ui.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-ui make[2]: Entering directory '/tmp/jaeger-tests' >>>> Elasticsearch image not loaded because SKIP_ES_EXTERNAL is true SKIP_ES_EXTERNAL=true ./tests/e2e/ui/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.16.8 True False 87m Cluster version is 4.16.8' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.16.8 True False 87m Cluster version is 4.16.8' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/ui/render.sh ++ export SUITE_DIR=./tests/e2e/ui ++ SUITE_DIR=./tests/e2e/ui ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/ui ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + start_test allinone + '[' 1 -ne 1 ']' + test_name=allinone + echo =========================================================================== =========================================================================== + info 'Rendering files for test allinone' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test allinone\e[0m' Rendering files for test allinone + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/ui/_build + '[' _build '!=' _build ']' + mkdir -p allinone + cd allinone + export GET_URL_COMMAND + export URL + export JAEGER_NAME=all-in-one-ui + JAEGER_NAME=all-in-one-ui + '[' true = true ']' + GET_URL_COMMAND='kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE' + URL='https://$(kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE)/search' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/ensure-ingress-host.sh.template -o ./ensure-ingress-host.sh + chmod +x ./ensure-ingress-host.sh + EXPECTED_CODE=200 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./01-curl.yaml + ASSERT_PRESENT=true + TRACKING_ID=MyTrackingId + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/test-ui-config.yaml.template -o ./04-test-ui-config.yaml + start_test production + '[' 1 -ne 1 ']' + test_name=production + echo =========================================================================== =========================================================================== + info 'Rendering files for test production' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test production\e[0m' Rendering files for test production + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/ui/_build/allinone + '[' allinone '!=' _build ']' + cd .. + mkdir -p production + cd production + export JAEGER_NAME=production-ui + JAEGER_NAME=production-ui + [[ true = true ]] + [[ true = true ]] + render_install_jaeger production-ui production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=production-ui + JAEGER_NAME=production-ui + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/ensure-ingress-host.sh.template -o ./ensure-ingress-host.sh + chmod +x ./ensure-ingress-host.sh + '[' true = true ']' + INSECURE=true + EXPECTED_CODE=403 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./02-check-forbbiden-access.yaml + EXPECTED_CODE=200 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./03-curl.yaml + INSECURE=true + EXPECTED_CODE=200 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./05-check-disabled-security.yaml + ASSERT_PRESENT=false + TRACKING_ID=MyTrackingId + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/test-ui-config.yaml.template -o ./06-check-NO-gaID.yaml + ASSERT_PRESENT=true + TRACKING_ID=MyTrackingId + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/test-ui-config.yaml.template -o ./08-check-gaID.yaml make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running ui E2E tests' Running ui E2E tests + cd tests/e2e/ui/_build + set +e + KUBECONFIG=/tmp/kubeconfig-2030369819 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-rosa-h-yd3i.24xa.s3.devshift.org:443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 3 tests === RUN kuttl/harness === RUN kuttl/harness/allinone === PAUSE kuttl/harness/allinone === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/production === PAUSE kuttl/harness/production === CONT kuttl/harness/allinone logger.go:42: 20:31:57 | allinone | Ignoring ensure-ingress-host.sh as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 20:31:57 | allinone | Creating namespace: kuttl-test-inspired-shiner logger.go:42: 20:31:57 | allinone/0-install | starting test step 0-install logger.go:42: 20:31:58 | allinone/0-install | Jaeger:kuttl-test-inspired-shiner/all-in-one-ui created logger.go:42: 20:31:59 | allinone/0-install | test step completed 0-install logger.go:42: 20:31:59 | allinone/1-curl | starting test step 1-curl logger.go:42: 20:31:59 | allinone/1-curl | running command: [./ensure-ingress-host.sh] logger.go:42: 20:31:59 | allinone/1-curl | Checking the Ingress host value was populated logger.go:42: 20:31:59 | allinone/1-curl | Try number 0 logger.go:42: 20:31:59 | allinone/1-curl | Hostname is all-in-one-ui-kuttl-test-inspired-shiner.apps.rosa.ci-rosa-h-yd3i.24xa.s3.devshift.org logger.go:42: 20:31:59 | allinone/1-curl | running command: [sh -c ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 200 true $NAMESPACE all-in-one-ui] logger.go:42: 20:32:00 | allinone/1-curl | Checking an expected HTTP response logger.go:42: 20:32:00 | allinone/1-curl | Running in OpenShift logger.go:42: 20:32:00 | allinone/1-curl | User not provided. Getting the token... logger.go:42: 20:32:01 | allinone/1-curl | Warning: resource jaegers/all-in-one-ui is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 20:32:09 | allinone/1-curl | Try number 1/30 the https://all-in-one-ui-kuttl-test-inspired-shiner.apps.rosa.ci-rosa-h-yd3i.24xa.s3.devshift.org/search logger.go:42: 20:32:09 | allinone/1-curl | curl response asserted properly logger.go:42: 20:32:10 | allinone/1-curl | test step completed 1-curl logger.go:42: 20:32:10 | allinone/2-delete | starting test step 2-delete logger.go:42: 20:32:10 | allinone/2-delete | Jaeger:kuttl-test-inspired-shiner/all-in-one-ui created logger.go:42: 20:32:10 | allinone/2-delete | test step completed 2-delete logger.go:42: 20:32:10 | allinone/3-install | starting test step 3-install logger.go:42: 20:32:10 | allinone/3-install | Jaeger:kuttl-test-inspired-shiner/all-in-one-ui updated logger.go:42: 20:32:10 | allinone/3-install | test step completed 3-install logger.go:42: 20:32:10 | allinone/4-test-ui-config | starting test step 4-test-ui-config logger.go:42: 20:32:10 | allinone/4-test-ui-config | running command: [./ensure-ingress-host.sh] logger.go:42: 20:32:11 | allinone/4-test-ui-config | Checking the Ingress host value was populated logger.go:42: 20:32:11 | allinone/4-test-ui-config | Try number 0 logger.go:42: 20:32:11 | allinone/4-test-ui-config | error: error executing jsonpath "{.items[0].status.ingress[0].host}": Error executing template: array index out of bounds: index 0, length 0. Printing more information for debugging the template: logger.go:42: 20:32:11 | allinone/4-test-ui-config | template was: logger.go:42: 20:32:11 | allinone/4-test-ui-config | {.items[0].status.ingress[0].host} logger.go:42: 20:32:11 | allinone/4-test-ui-config | object given to jsonpath engine was: logger.go:42: 20:32:11 | allinone/4-test-ui-config | map[string]interface {}{"apiVersion":"v1", "items":[]interface {}{}, "kind":"List", "metadata":map[string]interface {}{"resourceVersion":""}} logger.go:42: 20:32:11 | allinone/4-test-ui-config | logger.go:42: 20:32:11 | allinone/4-test-ui-config | logger.go:42: 20:32:21 | allinone/4-test-ui-config | Try number 1 logger.go:42: 20:32:21 | allinone/4-test-ui-config | Hostname is all-in-one-ui-kuttl-test-inspired-shiner.apps.rosa.ci-rosa-h-yd3i.24xa.s3.devshift.org logger.go:42: 20:32:21 | allinone/4-test-ui-config | running command: [sh -c ASSERT_PRESENT=true EXPECTED_CONTENT=MyTrackingId QUERY_HOSTNAME=https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search go run ../../../../cmd-utils/uiconfig/main.go] logger.go:42: 20:32:22 | allinone/4-test-ui-config | time="2024-08-29T20:32:22Z" level=info msg="Querying https://all-in-one-ui-kuttl-test-inspired-shiner.apps.rosa.ci-rosa-h-yd3i.24xa.s3.devshift.org/search..." logger.go:42: 20:32:22 | allinone/4-test-ui-config | time="2024-08-29T20:32:22Z" level=info msg="No secret provided for the Authorization header" logger.go:42: 20:32:22 | allinone/4-test-ui-config | time="2024-08-29T20:32:22Z" level=info msg="Polling to https://all-in-one-ui-kuttl-test-inspired-shiner.apps.rosa.ci-rosa-h-yd3i.24xa.s3.devshift.org/search" logger.go:42: 20:32:22 | allinone/4-test-ui-config | time="2024-08-29T20:32:22Z" level=info msg="Doing request number 0" logger.go:42: 20:32:22 | allinone/4-test-ui-config | time="2024-08-29T20:32:22Z" level=info msg="Content found and asserted!" logger.go:42: 20:32:22 | allinone/4-test-ui-config | time="2024-08-29T20:32:22Z" level=info msg="Success!" logger.go:42: 20:32:23 | allinone/4-test-ui-config | test step completed 4-test-ui-config logger.go:42: 20:32:23 | allinone | allinone events from ns kuttl-test-inspired-shiner: logger.go:42: 20:32:23 | allinone | 2024-08-29 20:31:58 +0000 UTC Normal Pod all-in-one-ui-7d8d89895-786nr Binding Scheduled Successfully assigned kuttl-test-inspired-shiner/all-in-one-ui-7d8d89895-786nr to ip-10-0-51-237.us-west-2.compute.internal default-scheduler logger.go:42: 20:32:23 | allinone | 2024-08-29 20:31:58 +0000 UTC Normal ReplicaSet.apps all-in-one-ui-7d8d89895 SuccessfulCreate Created pod: all-in-one-ui-7d8d89895-786nr replicaset-controller logger.go:42: 20:32:23 | allinone | 2024-08-29 20:31:58 +0000 UTC Normal Deployment.apps all-in-one-ui ScalingReplicaSet Scaled up replica set all-in-one-ui-7d8d89895 to 1 deployment-controller logger.go:42: 20:32:23 | allinone | 2024-08-29 20:31:59 +0000 UTC Normal Pod all-in-one-ui-7d8d89895-786nr AddedInterface Add eth0 [10.130.0.90/23] from ovn-kubernetes multus logger.go:42: 20:32:23 | allinone | 2024-08-29 20:31:59 +0000 UTC Normal Pod all-in-one-ui-7d8d89895-786nr.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:345d3b6c507df9f8a18948a594e7a03698085689c6965faf79b8aadc74a0ceaa" already present on machine kubelet logger.go:42: 20:32:23 | allinone | 2024-08-29 20:31:59 +0000 UTC Normal Pod all-in-one-ui-7d8d89895-786nr.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 20:32:23 | allinone | 2024-08-29 20:31:59 +0000 UTC Normal Pod all-in-one-ui-7d8d89895-786nr.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 20:32:23 | allinone | 2024-08-29 20:31:59 +0000 UTC Normal Pod all-in-one-ui-7d8d89895-786nr.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1613032062a68faa414cd5754a666b7b78ad0e52cf328eaaed355719423a55dd" already present on machine kubelet logger.go:42: 20:32:23 | allinone | 2024-08-29 20:31:59 +0000 UTC Normal Pod all-in-one-ui-7d8d89895-786nr.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 20:32:23 | allinone | 2024-08-29 20:31:59 +0000 UTC Normal Pod all-in-one-ui-7d8d89895-786nr.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 20:32:23 | allinone | 2024-08-29 20:32:02 +0000 UTC Normal ReplicaSet.apps all-in-one-ui-7d8d89895 SuccessfulDelete Deleted pod: all-in-one-ui-7d8d89895-786nr replicaset-controller logger.go:42: 20:32:23 | allinone | 2024-08-29 20:32:02 +0000 UTC Normal Deployment.apps all-in-one-ui ScalingReplicaSet Scaled down replica set all-in-one-ui-7d8d89895 to 0 from 1 deployment-controller logger.go:42: 20:32:23 | allinone | 2024-08-29 20:32:04 +0000 UTC Normal Pod all-in-one-ui-7d8d89895-786nr.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 20:32:23 | allinone | 2024-08-29 20:32:04 +0000 UTC Normal Pod all-in-one-ui-7d8d89895-786nr.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 20:32:23 | allinone | 2024-08-29 20:32:05 +0000 UTC Normal Pod all-in-one-ui-6954d59d98-mhzpk Binding Scheduled Successfully assigned kuttl-test-inspired-shiner/all-in-one-ui-6954d59d98-mhzpk to ip-10-0-51-237.us-west-2.compute.internal default-scheduler logger.go:42: 20:32:23 | allinone | 2024-08-29 20:32:05 +0000 UTC Normal Pod all-in-one-ui-6954d59d98-mhzpk AddedInterface Add eth0 [10.130.0.91/23] from ovn-kubernetes multus logger.go:42: 20:32:23 | allinone | 2024-08-29 20:32:05 +0000 UTC Normal Pod all-in-one-ui-6954d59d98-mhzpk.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:345d3b6c507df9f8a18948a594e7a03698085689c6965faf79b8aadc74a0ceaa" already present on machine kubelet logger.go:42: 20:32:23 | allinone | 2024-08-29 20:32:05 +0000 UTC Normal Pod all-in-one-ui-6954d59d98-mhzpk.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 20:32:23 | allinone | 2024-08-29 20:32:05 +0000 UTC Normal Pod all-in-one-ui-6954d59d98-mhzpk.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 20:32:23 | allinone | 2024-08-29 20:32:05 +0000 UTC Normal Pod all-in-one-ui-6954d59d98-mhzpk.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1613032062a68faa414cd5754a666b7b78ad0e52cf328eaaed355719423a55dd" already present on machine kubelet logger.go:42: 20:32:23 | allinone | 2024-08-29 20:32:05 +0000 UTC Normal Pod all-in-one-ui-6954d59d98-mhzpk.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 20:32:23 | allinone | 2024-08-29 20:32:05 +0000 UTC Normal Pod all-in-one-ui-6954d59d98-mhzpk.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 20:32:23 | allinone | 2024-08-29 20:32:05 +0000 UTC Normal ReplicaSet.apps all-in-one-ui-6954d59d98 SuccessfulCreate Created pod: all-in-one-ui-6954d59d98-mhzpk replicaset-controller logger.go:42: 20:32:23 | allinone | 2024-08-29 20:32:05 +0000 UTC Normal Deployment.apps all-in-one-ui ScalingReplicaSet Scaled up replica set all-in-one-ui-6954d59d98 to 1 deployment-controller logger.go:42: 20:32:23 | allinone | 2024-08-29 20:32:10 +0000 UTC Normal Pod all-in-one-ui-6954d59d98-mhzpk.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 20:32:23 | allinone | 2024-08-29 20:32:10 +0000 UTC Normal Pod all-in-one-ui-6954d59d98-mhzpk.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 20:32:23 | allinone | 2024-08-29 20:32:11 +0000 UTC Normal Pod all-in-one-ui-79df9b948b-bprp7 Binding Scheduled Successfully assigned kuttl-test-inspired-shiner/all-in-one-ui-79df9b948b-bprp7 to ip-10-0-51-237.us-west-2.compute.internal default-scheduler logger.go:42: 20:32:23 | allinone | 2024-08-29 20:32:11 +0000 UTC Normal Pod all-in-one-ui-79df9b948b-bprp7 AddedInterface Add eth0 [10.130.0.92/23] from ovn-kubernetes multus logger.go:42: 20:32:23 | allinone | 2024-08-29 20:32:11 +0000 UTC Normal Pod all-in-one-ui-79df9b948b-bprp7.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:345d3b6c507df9f8a18948a594e7a03698085689c6965faf79b8aadc74a0ceaa" already present on machine kubelet logger.go:42: 20:32:23 | allinone | 2024-08-29 20:32:11 +0000 UTC Normal Pod all-in-one-ui-79df9b948b-bprp7.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 20:32:23 | allinone | 2024-08-29 20:32:11 +0000 UTC Normal Pod all-in-one-ui-79df9b948b-bprp7.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 20:32:23 | allinone | 2024-08-29 20:32:11 +0000 UTC Normal Pod all-in-one-ui-79df9b948b-bprp7.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1613032062a68faa414cd5754a666b7b78ad0e52cf328eaaed355719423a55dd" already present on machine kubelet logger.go:42: 20:32:23 | allinone | 2024-08-29 20:32:11 +0000 UTC Normal Pod all-in-one-ui-79df9b948b-bprp7.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 20:32:23 | allinone | 2024-08-29 20:32:11 +0000 UTC Normal Pod all-in-one-ui-79df9b948b-bprp7.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 20:32:23 | allinone | 2024-08-29 20:32:11 +0000 UTC Normal ReplicaSet.apps all-in-one-ui-79df9b948b SuccessfulCreate Created pod: all-in-one-ui-79df9b948b-bprp7 replicaset-controller logger.go:42: 20:32:23 | allinone | 2024-08-29 20:32:11 +0000 UTC Normal Deployment.apps all-in-one-ui ScalingReplicaSet Scaled up replica set all-in-one-ui-79df9b948b to 1 deployment-controller logger.go:42: 20:32:23 | allinone | 2024-08-29 20:32:14 +0000 UTC Normal ReplicaSet.apps all-in-one-ui-79df9b948b SuccessfulDelete Deleted pod: all-in-one-ui-79df9b948b-bprp7 replicaset-controller logger.go:42: 20:32:23 | allinone | 2024-08-29 20:32:14 +0000 UTC Normal Deployment.apps all-in-one-ui ScalingReplicaSet Scaled down replica set all-in-one-ui-79df9b948b to 0 from 1 deployment-controller logger.go:42: 20:32:23 | allinone | 2024-08-29 20:32:15 +0000 UTC Warning Pod all-in-one-ui-79df9b948b-bprp7 FailedToRetrieveImagePullSecret Unable to retrieve some image pull secrets (all-in-one-ui-ui-proxy-dockercfg-zcxkb); attempting to pull the image may not succeed. kubelet logger.go:42: 20:32:23 | allinone | 2024-08-29 20:32:16 +0000 UTC Normal Pod all-in-one-ui-79df9b948b-bprp7.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 20:32:23 | allinone | 2024-08-29 20:32:16 +0000 UTC Normal Pod all-in-one-ui-79df9b948b-bprp7.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 20:32:23 | allinone | 2024-08-29 20:32:17 +0000 UTC Normal Pod all-in-one-ui-78d7b64fd4-t8kjb Binding Scheduled Successfully assigned kuttl-test-inspired-shiner/all-in-one-ui-78d7b64fd4-t8kjb to ip-10-0-51-237.us-west-2.compute.internal default-scheduler logger.go:42: 20:32:23 | allinone | 2024-08-29 20:32:17 +0000 UTC Normal Pod all-in-one-ui-78d7b64fd4-t8kjb AddedInterface Add eth0 [10.130.0.93/23] from ovn-kubernetes multus logger.go:42: 20:32:23 | allinone | 2024-08-29 20:32:17 +0000 UTC Normal Pod all-in-one-ui-78d7b64fd4-t8kjb.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:345d3b6c507df9f8a18948a594e7a03698085689c6965faf79b8aadc74a0ceaa" already present on machine kubelet logger.go:42: 20:32:23 | allinone | 2024-08-29 20:32:17 +0000 UTC Normal Pod all-in-one-ui-78d7b64fd4-t8kjb.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 20:32:23 | allinone | 2024-08-29 20:32:17 +0000 UTC Normal Pod all-in-one-ui-78d7b64fd4-t8kjb.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 20:32:23 | allinone | 2024-08-29 20:32:17 +0000 UTC Normal ReplicaSet.apps all-in-one-ui-78d7b64fd4 SuccessfulCreate Created pod: all-in-one-ui-78d7b64fd4-t8kjb replicaset-controller logger.go:42: 20:32:23 | allinone | 2024-08-29 20:32:17 +0000 UTC Normal Deployment.apps all-in-one-ui ScalingReplicaSet Scaled up replica set all-in-one-ui-78d7b64fd4 to 1 deployment-controller logger.go:42: 20:32:23 | allinone | Deleting namespace: kuttl-test-inspired-shiner === CONT kuttl/harness/production logger.go:42: 20:32:29 | production | Ignoring add-tracking-id.yaml as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 20:32:29 | production | Ignoring ensure-ingress-host.sh as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 20:32:29 | production | Creating namespace: kuttl-test-guided-gorilla logger.go:42: 20:32:29 | production/1-install | starting test step 1-install logger.go:42: 20:32:29 | production/1-install | Jaeger:kuttl-test-guided-gorilla/production-ui created logger.go:42: 20:42:30 | production/1-install | test step failed 1-install case.go:364: failed in step 1-install case.go:366: --- Deployment:kuttl-test-guided-gorilla/production-ui-collector +++ Deployment:kuttl-test-guided-gorilla/production-ui-collector @@ -1,10 +1,412 @@ apiVersion: apps/v1 kind: Deployment metadata: + annotations: + linkerd.io/inject: disabled + prometheus.io/port: "14269" + prometheus.io/scrape: "true" + labels: + app: jaeger + app.kubernetes.io/component: collector + app.kubernetes.io/instance: production-ui + app.kubernetes.io/managed-by: jaeger-operator + app.kubernetes.io/name: production-ui-collector + app.kubernetes.io/part-of: jaeger + managedFields: + - apiVersion: apps/v1 + fieldsType: FieldsV1 + fieldsV1: + f:metadata: + f:annotations: + .: {} + f:linkerd.io/inject: {} + f:prometheus.io/port: {} + f:prometheus.io/scrape: {} + f:labels: + .: {} + f:app: {} + f:app.kubernetes.io/component: {} + f:app.kubernetes.io/instance: {} + f:app.kubernetes.io/managed-by: {} + f:app.kubernetes.io/name: {} + f:app.kubernetes.io/part-of: {} + f:ownerReferences: + .: {} + k:{"uid":"c3f9e854-aff7-4b21-a63d-67afc142989c"}: {} + f:spec: + f:progressDeadlineSeconds: {} + f:replicas: {} + f:revisionHistoryLimit: {} + f:selector: {} + f:strategy: + f:type: {} + f:template: + f:metadata: + f:annotations: + .: {} + f:linkerd.io/inject: {} + f:prometheus.io/port: {} + f:prometheus.io/scrape: {} + f:sidecar.istio.io/inject: {} + f:labels: + .: {} + f:app: {} + f:app.kubernetes.io/component: {} + f:app.kubernetes.io/instance: {} + f:app.kubernetes.io/managed-by: {} + f:app.kubernetes.io/name: {} + f:app.kubernetes.io/part-of: {} + f:spec: + f:containers: + k:{"name":"jaeger-collector"}: + .: {} + f:args: {} + f:env: + .: {} + k:{"name":"COLLECTOR_OTLP_ENABLED"}: + .: {} + f:name: {} + f:value: {} + k:{"name":"COLLECTOR_ZIPKIN_HOST_PORT"}: + .: {} + f:name: {} + f:value: {} + k:{"name":"SPAN_STORAGE_TYPE"}: + .: {} + f:name: {} + f:value: {} + f:image: {} + f:imagePullPolicy: {} + f:livenessProbe: + .: {} + f:failureThreshold: {} + f:httpGet: + .: {} + f:path: {} + f:port: {} + f:scheme: {} + f:initialDelaySeconds: {} + f:periodSeconds: {} + f:successThreshold: {} + f:timeoutSeconds: {} + f:name: {} + f:ports: + .: {} + k:{"containerPort":4317,"protocol":"TCP"}: + .: {} + f:containerPort: {} + f:name: {} + f:protocol: {} + k:{"containerPort":4318,"protocol":"TCP"}: + .: {} + f:containerPort: {} + f:name: {} + f:protocol: {} + k:{"containerPort":9411,"protocol":"TCP"}: + .: {} + f:containerPort: {} + f:name: {} + f:protocol: {} + k:{"containerPort":14250,"protocol":"TCP"}: + .: {} + f:containerPort: {} + f:name: {} + f:protocol: {} + k:{"containerPort":14267,"protocol":"TCP"}: + .: {} + f:containerPort: {} + f:name: {} + f:protocol: {} + k:{"containerPort":14268,"protocol":"TCP"}: + .: {} + f:containerPort: {} + f:name: {} + f:protocol: {} + k:{"containerPort":14269,"protocol":"TCP"}: + .: {} + f:containerPort: {} + f:name: {} + f:protocol: {} + f:readinessProbe: + .: {} + f:failureThreshold: {} + f:httpGet: + .: {} + f:path: {} + f:port: {} + f:scheme: {} + f:initialDelaySeconds: {} + f:periodSeconds: {} + f:successThreshold: {} + f:timeoutSeconds: {} + f:resources: {} + f:terminationMessagePath: {} + f:terminationMessagePolicy: {} + f:volumeMounts: + .: {} + k:{"mountPath":"/certs"}: + .: {} + f:mountPath: {} + f:name: {} + f:readOnly: {} + k:{"mountPath":"/etc/jaeger/sampling"}: + .: {} + f:mountPath: {} + f:name: {} + f:readOnly: {} + k:{"mountPath":"/etc/pki/ca-trust/extracted/pem"}: + .: {} + f:mountPath: {} + f:name: {} + f:readOnly: {} + k:{"mountPath":"/etc/tls-config"}: + .: {} + f:mountPath: {} + f:name: {} + f:readOnly: {} + f:dnsPolicy: {} + f:enableServiceLinks: {} + f:restartPolicy: {} + f:schedulerName: {} + f:securityContext: {} + f:serviceAccount: {} + f:serviceAccountName: {} + f:terminationGracePeriodSeconds: {} + f:volumes: + .: {} + k:{"name":"certs"}: + .: {} + f:name: {} + f:secret: + .: {} + f:defaultMode: {} + f:secretName: {} + k:{"name":"production-ui-collector-tls-config-volume"}: + .: {} + f:name: {} + f:secret: + .: {} + f:defaultMode: {} + f:secretName: {} + k:{"name":"production-ui-sampling-configuration-volume"}: + .: {} + f:configMap: + .: {} + f:defaultMode: {} + f:items: {} + f:name: {} + f:name: {} + k:{"name":"production-ui-trusted-ca"}: + .: {} + f:configMap: + .: {} + f:defaultMode: {} + f:items: {} + f:name: {} + f:name: {} + manager: jaeger-operator + operation: Update + time: "2024-08-29T20:32:35Z" + - apiVersion: apps/v1 + fieldsType: FieldsV1 + fieldsV1: + f:metadata: + f:annotations: + f:deployment.kubernetes.io/revision: {} + f:status: + f:conditions: + .: {} + k:{"type":"Available"}: + .: {} + f:lastTransitionTime: {} + f:lastUpdateTime: {} + f:message: {} + f:reason: {} + f:status: {} + f:type: {} + k:{"type":"Progressing"}: + .: {} + f:lastTransitionTime: {} + f:lastUpdateTime: {} + f:message: {} + f:reason: {} + f:status: {} + f:type: {} + f:observedGeneration: {} + f:replicas: {} + f:unavailableReplicas: {} + f:updatedReplicas: {} + manager: kube-controller-manager + operation: Update + subresource: status + time: "2024-08-29T20:32:35Z" name: production-ui-collector namespace: kuttl-test-guided-gorilla + ownerReferences: + - apiVersion: jaegertracing.io/v1 + controller: true + kind: Jaeger + name: production-ui + uid: c3f9e854-aff7-4b21-a63d-67afc142989c spec: + progressDeadlineSeconds: 600 replicas: 1 + revisionHistoryLimit: 10 + selector: + matchLabels: + app: jaeger + app.kubernetes.io/component: collector + app.kubernetes.io/instance: production-ui + app.kubernetes.io/managed-by: jaeger-operator + app.kubernetes.io/name: production-ui-collector + app.kubernetes.io/part-of: jaeger + strategy: + type: Recreate + template: + metadata: + annotations: + linkerd.io/inject: disabled + prometheus.io/port: "14269" + prometheus.io/scrape: "true" + sidecar.istio.io/inject: "false" + creationTimestamp: null + labels: + app: jaeger + app.kubernetes.io/component: collector + app.kubernetes.io/instance: production-ui + app.kubernetes.io/managed-by: jaeger-operator + app.kubernetes.io/name: production-ui-collector + app.kubernetes.io/part-of: jaeger + spec: + containers: + - args: + - --collector.grpc.tls.cert=/etc/tls-config/tls.crt + - --collector.grpc.tls.enabled=true + - --collector.grpc.tls.key=/etc/tls-config/tls.key + - --sampling.strategies-file=/etc/jaeger/sampling/sampling.json + - --es.server-urls=https://elasticsearch.kuttl-test-guided-gorilla.svc.cluster.local:9200 + - --es.tls.enabled=true + - --es.tls.ca=/certs/ca + - --es.tls.cert=/certs/cert + - --es.tls.key=/certs/key + - --es.timeout=15s + - --es.num-shards=1 + - --es.num-replicas=0 + env: + - name: SPAN_STORAGE_TYPE + value: elasticsearch + - name: COLLECTOR_ZIPKIN_HOST_PORT + value: :9411 + - name: COLLECTOR_OTLP_ENABLED + value: "true" + image: registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:b90e0d264bb0b051c2a4b9e6d499f407fed58f834ab8ebf30263ee1cf7c7ea3c + imagePullPolicy: IfNotPresent + livenessProbe: + failureThreshold: 5 + httpGet: + path: / + port: 14269 + scheme: HTTP + initialDelaySeconds: 5 + periodSeconds: 15 + successThreshold: 1 + timeoutSeconds: 1 + name: jaeger-collector + ports: + - containerPort: 9411 + name: zipkin + protocol: TCP + - containerPort: 14267 + name: c-tchan-trft + protocol: TCP + - containerPort: 14268 + name: c-binary-trft + protocol: TCP + - containerPort: 14269 + name: admin-http + protocol: TCP + - containerPort: 14250 + name: grpc + protocol: TCP + - containerPort: 4317 + name: grpc-otlp + protocol: TCP + - containerPort: 4318 + name: http-otlp + protocol: TCP + readinessProbe: + failureThreshold: 3 + httpGet: + path: / + port: 14269 + scheme: HTTP + initialDelaySeconds: 1 + periodSeconds: 10 + successThreshold: 1 + timeoutSeconds: 1 + resources: {} + terminationMessagePath: /dev/termination-log + terminationMessagePolicy: File + volumeMounts: + - mountPath: /etc/jaeger/sampling + name: production-ui-sampling-configuration-volume + readOnly: true + - mountPath: /etc/tls-config + name: production-ui-collector-tls-config-volume + readOnly: true + - mountPath: /etc/pki/ca-trust/extracted/pem + name: production-ui-trusted-ca + readOnly: true + - mountPath: /certs + name: certs + readOnly: true + dnsPolicy: ClusterFirst + enableServiceLinks: false + restartPolicy: Always + schedulerName: default-scheduler + securityContext: {} + serviceAccount: production-ui + serviceAccountName: production-ui + terminationGracePeriodSeconds: 30 + volumes: + - configMap: + defaultMode: 420 + items: + - key: sampling + path: sampling.json + name: production-ui-sampling-configuration + name: production-ui-sampling-configuration-volume + - name: production-ui-collector-tls-config-volume + secret: + defaultMode: 420 + secretName: production-ui-collector-headless-tls + - configMap: + defaultMode: 420 + items: + - key: ca-bundle.crt + path: tls-ca-bundle.pem + name: production-ui-trusted-ca + name: production-ui-trusted-ca + - name: certs + secret: + defaultMode: 420 + secretName: production-ui-jaeger-elasticsearch status: - readyReplicas: 1 + conditions: + - lastTransitionTime: "2024-08-29T20:32:35Z" + lastUpdateTime: "2024-08-29T20:32:35Z" + message: Deployment does not have minimum availability. + reason: MinimumReplicasUnavailable + status: "False" + type: Available + - lastTransitionTime: "2024-08-29T20:32:35Z" + lastUpdateTime: "2024-08-29T20:32:35Z" + message: ReplicaSet "production-ui-collector-5cf4d46d9c" is progressing. + reason: ReplicaSetUpdated + status: "True" + type: Progressing + observedGeneration: 1 + replicas: 1 + unavailableReplicas: 1 + updatedReplicas: 1 case.go:366: resource Deployment:kuttl-test-guided-gorilla/production-ui-collector: .status.readyReplicas: key is missing from map case.go:366: --- Deployment:kuttl-test-guided-gorilla/production-ui-query +++ Deployment:kuttl-test-guided-gorilla/production-ui-query @@ -1,10 +1,508 @@ apiVersion: apps/v1 kind: Deployment metadata: + annotations: + linkerd.io/inject: disabled + prometheus.io/port: "16687" + prometheus.io/scrape: "true" + sidecar.jaegertracing.io/inject: production-ui + sidecar.jaegertracing.io/revision: "1" + labels: + app: jaeger + app.kubernetes.io/component: query + app.kubernetes.io/instance: production-ui + app.kubernetes.io/managed-by: jaeger-operator + app.kubernetes.io/name: production-ui-query + app.kubernetes.io/part-of: jaeger + sidecar.jaegertracing.io/injected: production-ui + managedFields: + - apiVersion: apps/v1 + fieldsType: FieldsV1 + fieldsV1: + f:metadata: + f:annotations: + .: {} + f:linkerd.io/inject: {} + f:prometheus.io/port: {} + f:prometheus.io/scrape: {} + f:sidecar.jaegertracing.io/inject: {} + f:sidecar.jaegertracing.io/revision: {} + f:labels: + .: {} + f:app: {} + f:app.kubernetes.io/component: {} + f:app.kubernetes.io/instance: {} + f:app.kubernetes.io/managed-by: {} + f:app.kubernetes.io/name: {} + f:app.kubernetes.io/part-of: {} + f:ownerReferences: + .: {} + k:{"uid":"c3f9e854-aff7-4b21-a63d-67afc142989c"}: {} + f:spec: + f:progressDeadlineSeconds: {} + f:replicas: {} + f:revisionHistoryLimit: {} + f:selector: {} + f:strategy: + f:type: {} + f:template: + f:metadata: + f:annotations: + .: {} + f:linkerd.io/inject: {} + f:prometheus.io/port: {} + f:prometheus.io/scrape: {} + f:sidecar.istio.io/inject: {} + f:sidecar.jaegertracing.io/inject: {} + f:labels: + .: {} + f:app: {} + f:app.kubernetes.io/component: {} + f:app.kubernetes.io/instance: {} + f:app.kubernetes.io/managed-by: {} + f:app.kubernetes.io/name: {} + f:app.kubernetes.io/part-of: {} + f:spec: + f:containers: + k:{"name":"jaeger-query"}: + .: {} + f:args: {} + f:env: + .: {} + k:{"name":"JAEGER_DISABLED"}: + .: {} + f:name: {} + f:value: {} + k:{"name":"METRICS_STORAGE_TYPE"}: + .: {} + f:name: {} + k:{"name":"SPAN_STORAGE_TYPE"}: + .: {} + f:name: {} + f:value: {} + f:image: {} + f:imagePullPolicy: {} + f:livenessProbe: + .: {} + f:failureThreshold: {} + f:httpGet: + .: {} + f:path: {} + f:port: {} + f:scheme: {} + f:initialDelaySeconds: {} + f:periodSeconds: {} + f:successThreshold: {} + f:timeoutSeconds: {} + f:name: {} + f:ports: + .: {} + k:{"containerPort":16685,"protocol":"TCP"}: + .: {} + f:containerPort: {} + f:name: {} + f:protocol: {} + k:{"containerPort":16686,"protocol":"TCP"}: + .: {} + f:containerPort: {} + f:name: {} + f:protocol: {} + k:{"containerPort":16687,"protocol":"TCP"}: + .: {} + f:containerPort: {} + f:name: {} + f:protocol: {} + f:readinessProbe: + .: {} + f:failureThreshold: {} + f:httpGet: + .: {} + f:path: {} + f:port: {} + f:scheme: {} + f:initialDelaySeconds: {} + f:periodSeconds: {} + f:successThreshold: {} + f:timeoutSeconds: {} + f:resources: {} + f:terminationMessagePath: {} + f:terminationMessagePolicy: {} + f:volumeMounts: + .: {} + k:{"mountPath":"/certs"}: + .: {} + f:mountPath: {} + f:name: {} + f:readOnly: {} + k:{"mountPath":"/etc/config"}: + .: {} + f:mountPath: {} + f:name: {} + f:readOnly: {} + k:{"mountPath":"/etc/pki/ca-trust/extracted/pem"}: + .: {} + f:mountPath: {} + f:name: {} + f:readOnly: {} + k:{"name":"oauth-proxy"}: + .: {} + f:args: {} + f:image: {} + f:imagePullPolicy: {} + f:name: {} + f:ports: + .: {} + k:{"containerPort":8443,"protocol":"TCP"}: + .: {} + f:containerPort: {} + f:name: {} + f:protocol: {} + f:resources: {} + f:terminationMessagePath: {} + f:terminationMessagePolicy: {} + f:volumeMounts: + .: {} + k:{"mountPath":"/etc/pki/ca-trust/extracted/pem"}: + .: {} + f:mountPath: {} + f:name: {} + f:readOnly: {} + k:{"mountPath":"/etc/tls/private"}: + .: {} + f:mountPath: {} + f:name: {} + f:dnsPolicy: {} + f:enableServiceLinks: {} + f:restartPolicy: {} + f:schedulerName: {} + f:securityContext: {} + f:serviceAccount: {} + f:serviceAccountName: {} + f:terminationGracePeriodSeconds: {} + f:volumes: + .: {} + k:{"name":"certs"}: + .: {} + f:name: {} + f:secret: + .: {} + f:defaultMode: {} + f:secretName: {} + k:{"name":"production-ui-trusted-ca"}: + .: {} + f:configMap: + .: {} + f:defaultMode: {} + f:items: {} + f:name: {} + f:name: {} + k:{"name":"production-ui-ui-configuration-volume"}: + .: {} + f:configMap: + .: {} + f:defaultMode: {} + f:items: {} + f:name: {} + f:name: {} + k:{"name":"production-ui-ui-oauth-proxy-tls"}: + .: {} + f:name: {} + f:secret: + .: {} + f:defaultMode: {} + f:secretName: {} + manager: jaeger-operator + operation: Update + time: "2024-08-29T20:37:35Z" + - apiVersion: apps/v1 + fieldsType: FieldsV1 + fieldsV1: + f:metadata: + f:annotations: + f:deployment.kubernetes.io/revision: {} + f:status: + f:conditions: + .: {} + k:{"type":"Available"}: + .: {} + f:lastTransitionTime: {} + f:lastUpdateTime: {} + f:message: {} + f:reason: {} + f:status: {} + f:type: {} + k:{"type":"Progressing"}: + .: {} + f:lastTransitionTime: {} + f:lastUpdateTime: {} + f:message: {} + f:reason: {} + f:status: {} + f:type: {} + f:observedGeneration: {} + f:replicas: {} + f:unavailableReplicas: {} + f:updatedReplicas: {} + manager: kube-controller-manager + operation: Update + subresource: status + time: "2024-08-29T20:37:35Z" name: production-ui-query namespace: kuttl-test-guided-gorilla + ownerReferences: + - apiVersion: jaegertracing.io/v1 + controller: true + kind: Jaeger + name: production-ui + uid: c3f9e854-aff7-4b21-a63d-67afc142989c spec: + progressDeadlineSeconds: 600 replicas: 1 + revisionHistoryLimit: 10 + selector: + matchLabels: + app: jaeger + app.kubernetes.io/component: query + app.kubernetes.io/instance: production-ui + app.kubernetes.io/managed-by: jaeger-operator + app.kubernetes.io/name: production-ui-query + app.kubernetes.io/part-of: jaeger + strategy: + type: Recreate + template: + metadata: + annotations: + linkerd.io/inject: disabled + prometheus.io/port: "16687" + prometheus.io/scrape: "true" + sidecar.istio.io/inject: "false" + sidecar.jaegertracing.io/inject: production-ui + creationTimestamp: null + labels: + app: jaeger + app.kubernetes.io/component: query + app.kubernetes.io/instance: production-ui + app.kubernetes.io/managed-by: jaeger-operator + app.kubernetes.io/name: production-ui-query + app.kubernetes.io/part-of: jaeger + spec: + containers: + - args: + - --query.ui-config=/etc/config/ui.json + - --es.server-urls=https://elasticsearch.kuttl-test-guided-gorilla.svc.cluster.local:9200 + - --es.tls.enabled=true + - --es.tls.ca=/certs/ca + - --es.tls.cert=/certs/cert + - --es.tls.key=/certs/key + - --es.timeout=15s + - --es.num-shards=1 + - --es.num-replicas=0 + env: + - name: SPAN_STORAGE_TYPE + value: elasticsearch + - name: METRICS_STORAGE_TYPE + - name: JAEGER_DISABLED + value: "false" + - name: JAEGER_SERVICE_NAME + value: production-ui.kuttl-test-guided-gorilla + - name: JAEGER_PROPAGATION + value: jaeger,b3,w3c + image: registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:831f83e8396d61822fb2dd576e711e0f10c1f8c7bd03bf0d4bec49dfb90cacca + imagePullPolicy: IfNotPresent + livenessProbe: + failureThreshold: 5 + httpGet: + path: / + port: 16687 + scheme: HTTP + initialDelaySeconds: 5 + periodSeconds: 15 + successThreshold: 1 + timeoutSeconds: 1 + name: jaeger-query + ports: + - containerPort: 16685 + name: grpc-query + protocol: TCP + - containerPort: 16686 + name: query + protocol: TCP + - containerPort: 16687 + name: admin-http + protocol: TCP + readinessProbe: + failureThreshold: 3 + httpGet: + path: / + port: 16687 + scheme: HTTP + initialDelaySeconds: 1 + periodSeconds: 10 + successThreshold: 1 + timeoutSeconds: 1 + resources: {} + terminationMessagePath: /dev/termination-log + terminationMessagePolicy: File + volumeMounts: + - mountPath: /etc/config + name: production-ui-ui-configuration-volume + readOnly: true + - mountPath: /etc/pki/ca-trust/extracted/pem + name: production-ui-trusted-ca + readOnly: true + - mountPath: /certs + name: certs + readOnly: true + - args: + - --cookie-secret=mNAtPPGdu6qF5NE1ui4fyN + - --https-address=:8443 + - '--openshift-sar={"namespace": "kuttl-test-guided-gorilla", "resource": + "pods", "verb": "get"}' + - --openshift-service-account=production-ui-ui-proxy + - --provider=openshift + - --tls-cert=/etc/tls/private/tls.crt + - --tls-key=/etc/tls/private/tls.key + - --upstream=http://localhost:16686 + env: + - name: JAEGER_SERVICE_NAME + value: production-ui.kuttl-test-guided-gorilla + - name: JAEGER_PROPAGATION + value: jaeger,b3,w3c + image: registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1613032062a68faa414cd5754a666b7b78ad0e52cf328eaaed355719423a55dd + imagePullPolicy: IfNotPresent + name: oauth-proxy + ports: + - containerPort: 8443 + name: public + protocol: TCP + resources: {} + terminationMessagePath: /dev/termination-log + terminationMessagePolicy: File + volumeMounts: + - mountPath: /etc/tls/private + name: production-ui-ui-oauth-proxy-tls + - mountPath: /etc/pki/ca-trust/extracted/pem + name: production-ui-trusted-ca + readOnly: true + - args: + - --agent.tags=cluster=undefined,deployment.name=production-ui-query,host.ip=${HOST_IP:},pod.name=${POD_NAME:},pod.namespace=kuttl-test-guided-gorilla + - --reporter.grpc.host-port=dns:///production-ui-collector-headless.kuttl-test-guided-gorilla.svc:14250 + - --reporter.grpc.tls.ca=/etc/pki/ca-trust/source/service-ca/service-ca.crt + - --reporter.grpc.tls.enabled=true + env: + - name: POD_NAME + valueFrom: + fieldRef: + apiVersion: v1 + fieldPath: metadata.name + - name: HOST_IP + valueFrom: + fieldRef: + apiVersion: v1 + fieldPath: status.hostIP + image: registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:5fcc9d365678f54c0f52005318dbbc32b87f743cae7e1a84568e769653f3b128 + imagePullPolicy: IfNotPresent + livenessProbe: + failureThreshold: 5 + httpGet: + path: / + port: 14271 + scheme: HTTP + initialDelaySeconds: 5 + periodSeconds: 15 + successThreshold: 1 + timeoutSeconds: 1 + name: jaeger-agent + ports: + - containerPort: 5775 + name: zk-compact-trft + protocol: UDP + - containerPort: 5778 + name: config-rest + protocol: TCP + - containerPort: 6831 + name: jg-compact-trft + protocol: UDP + - containerPort: 6832 + name: jg-binary-trft + protocol: UDP + - containerPort: 14271 + name: admin-http + protocol: TCP + readinessProbe: + failureThreshold: 3 + httpGet: + path: / + port: 14271 + scheme: HTTP + initialDelaySeconds: 1 + periodSeconds: 10 + successThreshold: 1 + timeoutSeconds: 1 + resources: {} + terminationMessagePath: /dev/termination-log + terminationMessagePolicy: File + volumeMounts: + - mountPath: /etc/pki/ca-trust/extracted/pem + name: production-ui-trusted-ca + readOnly: true + - mountPath: /etc/pki/ca-trust/source/service-ca + name: production-ui-service-ca + readOnly: true + dnsPolicy: ClusterFirst + enableServiceLinks: false + restartPolicy: Always + schedulerName: default-scheduler + securityContext: {} + serviceAccount: production-ui-ui-proxy + serviceAccountName: production-ui-ui-proxy + terminationGracePeriodSeconds: 30 + volumes: + - configMap: + defaultMode: 420 + items: + - key: ui + path: ui.json + name: production-ui-ui-configuration + name: production-ui-ui-configuration-volume + - configMap: + defaultMode: 420 + items: + - key: ca-bundle.crt + path: tls-ca-bundle.pem + name: production-ui-trusted-ca + name: production-ui-trusted-ca + - name: production-ui-ui-oauth-proxy-tls + secret: + defaultMode: 420 + secretName: production-ui-ui-oauth-proxy-tls + - name: certs + secret: + defaultMode: 420 + secretName: production-ui-jaeger-elasticsearch + - configMap: + defaultMode: 420 + items: + - key: service-ca.crt + path: service-ca.crt + name: production-ui-service-ca + name: production-ui-service-ca status: - readyReplicas: 1 + conditions: + - lastTransitionTime: "2024-08-29T20:32:35Z" + lastUpdateTime: "2024-08-29T20:32:35Z" + message: Deployment does not have minimum availability. + reason: MinimumReplicasUnavailable + status: "False" + type: Available + - lastTransitionTime: "2024-08-29T20:32:35Z" + lastUpdateTime: "2024-08-29T20:32:35Z" + message: ReplicaSet "production-ui-query-6c48fc9bf4" is progressing. + reason: ReplicaSetUpdated + status: "True" + type: Progressing + observedGeneration: 3 + replicas: 1 + unavailableReplicas: 1 + updatedReplicas: 1 case.go:366: resource Deployment:kuttl-test-guided-gorilla/production-ui-query: .status.readyReplicas: key is missing from map logger.go:42: 20:42:30 | production | production events from ns kuttl-test-guided-gorilla: logger.go:42: 20:42:30 | production | 2024-08-29 20:32:35 +0000 UTC Normal Pod production-ui-collector-5cf4d46d9c-49k5c Binding Scheduled Successfully assigned kuttl-test-guided-gorilla/production-ui-collector-5cf4d46d9c-49k5c to ip-10-0-51-237.us-west-2.compute.internal default-scheduler logger.go:42: 20:42:30 | production | 2024-08-29 20:32:35 +0000 UTC Normal ReplicaSet.apps production-ui-collector-5cf4d46d9c SuccessfulCreate Created pod: production-ui-collector-5cf4d46d9c-49k5c replicaset-controller logger.go:42: 20:42:30 | production | 2024-08-29 20:32:35 +0000 UTC Normal Deployment.apps production-ui-collector ScalingReplicaSet Scaled up replica set production-ui-collector-5cf4d46d9c to 1 deployment-controller logger.go:42: 20:42:30 | production | 2024-08-29 20:32:35 +0000 UTC Normal Pod production-ui-query-6c48fc9bf4-zsbl8 Binding Scheduled Successfully assigned kuttl-test-guided-gorilla/production-ui-query-6c48fc9bf4-zsbl8 to ip-10-0-51-237.us-west-2.compute.internal default-scheduler logger.go:42: 20:42:30 | production | 2024-08-29 20:32:35 +0000 UTC Normal Pod production-ui-query-6c48fc9bf4-zsbl8 AddedInterface Add eth0 [10.130.0.95/23] from ovn-kubernetes multus logger.go:42: 20:42:30 | production | 2024-08-29 20:32:35 +0000 UTC Normal Pod production-ui-query-6c48fc9bf4-zsbl8.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:831f83e8396d61822fb2dd576e711e0f10c1f8c7bd03bf0d4bec49dfb90cacca" already present on machine kubelet logger.go:42: 20:42:30 | production | 2024-08-29 20:32:35 +0000 UTC Normal Pod production-ui-query-6c48fc9bf4-zsbl8.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 20:42:30 | production | 2024-08-29 20:32:35 +0000 UTC Normal Pod production-ui-query-6c48fc9bf4-zsbl8.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 20:42:30 | production | 2024-08-29 20:32:35 +0000 UTC Normal Pod production-ui-query-6c48fc9bf4-zsbl8.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:1613032062a68faa414cd5754a666b7b78ad0e52cf328eaaed355719423a55dd" already present on machine kubelet logger.go:42: 20:42:30 | production | 2024-08-29 20:32:35 +0000 UTC Normal Pod production-ui-query-6c48fc9bf4-zsbl8.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 20:42:30 | production | 2024-08-29 20:32:35 +0000 UTC Normal Pod production-ui-query-6c48fc9bf4-zsbl8.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 20:42:30 | production | 2024-08-29 20:32:35 +0000 UTC Normal ReplicaSet.apps production-ui-query-6c48fc9bf4 SuccessfulCreate Created pod: production-ui-query-6c48fc9bf4-zsbl8 replicaset-controller logger.go:42: 20:42:30 | production | 2024-08-29 20:32:35 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled up replica set production-ui-query-6c48fc9bf4 to 1 deployment-controller logger.go:42: 20:42:30 | production | 2024-08-29 20:32:36 +0000 UTC Normal Pod production-ui-collector-5cf4d46d9c-49k5c AddedInterface Add eth0 [10.130.0.94/23] from ovn-kubernetes multus logger.go:42: 20:42:30 | production | 2024-08-29 20:32:36 +0000 UTC Normal Pod production-ui-collector-5cf4d46d9c-49k5c.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:b90e0d264bb0b051c2a4b9e6d499f407fed58f834ab8ebf30263ee1cf7c7ea3c" already present on machine kubelet logger.go:42: 20:42:30 | production | 2024-08-29 20:32:36 +0000 UTC Normal Pod production-ui-collector-5cf4d46d9c-49k5c.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 20:42:30 | production | 2024-08-29 20:32:36 +0000 UTC Normal Pod production-ui-collector-5cf4d46d9c-49k5c.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 20:42:30 | production | 2024-08-29 20:32:36 +0000 UTC Normal Pod production-ui-query-6c48fc9bf4-zsbl8.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:5fcc9d365678f54c0f52005318dbbc32b87f743cae7e1a84568e769653f3b128" already present on machine kubelet logger.go:42: 20:42:30 | production | 2024-08-29 20:32:36 +0000 UTC Normal Pod production-ui-query-6c48fc9bf4-zsbl8.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 20:42:30 | production | 2024-08-29 20:32:36 +0000 UTC Normal Pod production-ui-query-6c48fc9bf4-zsbl8.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 20:42:30 | production | 2024-08-29 20:32:37 +0000 UTC Warning Pod production-ui-query-6c48fc9bf4-zsbl8.spec.containers{jaeger-query} Unhealthy Readiness probe failed: HTTP probe failed with statuscode: 503 kubelet logger.go:42: 20:42:30 | production | 2024-08-29 20:32:38 +0000 UTC Warning Pod production-ui-collector-5cf4d46d9c-49k5c.spec.containers{jaeger-collector} Unhealthy Readiness probe failed: HTTP probe failed with statuscode: 503 kubelet logger.go:42: 20:42:30 | production | 2024-08-29 20:32:47 +0000 UTC Warning Pod production-ui-query-6c48fc9bf4-zsbl8.spec.containers{jaeger-query} BackOff Back-off restarting failed container jaeger-query in pod production-ui-query-6c48fc9bf4-zsbl8_kuttl-test-guided-gorilla(f8a5e91d-d2f6-4515-8006-ef1cf7e8deb1) kubelet logger.go:42: 20:42:30 | production | 2024-08-29 20:32:48 +0000 UTC Warning Pod production-ui-collector-5cf4d46d9c-49k5c.spec.containers{jaeger-collector} BackOff Back-off restarting failed container jaeger-collector in pod production-ui-collector-5cf4d46d9c-49k5c_kuttl-test-guided-gorilla(878922fd-0977-4fd9-affc-a06408bfdf7d) kubelet logger.go:42: 20:42:30 | production | 2024-08-29 20:32:50 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling production-ui-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 20:42:30 | production | 2024-08-29 20:32:50 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling production-ui-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 20:42:30 | production | 2024-08-29 20:32:50 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling production-ui-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 20:42:30 | production | Deleting namespace: kuttl-test-guided-gorilla === CONT kuttl/harness/artifacts logger.go:42: 20:42:36 | artifacts | Creating namespace: kuttl-test-witty-doe logger.go:42: 20:42:36 | artifacts | artifacts events from ns kuttl-test-witty-doe: logger.go:42: 20:42:36 | artifacts | Deleting namespace: kuttl-test-witty-doe === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- FAIL: kuttl (645.72s) --- FAIL: kuttl/harness (0.00s) --- PASS: kuttl/harness/allinone (31.63s) --- FAIL: kuttl/harness/production (607.31s) --- PASS: kuttl/harness/artifacts (6.09s) FAIL + exit_code=1 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name ui --report --output /logs/artifacts/ui.xml ./artifacts/kuttl-report.xml time="2024-08-29T20:42:43Z" level=debug msg="Setting a new name for the test suites" time="2024-08-29T20:42:43Z" level=debug msg="Removing 'artifacts' TestCase" time="2024-08-29T20:42:43Z" level=debug msg="normalizing test case names" time="2024-08-29T20:42:43Z" level=debug msg="ui/allinone -> ui_allinone" time="2024-08-29T20:42:43Z" level=debug msg="ui/production -> ui_production" time="2024-08-29T20:42:43Z" level=debug msg="ui/artifacts -> ui_artifacts" +---------------+--------+ | NAME | RESULT | +---------------+--------+ | ui_allinone | passed | | ui_production | failed | | ui_artifacts | passed | +---------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh upgrade false true + '[' 3 -ne 3 ']' + test_suite_name=upgrade + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/upgrade.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-upgrade make[2]: Entering directory '/tmp/jaeger-tests' make docker JAEGER_VERSION=1.60.1 IMG="quay.io//jaeger-operator:next" make[3]: Entering directory '/tmp/jaeger-tests' [ ! -z "true" ] || docker build --build-arg=GOPROXY= --build-arg=VERSION="1.60.1" --build-arg=JAEGER_VERSION=1.60.1 --build-arg=TARGETARCH= --build-arg VERSION_DATE=2024-08-29T20:42:43Z --build-arg VERSION_PKG="github.com/jaegertracing/jaeger-operator/pkg/version" -t "quay.io//jaeger-operator:next" . make[3]: Leaving directory '/tmp/jaeger-tests' touch build-e2e-upgrade-image SKIP_ES_EXTERNAL=true IMG=quay.io//jaeger-operator:"1.60.1" JAEGER_OPERATOR_VERSION="1.60.1" JAEGER_VERSION="1.60.0" ./tests/e2e/upgrade/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.16.8 True False 98m Cluster version is 4.16.8' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.16.8 True False 98m Cluster version is 4.16.8' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/upgrade/render.sh ++ export SUITE_DIR=./tests/e2e/upgrade ++ SUITE_DIR=./tests/e2e/upgrade ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/upgrade ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ xargs -I '{}' cp -r '{}' _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + export JAEGER_NAME + '[' true = true ']' + skip_test upgrade 'Test not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=upgrade + message='Test not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/upgrade/_build + '[' _build '!=' _build ']' + rm -rf upgrade + warning 'upgrade: Test not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: upgrade: Test not supported in OpenShift\e[0m' WAR: upgrade: Test not supported in OpenShift + '[' true = true ']' + skip_test upgrade-from-latest-release 'Test not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=upgrade-from-latest-release + message='Test not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/upgrade/_build + '[' _build '!=' _build ']' + rm -rf upgrade-from-latest-release + warning 'upgrade-from-latest-release: Test not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: upgrade-from-latest-release: Test not supported in OpenShift\e[0m' WAR: upgrade-from-latest-release: Test not supported in OpenShift make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running upgrade E2E tests' Running upgrade E2E tests + cd tests/e2e/upgrade/_build + set +e + KUBECONFIG=/tmp/kubeconfig-2030369819 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-rosa-h-yd3i.24xa.s3.devshift.org:443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 1 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === CONT kuttl/harness/artifacts logger.go:42: 20:42:45 | artifacts | Creating namespace: kuttl-test-concise-mole logger.go:42: 20:42:45 | artifacts | artifacts events from ns kuttl-test-concise-mole: logger.go:42: 20:42:45 | artifacts | Deleting namespace: kuttl-test-concise-mole === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (6.72s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (6.10s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name upgrade --report --output /logs/artifacts/upgrade.xml ./artifacts/kuttl-report.xml time="2024-08-29T20:42:51Z" level=debug msg="Setting a new name for the test suites" time="2024-08-29T20:42:51Z" level=debug msg="Removing 'artifacts' TestCase" time="2024-08-29T20:42:51Z" level=debug msg="normalizing test case names" time="2024-08-29T20:42:51Z" level=debug msg="upgrade/artifacts -> upgrade_artifacts" +-------------------+--------+ | NAME | RESULT | +-------------------+--------+ | upgrade_artifacts | passed | +-------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests'