% Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 0 0 0 0 0 0 0 0 --:--:-- 0:00:01 --:--:-- 0 0 0 0 0 0 0 0 0 --:--:-- 0:00:02 --:--:-- 0 100 3831 100 3831 0 0 1266 0 0:00:03 0:00:03 --:--:-- 1266 % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 9315 100 9315 0 0 24839 0 --:--:-- --:--:-- --:--:-- 24840 % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 2607 100 2607 0 0 16872 0 --:--:-- --:--:-- --:--:-- 16928 % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 1953 100 1953 0 0 8887 0 --:--:-- --:--:-- --:--:-- 8917 % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 350 100 350 0 0 2152 0 --:--:-- --:--:-- --:--:-- 2160 Installing kuttl Try 0... curl -sLo /tmp/jaeger-tests/hack/install/../../bin/kubectl-kuttl https://github.com/kudobuilder/kuttl/releases/download/v0.15.0/kubectl-kuttl_0.15.0_linux_x86_64 KUBECONFIG file is: /tmp/kubeconfig-2486720052 for suite in sidecar streaming upgrade miscellaneous ui elasticsearch examples generate; do \ make run-e2e-tests-$suite ; \ done make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh sidecar false true + '[' 3 -ne 3 ']' + test_suite_name=sidecar + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/sidecar.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-sidecar make[2]: Entering directory '/tmp/jaeger-tests' ./tests/e2e/sidecar/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.29 True False 12m Cluster version is 4.15.29' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.29 True False 12m Cluster version is 4.15.29' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/sidecar/render.sh ++ export SUITE_DIR=./tests/e2e/sidecar ++ SUITE_DIR=./tests/e2e/sidecar ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/sidecar ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + jaeger_service_name=order + start_test sidecar-deployment + '[' 1 -ne 1 ']' + test_name=sidecar-deployment + echo =========================================================================== =========================================================================== + info 'Rendering files for test sidecar-deployment' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test sidecar-deployment\e[0m' Rendering files for test sidecar-deployment + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/sidecar/_build + '[' _build '!=' _build ']' + mkdir -p sidecar-deployment + cd sidecar-deployment + render_install_vertx 01 + '[' 1 -ne 1 ']' + test_step=01 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-assert.yaml.template -o ./01-assert.yaml + render_find_service agent-as-sidecar allInOne order 00 03 + '[' 5 -ne 5 ']' + jaeger=agent-as-sidecar + deployment_strategy=allInOne + service_name=order + job_number=00 + test_step=03 + export JAEGER_NAME=agent-as-sidecar + JAEGER_NAME=agent-as-sidecar + export JOB_NUMBER=00 + JOB_NUMBER=00 + export SERVICE_NAME=order + SERVICE_NAME=order + export JAEGER_QUERY_ENDPOINT + '[' true = true ']' + '[' allInOne '!=' allInOne ']' + template=/tmp/jaeger-tests/tests/templates/find-service.yaml.template + JAEGER_QUERY_ENDPOINT=http://agent-as-sidecar-query:16686 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/find-service.yaml.template -o ./03-find-service.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-find-service.yaml.template -o ./03-assert.yaml + unset JAEGER_NAME + unset SERVICE_NAME + unset JOB_NUMBER + unset JAEGER_COLLECTOR_ENDPOINT + render_find_service agent-as-sidecar2 allInOne order 01 06 + '[' 5 -ne 5 ']' + jaeger=agent-as-sidecar2 + deployment_strategy=allInOne + service_name=order + job_number=01 + test_step=06 + export JAEGER_NAME=agent-as-sidecar2 + JAEGER_NAME=agent-as-sidecar2 + export JOB_NUMBER=01 + JOB_NUMBER=01 + export SERVICE_NAME=order + SERVICE_NAME=order + export JAEGER_QUERY_ENDPOINT + '[' true = true ']' + '[' allInOne '!=' allInOne ']' + template=/tmp/jaeger-tests/tests/templates/find-service.yaml.template + JAEGER_QUERY_ENDPOINT=http://agent-as-sidecar2-query:16686 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/find-service.yaml.template -o ./06-find-service.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-find-service.yaml.template -o ./06-assert.yaml + unset JAEGER_NAME + unset SERVICE_NAME + unset JOB_NUMBER + unset JAEGER_COLLECTOR_ENDPOINT + start_test sidecar-namespace + '[' 1 -ne 1 ']' + test_name=sidecar-namespace + echo =========================================================================== =========================================================================== + info 'Rendering files for test sidecar-namespace' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test sidecar-namespace\e[0m' Rendering files for test sidecar-namespace + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/sidecar/_build/sidecar-deployment + '[' sidecar-deployment '!=' _build ']' + cd .. + mkdir -p sidecar-namespace + cd sidecar-namespace + render_install_vertx 01 + '[' 1 -ne 1 ']' + test_step=01 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-assert.yaml.template -o ./01-assert.yaml + render_find_service agent-as-sidecar allInOne order 00 03 + '[' 5 -ne 5 ']' + jaeger=agent-as-sidecar + deployment_strategy=allInOne + service_name=order + job_number=00 + test_step=03 + export JAEGER_NAME=agent-as-sidecar + JAEGER_NAME=agent-as-sidecar + export JOB_NUMBER=00 + JOB_NUMBER=00 + export SERVICE_NAME=order + SERVICE_NAME=order + export JAEGER_QUERY_ENDPOINT + '[' true = true ']' + '[' allInOne '!=' allInOne ']' + template=/tmp/jaeger-tests/tests/templates/find-service.yaml.template + JAEGER_QUERY_ENDPOINT=http://agent-as-sidecar-query:16686 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/find-service.yaml.template -o ./03-find-service.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-find-service.yaml.template -o ./03-assert.yaml + unset JAEGER_NAME + unset SERVICE_NAME + unset JOB_NUMBER + unset JAEGER_COLLECTOR_ENDPOINT + render_find_service agent-as-sidecar2 allInOne order 01 06 + '[' 5 -ne 5 ']' + jaeger=agent-as-sidecar2 + deployment_strategy=allInOne + service_name=order + job_number=01 + test_step=06 + export JAEGER_NAME=agent-as-sidecar2 + JAEGER_NAME=agent-as-sidecar2 + export JOB_NUMBER=01 + JOB_NUMBER=01 + export SERVICE_NAME=order + SERVICE_NAME=order + export JAEGER_QUERY_ENDPOINT + '[' true = true ']' + '[' allInOne '!=' allInOne ']' + template=/tmp/jaeger-tests/tests/templates/find-service.yaml.template + JAEGER_QUERY_ENDPOINT=http://agent-as-sidecar2-query:16686 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/find-service.yaml.template -o ./06-find-service.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-find-service.yaml.template -o ./06-assert.yaml + unset JAEGER_NAME + unset SERVICE_NAME + unset JOB_NUMBER + unset JAEGER_COLLECTOR_ENDPOINT + start_test sidecar-skip-webhook + '[' 1 -ne 1 ']' + test_name=sidecar-skip-webhook + echo =========================================================================== =========================================================================== + info 'Rendering files for test sidecar-skip-webhook' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test sidecar-skip-webhook\e[0m' Rendering files for test sidecar-skip-webhook + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/sidecar/_build/sidecar-namespace + '[' sidecar-namespace '!=' _build ']' + cd .. + mkdir -p sidecar-skip-webhook + cd sidecar-skip-webhook + render_install_vertx 01 + '[' 1 -ne 1 ']' + test_step=01 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-assert.yaml.template -o ./01-assert.yaml make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running sidecar E2E tests' Running sidecar E2E tests + cd tests/e2e/sidecar/_build + set +e + KUBECONFIG=/tmp/kubeconfig-2486720052 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-rosa-h-u0tm.f61o.s3.devshift.org:443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 4 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/sidecar-deployment === PAUSE kuttl/harness/sidecar-deployment === RUN kuttl/harness/sidecar-namespace === PAUSE kuttl/harness/sidecar-namespace === RUN kuttl/harness/sidecar-skip-webhook === PAUSE kuttl/harness/sidecar-skip-webhook === CONT kuttl/harness/artifacts logger.go:42: 07:01:22 | artifacts | Creating namespace: kuttl-test-discrete-muskox logger.go:42: 07:01:22 | artifacts | artifacts events from ns kuttl-test-discrete-muskox: logger.go:42: 07:01:22 | artifacts | Deleting namespace: kuttl-test-discrete-muskox === CONT kuttl/harness/sidecar-namespace logger.go:42: 07:01:28 | sidecar-namespace | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:01:28 | sidecar-namespace | Creating namespace: kuttl-test-ultimate-shrimp logger.go:42: 07:01:28 | sidecar-namespace/0-install | starting test step 0-install logger.go:42: 07:01:28 | sidecar-namespace/0-install | Jaeger:kuttl-test-ultimate-shrimp/agent-as-sidecar created logger.go:42: 07:01:35 | sidecar-namespace/0-install | test step completed 0-install logger.go:42: 07:01:35 | sidecar-namespace/1-install | starting test step 1-install logger.go:42: 07:01:36 | sidecar-namespace/1-install | Deployment:kuttl-test-ultimate-shrimp/vertx-create-span-sidecar created logger.go:42: 07:01:43 | sidecar-namespace/1-install | test step completed 1-install logger.go:42: 07:01:43 | sidecar-namespace/2-enable-injection | starting test step 2-enable-injection logger.go:42: 07:01:43 | sidecar-namespace/2-enable-injection | running command: [sh -c kubectl annotate --overwrite namespaces $NAMESPACE "sidecar.jaegertracing.io/inject"="true"] logger.go:42: 07:01:44 | sidecar-namespace/2-enable-injection | namespace/kuttl-test-ultimate-shrimp annotated logger.go:42: 07:01:49 | sidecar-namespace/2-enable-injection | test step completed 2-enable-injection logger.go:42: 07:01:49 | sidecar-namespace/3-find-service | starting test step 3-find-service logger.go:42: 07:01:49 | sidecar-namespace/3-find-service | Job:kuttl-test-ultimate-shrimp/00-find-service created logger.go:42: 07:02:04 | sidecar-namespace/3-find-service | test step completed 3-find-service logger.go:42: 07:02:04 | sidecar-namespace/4-other-instance | starting test step 4-other-instance logger.go:42: 07:02:04 | sidecar-namespace/4-other-instance | Jaeger:kuttl-test-ultimate-shrimp/agent-as-sidecar2 created logger.go:42: 07:02:08 | sidecar-namespace/4-other-instance | test step completed 4-other-instance logger.go:42: 07:02:08 | sidecar-namespace/5-delete-first-instance | starting test step 5-delete-first-instance logger.go:42: 07:02:08 | sidecar-namespace/5-delete-first-instance | test step completed 5-delete-first-instance logger.go:42: 07:02:08 | sidecar-namespace/6-find-service | starting test step 6-find-service logger.go:42: 07:02:08 | sidecar-namespace/6-find-service | Job:kuttl-test-ultimate-shrimp/01-find-service created logger.go:42: 07:02:21 | sidecar-namespace/6-find-service | test step completed 6-find-service logger.go:42: 07:02:21 | sidecar-namespace/7-disable-injection | starting test step 7-disable-injection logger.go:42: 07:02:21 | sidecar-namespace/7-disable-injection | running command: [sh -c kubectl annotate --overwrite namespaces $NAMESPACE "sidecar.jaegertracing.io/inject"="false"] logger.go:42: 07:02:21 | sidecar-namespace/7-disable-injection | namespace/kuttl-test-ultimate-shrimp annotated logger.go:42: 07:02:23 | sidecar-namespace/7-disable-injection | test step completed 7-disable-injection logger.go:42: 07:02:23 | sidecar-namespace | sidecar-namespace events from ns kuttl-test-ultimate-shrimp: logger.go:42: 07:02:23 | sidecar-namespace | 2024-09-12 07:01:29 +0000 UTC Normal Pod agent-as-sidecar-6ddd965c7-h28cf Binding Scheduled Successfully assigned kuttl-test-ultimate-shrimp/agent-as-sidecar-6ddd965c7-h28cf to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:02:23 | sidecar-namespace | 2024-09-12 07:01:29 +0000 UTC Warning Pod agent-as-sidecar-6ddd965c7-h28cf FailedMount MountVolume.SetUp failed for volume "agent-as-sidecar-collector-tls-config-volume" : secret "agent-as-sidecar-collector-headless-tls" not found kubelet logger.go:42: 07:02:23 | sidecar-namespace | 2024-09-12 07:01:29 +0000 UTC Normal ReplicaSet.apps agent-as-sidecar-6ddd965c7 SuccessfulCreate Created pod: agent-as-sidecar-6ddd965c7-h28cf replicaset-controller logger.go:42: 07:02:23 | sidecar-namespace | 2024-09-12 07:01:29 +0000 UTC Normal Deployment.apps agent-as-sidecar ScalingReplicaSet Scaled up replica set agent-as-sidecar-6ddd965c7 to 1 deployment-controller logger.go:42: 07:02:23 | sidecar-namespace | 2024-09-12 07:01:30 +0000 UTC Normal Pod agent-as-sidecar-6ddd965c7-h28cf AddedInterface Add eth0 [10.130.0.11/23] from ovn-kubernetes logger.go:42: 07:02:23 | sidecar-namespace | 2024-09-12 07:01:30 +0000 UTC Normal Pod agent-as-sidecar-6ddd965c7-h28cf.spec.containers{jaeger} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" kubelet logger.go:42: 07:02:23 | sidecar-namespace | 2024-09-12 07:01:33 +0000 UTC Normal Pod agent-as-sidecar-6ddd965c7-h28cf.spec.containers{jaeger} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" in 2.915s (2.915s including waiting) kubelet logger.go:42: 07:02:23 | sidecar-namespace | 2024-09-12 07:01:33 +0000 UTC Normal Pod agent-as-sidecar-6ddd965c7-h28cf.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:02:23 | sidecar-namespace | 2024-09-12 07:01:33 +0000 UTC Normal Pod agent-as-sidecar-6ddd965c7-h28cf.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:02:23 | sidecar-namespace | 2024-09-12 07:01:36 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-ppshn Binding Scheduled Successfully assigned kuttl-test-ultimate-shrimp/vertx-create-span-sidecar-84d458b68c-ppshn to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:02:23 | sidecar-namespace | 2024-09-12 07:01:36 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-ppshn AddedInterface Add eth0 [10.130.0.12/23] from ovn-kubernetes logger.go:42: 07:02:23 | sidecar-namespace | 2024-09-12 07:01:36 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-ppshn.spec.containers{vertx-create-span-sidecar} Pulling Pulling image "jaegertracing/vertx-create-span:operator-e2e-tests" kubelet logger.go:42: 07:02:23 | sidecar-namespace | 2024-09-12 07:01:36 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-84d458b68c SuccessfulCreate Created pod: vertx-create-span-sidecar-84d458b68c-ppshn replicaset-controller logger.go:42: 07:02:23 | sidecar-namespace | 2024-09-12 07:01:36 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-84d458b68c to 1 deployment-controller logger.go:42: 07:02:23 | sidecar-namespace | 2024-09-12 07:01:42 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-ppshn.spec.containers{vertx-create-span-sidecar} Pulled Successfully pulled image "jaegertracing/vertx-create-span:operator-e2e-tests" in 6.448s (6.448s including waiting) kubelet logger.go:42: 07:02:23 | sidecar-namespace | 2024-09-12 07:01:43 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-ppshn.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:02:23 | sidecar-namespace | 2024-09-12 07:01:43 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-ppshn.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:02:23 | sidecar-namespace | 2024-09-12 07:01:44 +0000 UTC Normal Pod vertx-create-span-sidecar-9cc6f8685-98d26 Binding Scheduled Successfully assigned kuttl-test-ultimate-shrimp/vertx-create-span-sidecar-9cc6f8685-98d26 to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:02:23 | sidecar-namespace | 2024-09-12 07:01:44 +0000 UTC Normal Pod vertx-create-span-sidecar-9cc6f8685-98d26 AddedInterface Add eth0 [10.130.0.13/23] from ovn-kubernetes logger.go:42: 07:02:23 | sidecar-namespace | 2024-09-12 07:01:44 +0000 UTC Normal Pod vertx-create-span-sidecar-9cc6f8685-98d26.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:02:23 | sidecar-namespace | 2024-09-12 07:01:44 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-9cc6f8685 SuccessfulCreate Created pod: vertx-create-span-sidecar-9cc6f8685-98d26 replicaset-controller logger.go:42: 07:02:23 | sidecar-namespace | 2024-09-12 07:01:44 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-9cc6f8685 to 1 deployment-controller logger.go:42: 07:02:23 | sidecar-namespace | 2024-09-12 07:01:45 +0000 UTC Normal Pod vertx-create-span-sidecar-9cc6f8685-98d26.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:02:23 | sidecar-namespace | 2024-09-12 07:01:45 +0000 UTC Normal Pod vertx-create-span-sidecar-9cc6f8685-98d26.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:02:23 | sidecar-namespace | 2024-09-12 07:01:45 +0000 UTC Normal Pod vertx-create-span-sidecar-9cc6f8685-98d26.spec.containers{jaeger-agent} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" kubelet logger.go:42: 07:02:23 | sidecar-namespace | 2024-09-12 07:01:48 +0000 UTC Normal Pod vertx-create-span-sidecar-9cc6f8685-98d26.spec.containers{jaeger-agent} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" in 3.538s (3.538s including waiting) kubelet logger.go:42: 07:02:23 | sidecar-namespace | 2024-09-12 07:01:48 +0000 UTC Normal Pod vertx-create-span-sidecar-9cc6f8685-98d26.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:02:23 | sidecar-namespace | 2024-09-12 07:01:48 +0000 UTC Normal Pod vertx-create-span-sidecar-9cc6f8685-98d26.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:02:23 | sidecar-namespace | 2024-09-12 07:01:49 +0000 UTC Normal Pod 00-find-service-m8bdj Binding Scheduled Successfully assigned kuttl-test-ultimate-shrimp/00-find-service-m8bdj to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:02:23 | sidecar-namespace | 2024-09-12 07:01:49 +0000 UTC Normal Pod 00-find-service-m8bdj AddedInterface Add eth0 [10.130.0.14/23] from ovn-kubernetes logger.go:42: 07:02:23 | sidecar-namespace | 2024-09-12 07:01:49 +0000 UTC Normal Pod 00-find-service-m8bdj.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:02:23 | sidecar-namespace | 2024-09-12 07:01:49 +0000 UTC Normal Job.batch 00-find-service SuccessfulCreate Created pod: 00-find-service-m8bdj job-controller logger.go:42: 07:02:23 | sidecar-namespace | 2024-09-12 07:01:51 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-ppshn.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.130.0.12:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:02:23 | sidecar-namespace | 2024-09-12 07:01:51 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-ppshn.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.130.0.12:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:02:23 | sidecar-namespace | 2024-09-12 07:01:52 +0000 UTC Normal Pod 00-find-service-m8bdj.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 2.996s (2.996s including waiting) kubelet logger.go:42: 07:02:23 | sidecar-namespace | 2024-09-12 07:01:52 +0000 UTC Normal Pod 00-find-service-m8bdj.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:02:23 | sidecar-namespace | 2024-09-12 07:01:52 +0000 UTC Normal Pod 00-find-service-m8bdj.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:02:23 | sidecar-namespace | 2024-09-12 07:01:53 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-ppshn.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 07:02:23 | sidecar-namespace | 2024-09-12 07:01:53 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-ppshn.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.130.0.12:8080/": read tcp 10.130.0.2:35276->10.130.0.12:8080: read: connection reset by peer kubelet logger.go:42: 07:02:23 | sidecar-namespace | 2024-09-12 07:01:53 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-ppshn.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.130.0.12:8080/": dial tcp 10.130.0.12:8080: connect: connection refused kubelet logger.go:42: 07:02:23 | sidecar-namespace | 2024-09-12 07:01:53 +0000 UTC Warning Pod vertx-create-span-sidecar-9cc6f8685-98d26.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.130.0.13:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:02:23 | sidecar-namespace | 2024-09-12 07:01:53 +0000 UTC Warning Pod vertx-create-span-sidecar-9cc6f8685-98d26.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.130.0.13:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:02:23 | sidecar-namespace | 2024-09-12 07:01:54 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-ppshn.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:02:23 | sidecar-namespace | 2024-09-12 07:01:56 +0000 UTC Normal Pod vertx-create-span-sidecar-9cc6f8685-98d26.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 07:02:23 | sidecar-namespace | 2024-09-12 07:01:56 +0000 UTC Warning Pod vertx-create-span-sidecar-9cc6f8685-98d26.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.130.0.13:8080/": read tcp 10.130.0.2:49268->10.130.0.13:8080: read: connection reset by peer kubelet logger.go:42: 07:02:23 | sidecar-namespace | 2024-09-12 07:01:56 +0000 UTC Warning Pod vertx-create-span-sidecar-9cc6f8685-98d26.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.130.0.13:8080/": dial tcp 10.130.0.13:8080: connect: connection refused kubelet logger.go:42: 07:02:23 | sidecar-namespace | 2024-09-12 07:02:03 +0000 UTC Normal Job.batch 00-find-service Completed Job completed job-controller logger.go:42: 07:02:23 | sidecar-namespace | 2024-09-12 07:02:06 +0000 UTC Normal Pod agent-as-sidecar2-65ff4667bf-qpdhw Binding Scheduled Successfully assigned kuttl-test-ultimate-shrimp/agent-as-sidecar2-65ff4667bf-qpdhw to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:02:23 | sidecar-namespace | 2024-09-12 07:02:06 +0000 UTC Normal Pod agent-as-sidecar2-65ff4667bf-qpdhw AddedInterface Add eth0 [10.130.0.15/23] from ovn-kubernetes logger.go:42: 07:02:23 | sidecar-namespace | 2024-09-12 07:02:06 +0000 UTC Normal Pod agent-as-sidecar2-65ff4667bf-qpdhw.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" already present on machine kubelet logger.go:42: 07:02:23 | sidecar-namespace | 2024-09-12 07:02:06 +0000 UTC Normal Pod agent-as-sidecar2-65ff4667bf-qpdhw.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:02:23 | sidecar-namespace | 2024-09-12 07:02:06 +0000 UTC Normal Pod agent-as-sidecar2-65ff4667bf-qpdhw.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:02:23 | sidecar-namespace | 2024-09-12 07:02:06 +0000 UTC Normal ReplicaSet.apps agent-as-sidecar2-65ff4667bf SuccessfulCreate Created pod: agent-as-sidecar2-65ff4667bf-qpdhw replicaset-controller logger.go:42: 07:02:23 | sidecar-namespace | 2024-09-12 07:02:06 +0000 UTC Normal Deployment.apps agent-as-sidecar2 ScalingReplicaSet Scaled up replica set agent-as-sidecar2-65ff4667bf to 1 deployment-controller logger.go:42: 07:02:23 | sidecar-namespace | 2024-09-12 07:02:08 +0000 UTC Normal Pod 01-find-service-9c45d Binding Scheduled Successfully assigned kuttl-test-ultimate-shrimp/01-find-service-9c45d to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:02:23 | sidecar-namespace | 2024-09-12 07:02:08 +0000 UTC Normal Job.batch 01-find-service SuccessfulCreate Created pod: 01-find-service-9c45d job-controller logger.go:42: 07:02:23 | sidecar-namespace | 2024-09-12 07:02:08 +0000 UTC Normal Pod agent-as-sidecar-6ddd965c7-h28cf.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:02:23 | sidecar-namespace | 2024-09-12 07:02:08 +0000 UTC Normal Pod vertx-create-span-sidecar-5586fbb949-9v88p Binding Scheduled Successfully assigned kuttl-test-ultimate-shrimp/vertx-create-span-sidecar-5586fbb949-9v88p to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:02:23 | sidecar-namespace | 2024-09-12 07:02:08 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-5586fbb949 SuccessfulCreate Created pod: vertx-create-span-sidecar-5586fbb949-9v88p replicaset-controller logger.go:42: 07:02:23 | sidecar-namespace | 2024-09-12 07:02:08 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-84d458b68c SuccessfulDelete Deleted pod: vertx-create-span-sidecar-84d458b68c-ppshn replicaset-controller logger.go:42: 07:02:23 | sidecar-namespace | 2024-09-12 07:02:08 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled down replica set vertx-create-span-sidecar-84d458b68c to 0 from 1 deployment-controller logger.go:42: 07:02:23 | sidecar-namespace | 2024-09-12 07:02:08 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-5586fbb949 to 1 from 0 deployment-controller logger.go:42: 07:02:23 | sidecar-namespace | 2024-09-12 07:02:09 +0000 UTC Normal Pod 01-find-service-9c45d AddedInterface Add eth0 [10.130.0.16/23] from ovn-kubernetes logger.go:42: 07:02:23 | sidecar-namespace | 2024-09-12 07:02:09 +0000 UTC Normal Pod 01-find-service-9c45d.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:02:23 | sidecar-namespace | 2024-09-12 07:02:09 +0000 UTC Normal Pod vertx-create-span-sidecar-5586fbb949-9v88p AddedInterface Add eth0 [10.130.0.17/23] from ovn-kubernetes logger.go:42: 07:02:23 | sidecar-namespace | 2024-09-12 07:02:09 +0000 UTC Normal Pod vertx-create-span-sidecar-5586fbb949-9v88p.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:02:23 | sidecar-namespace | 2024-09-12 07:02:09 +0000 UTC Normal Pod vertx-create-span-sidecar-5586fbb949-9v88p.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:02:23 | sidecar-namespace | 2024-09-12 07:02:09 +0000 UTC Normal Pod vertx-create-span-sidecar-5586fbb949-9v88p.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:02:23 | sidecar-namespace | 2024-09-12 07:02:09 +0000 UTC Normal Pod vertx-create-span-sidecar-5586fbb949-9v88p.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" already present on machine kubelet logger.go:42: 07:02:23 | sidecar-namespace | 2024-09-12 07:02:09 +0000 UTC Normal Pod vertx-create-span-sidecar-5586fbb949-9v88p.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:02:23 | sidecar-namespace | 2024-09-12 07:02:09 +0000 UTC Normal Pod vertx-create-span-sidecar-5586fbb949-9v88p.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:02:23 | sidecar-namespace | 2024-09-12 07:02:10 +0000 UTC Normal Pod 01-find-service-9c45d.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 843ms (843ms including waiting) kubelet logger.go:42: 07:02:23 | sidecar-namespace | 2024-09-12 07:02:10 +0000 UTC Normal Pod 01-find-service-9c45d.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:02:23 | sidecar-namespace | 2024-09-12 07:02:10 +0000 UTC Normal Pod 01-find-service-9c45d.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:02:23 | sidecar-namespace | 2024-09-12 07:02:17 +0000 UTC Warning Pod vertx-create-span-sidecar-5586fbb949-9v88p.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.130.0.17:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:02:23 | sidecar-namespace | 2024-09-12 07:02:17 +0000 UTC Warning Pod vertx-create-span-sidecar-5586fbb949-9v88p.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.130.0.17:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:02:23 | sidecar-namespace | 2024-09-12 07:02:19 +0000 UTC Normal Pod vertx-create-span-sidecar-5586fbb949-9v88p.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 07:02:23 | sidecar-namespace | 2024-09-12 07:02:19 +0000 UTC Warning Pod vertx-create-span-sidecar-5586fbb949-9v88p.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.130.0.17:8080/": read tcp 10.130.0.2:38552->10.130.0.17:8080: read: connection reset by peer kubelet logger.go:42: 07:02:23 | sidecar-namespace | 2024-09-12 07:02:19 +0000 UTC Warning Pod vertx-create-span-sidecar-5586fbb949-9v88p.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.130.0.17:8080/": dial tcp 10.130.0.17:8080: connect: connection refused kubelet logger.go:42: 07:02:23 | sidecar-namespace | 2024-09-12 07:02:20 +0000 UTC Normal Job.batch 01-find-service Completed Job completed job-controller logger.go:42: 07:02:23 | sidecar-namespace | 2024-09-12 07:02:21 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-9cc6f8685 SuccessfulDelete Deleted pod: vertx-create-span-sidecar-9cc6f8685-98d26 replicaset-controller logger.go:42: 07:02:23 | sidecar-namespace | 2024-09-12 07:02:21 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled down replica set vertx-create-span-sidecar-9cc6f8685 to 0 from 1 deployment-controller logger.go:42: 07:02:23 | sidecar-namespace | 2024-09-12 07:02:22 +0000 UTC Normal Pod vertx-create-span-sidecar-5696c45b75-vpvhk Binding Scheduled Successfully assigned kuttl-test-ultimate-shrimp/vertx-create-span-sidecar-5696c45b75-vpvhk to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:02:23 | sidecar-namespace | 2024-09-12 07:02:22 +0000 UTC Normal Pod vertx-create-span-sidecar-5696c45b75-vpvhk AddedInterface Add eth0 [10.130.0.18/23] from ovn-kubernetes logger.go:42: 07:02:23 | sidecar-namespace | 2024-09-12 07:02:22 +0000 UTC Normal Pod vertx-create-span-sidecar-5696c45b75-vpvhk.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:02:23 | sidecar-namespace | 2024-09-12 07:02:22 +0000 UTC Normal Pod vertx-create-span-sidecar-5696c45b75-vpvhk.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:02:23 | sidecar-namespace | 2024-09-12 07:02:22 +0000 UTC Normal Pod vertx-create-span-sidecar-5696c45b75-vpvhk.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:02:23 | sidecar-namespace | 2024-09-12 07:02:22 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-5696c45b75 SuccessfulCreate Created pod: vertx-create-span-sidecar-5696c45b75-vpvhk replicaset-controller logger.go:42: 07:02:23 | sidecar-namespace | 2024-09-12 07:02:22 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-5696c45b75 to 1 from 0 deployment-controller logger.go:42: 07:02:23 | sidecar-namespace | Deleting namespace: kuttl-test-ultimate-shrimp === CONT kuttl/harness/sidecar-skip-webhook logger.go:42: 07:02:30 | sidecar-skip-webhook | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:02:30 | sidecar-skip-webhook | Creating namespace: kuttl-test-sweeping-coral logger.go:42: 07:02:30 | sidecar-skip-webhook/0-install | starting test step 0-install logger.go:42: 07:02:31 | sidecar-skip-webhook/0-install | Jaeger:kuttl-test-sweeping-coral/agent-as-sidecar created logger.go:42: 07:02:34 | sidecar-skip-webhook/0-install | test step completed 0-install logger.go:42: 07:02:34 | sidecar-skip-webhook/1-install | starting test step 1-install logger.go:42: 07:02:34 | sidecar-skip-webhook/1-install | Deployment:kuttl-test-sweeping-coral/vertx-create-span-sidecar created logger.go:42: 07:02:36 | sidecar-skip-webhook/1-install | test step completed 1-install logger.go:42: 07:02:36 | sidecar-skip-webhook/2-add-anotation-and-label | starting test step 2-add-anotation-and-label logger.go:42: 07:02:36 | sidecar-skip-webhook/2-add-anotation-and-label | running command: [kubectl label deployment vertx-create-span-sidecar app.kubernetes.io/name=jaeger-operator --namespace kuttl-test-sweeping-coral] logger.go:42: 07:02:36 | sidecar-skip-webhook/2-add-anotation-and-label | deployment.apps/vertx-create-span-sidecar labeled logger.go:42: 07:02:36 | sidecar-skip-webhook/2-add-anotation-and-label | running command: [kubectl annotate --overwrite deployment vertx-create-span-sidecar sidecar.jaegertracing.io/inject=true --namespace kuttl-test-sweeping-coral] logger.go:42: 07:02:36 | sidecar-skip-webhook/2-add-anotation-and-label | deployment.apps/vertx-create-span-sidecar annotated logger.go:42: 07:02:37 | sidecar-skip-webhook/2-add-anotation-and-label | test step completed 2-add-anotation-and-label logger.go:42: 07:02:37 | sidecar-skip-webhook/3-remove-label | starting test step 3-remove-label logger.go:42: 07:02:37 | sidecar-skip-webhook/3-remove-label | running command: [kubectl label deployment vertx-create-span-sidecar app.kubernetes.io/name- --namespace kuttl-test-sweeping-coral] logger.go:42: 07:02:37 | sidecar-skip-webhook/3-remove-label | deployment.apps/vertx-create-span-sidecar unlabeled logger.go:42: 07:02:38 | sidecar-skip-webhook/3-remove-label | test step completed 3-remove-label logger.go:42: 07:02:38 | sidecar-skip-webhook | sidecar-skip-webhook events from ns kuttl-test-sweeping-coral: logger.go:42: 07:02:38 | sidecar-skip-webhook | 2024-09-12 07:02:31 +0000 UTC Normal Pod agent-as-sidecar-c9f67f896-ckdkt Binding Scheduled Successfully assigned kuttl-test-sweeping-coral/agent-as-sidecar-c9f67f896-ckdkt to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:02:38 | sidecar-skip-webhook | 2024-09-12 07:02:31 +0000 UTC Normal ReplicaSet.apps agent-as-sidecar-c9f67f896 SuccessfulCreate Created pod: agent-as-sidecar-c9f67f896-ckdkt replicaset-controller logger.go:42: 07:02:38 | sidecar-skip-webhook | 2024-09-12 07:02:31 +0000 UTC Normal Deployment.apps agent-as-sidecar ScalingReplicaSet Scaled up replica set agent-as-sidecar-c9f67f896 to 1 deployment-controller logger.go:42: 07:02:38 | sidecar-skip-webhook | 2024-09-12 07:02:32 +0000 UTC Normal Pod agent-as-sidecar-c9f67f896-ckdkt AddedInterface Add eth0 [10.130.0.19/23] from ovn-kubernetes logger.go:42: 07:02:38 | sidecar-skip-webhook | 2024-09-12 07:02:32 +0000 UTC Normal Pod agent-as-sidecar-c9f67f896-ckdkt.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" already present on machine kubelet logger.go:42: 07:02:38 | sidecar-skip-webhook | 2024-09-12 07:02:32 +0000 UTC Normal Pod agent-as-sidecar-c9f67f896-ckdkt.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:02:38 | sidecar-skip-webhook | 2024-09-12 07:02:32 +0000 UTC Normal Pod agent-as-sidecar-c9f67f896-ckdkt.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:02:38 | sidecar-skip-webhook | 2024-09-12 07:02:34 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-8xrm4 Binding Scheduled Successfully assigned kuttl-test-sweeping-coral/vertx-create-span-sidecar-84d458b68c-8xrm4 to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:02:38 | sidecar-skip-webhook | 2024-09-12 07:02:34 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-84d458b68c SuccessfulCreate Created pod: vertx-create-span-sidecar-84d458b68c-8xrm4 replicaset-controller logger.go:42: 07:02:38 | sidecar-skip-webhook | 2024-09-12 07:02:34 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-84d458b68c to 1 deployment-controller logger.go:42: 07:02:38 | sidecar-skip-webhook | 2024-09-12 07:02:35 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-8xrm4 AddedInterface Add eth0 [10.130.0.20/23] from ovn-kubernetes logger.go:42: 07:02:38 | sidecar-skip-webhook | 2024-09-12 07:02:35 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-8xrm4.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:02:38 | sidecar-skip-webhook | 2024-09-12 07:02:35 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-8xrm4.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:02:38 | sidecar-skip-webhook | 2024-09-12 07:02:35 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-8xrm4.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:02:38 | sidecar-skip-webhook | 2024-09-12 07:02:36 +0000 UTC Normal Pod vertx-create-span-sidecar-68d6ddffcd-clfkg Binding Scheduled Successfully assigned kuttl-test-sweeping-coral/vertx-create-span-sidecar-68d6ddffcd-clfkg to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:02:38 | sidecar-skip-webhook | 2024-09-12 07:02:36 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-68d6ddffcd SuccessfulCreate Created pod: vertx-create-span-sidecar-68d6ddffcd-clfkg replicaset-controller logger.go:42: 07:02:38 | sidecar-skip-webhook | 2024-09-12 07:02:36 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-68d6ddffcd to 1 deployment-controller logger.go:42: 07:02:38 | sidecar-skip-webhook | 2024-09-12 07:02:37 +0000 UTC Normal Pod vertx-create-span-sidecar-68d6ddffcd-clfkg AddedInterface Add eth0 [10.130.0.21/23] from ovn-kubernetes logger.go:42: 07:02:38 | sidecar-skip-webhook | 2024-09-12 07:02:37 +0000 UTC Normal Pod vertx-create-span-sidecar-68d6ddffcd-clfkg.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:02:38 | sidecar-skip-webhook | 2024-09-12 07:02:37 +0000 UTC Normal Pod vertx-create-span-sidecar-68d6ddffcd-clfkg.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:02:38 | sidecar-skip-webhook | 2024-09-12 07:02:37 +0000 UTC Normal Pod vertx-create-span-sidecar-68d6ddffcd-clfkg.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:02:38 | sidecar-skip-webhook | 2024-09-12 07:02:37 +0000 UTC Normal Pod vertx-create-span-sidecar-68d6ddffcd-clfkg.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" already present on machine kubelet logger.go:42: 07:02:38 | sidecar-skip-webhook | 2024-09-12 07:02:37 +0000 UTC Normal Pod vertx-create-span-sidecar-68d6ddffcd-clfkg.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:02:38 | sidecar-skip-webhook | 2024-09-12 07:02:37 +0000 UTC Normal Pod vertx-create-span-sidecar-68d6ddffcd-clfkg.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:02:39 | sidecar-skip-webhook | Deleting namespace: kuttl-test-sweeping-coral === CONT kuttl/harness/sidecar-deployment logger.go:42: 07:02:45 | sidecar-deployment | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:02:45 | sidecar-deployment | Creating namespace: kuttl-test-nice-ray logger.go:42: 07:02:45 | sidecar-deployment/0-install | starting test step 0-install logger.go:42: 07:02:45 | sidecar-deployment/0-install | Jaeger:kuttl-test-nice-ray/agent-as-sidecar created logger.go:42: 07:02:49 | sidecar-deployment/0-install | test step completed 0-install logger.go:42: 07:02:49 | sidecar-deployment/1-install | starting test step 1-install logger.go:42: 07:02:49 | sidecar-deployment/1-install | Deployment:kuttl-test-nice-ray/vertx-create-span-sidecar created logger.go:42: 07:02:51 | sidecar-deployment/1-install | test step completed 1-install logger.go:42: 07:02:51 | sidecar-deployment/2-enable-injection | starting test step 2-enable-injection logger.go:42: 07:02:51 | sidecar-deployment/2-enable-injection | running command: [kubectl annotate --overwrite deployment vertx-create-span-sidecar sidecar.jaegertracing.io/inject=true --namespace kuttl-test-nice-ray] logger.go:42: 07:02:52 | sidecar-deployment/2-enable-injection | deployment.apps/vertx-create-span-sidecar annotated logger.go:42: 07:02:53 | sidecar-deployment/2-enable-injection | test step completed 2-enable-injection logger.go:42: 07:02:53 | sidecar-deployment/3-find-service | starting test step 3-find-service logger.go:42: 07:02:53 | sidecar-deployment/3-find-service | Job:kuttl-test-nice-ray/00-find-service created logger.go:42: 07:03:06 | sidecar-deployment/3-find-service | test step completed 3-find-service logger.go:42: 07:03:06 | sidecar-deployment/4-other-instance | starting test step 4-other-instance logger.go:42: 07:03:06 | sidecar-deployment/4-other-instance | Jaeger:kuttl-test-nice-ray/agent-as-sidecar2 created logger.go:42: 07:03:10 | sidecar-deployment/4-other-instance | test step completed 4-other-instance logger.go:42: 07:03:10 | sidecar-deployment/5-delete-first-instance | starting test step 5-delete-first-instance logger.go:42: 07:03:10 | sidecar-deployment/5-delete-first-instance | test step completed 5-delete-first-instance logger.go:42: 07:03:10 | sidecar-deployment/6-find-service | starting test step 6-find-service logger.go:42: 07:03:11 | sidecar-deployment/6-find-service | Job:kuttl-test-nice-ray/01-find-service created logger.go:42: 07:03:31 | sidecar-deployment/6-find-service | test step completed 6-find-service logger.go:42: 07:03:31 | sidecar-deployment/7-disable-injection | starting test step 7-disable-injection logger.go:42: 07:03:31 | sidecar-deployment/7-disable-injection | running command: [kubectl annotate --overwrite deployment vertx-create-span-sidecar sidecar.jaegertracing.io/inject=false --namespace kuttl-test-nice-ray] logger.go:42: 07:03:31 | sidecar-deployment/7-disable-injection | deployment.apps/vertx-create-span-sidecar annotated logger.go:42: 07:03:34 | sidecar-deployment/7-disable-injection | test step completed 7-disable-injection logger.go:42: 07:03:34 | sidecar-deployment | sidecar-deployment events from ns kuttl-test-nice-ray: logger.go:42: 07:03:34 | sidecar-deployment | 2024-09-12 07:02:46 +0000 UTC Normal Pod agent-as-sidecar-78c9bf8b96-99vz9 Binding Scheduled Successfully assigned kuttl-test-nice-ray/agent-as-sidecar-78c9bf8b96-99vz9 to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:03:34 | sidecar-deployment | 2024-09-12 07:02:46 +0000 UTC Normal Pod agent-as-sidecar-78c9bf8b96-99vz9 AddedInterface Add eth0 [10.130.0.22/23] from ovn-kubernetes logger.go:42: 07:03:34 | sidecar-deployment | 2024-09-12 07:02:46 +0000 UTC Normal Pod agent-as-sidecar-78c9bf8b96-99vz9.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" already present on machine kubelet logger.go:42: 07:03:34 | sidecar-deployment | 2024-09-12 07:02:46 +0000 UTC Normal Pod agent-as-sidecar-78c9bf8b96-99vz9.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:03:34 | sidecar-deployment | 2024-09-12 07:02:46 +0000 UTC Normal Pod agent-as-sidecar-78c9bf8b96-99vz9.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:03:34 | sidecar-deployment | 2024-09-12 07:02:46 +0000 UTC Normal ReplicaSet.apps agent-as-sidecar-78c9bf8b96 SuccessfulCreate Created pod: agent-as-sidecar-78c9bf8b96-99vz9 replicaset-controller logger.go:42: 07:03:34 | sidecar-deployment | 2024-09-12 07:02:46 +0000 UTC Normal Deployment.apps agent-as-sidecar ScalingReplicaSet Scaled up replica set agent-as-sidecar-78c9bf8b96 to 1 deployment-controller logger.go:42: 07:03:34 | sidecar-deployment | 2024-09-12 07:02:49 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-n4xbx Binding Scheduled Successfully assigned kuttl-test-nice-ray/vertx-create-span-sidecar-84d458b68c-n4xbx to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:03:34 | sidecar-deployment | 2024-09-12 07:02:49 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-n4xbx AddedInterface Add eth0 [10.130.0.23/23] from ovn-kubernetes logger.go:42: 07:03:34 | sidecar-deployment | 2024-09-12 07:02:49 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-n4xbx.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:03:34 | sidecar-deployment | 2024-09-12 07:02:49 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-n4xbx.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:03:34 | sidecar-deployment | 2024-09-12 07:02:49 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-n4xbx.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:03:34 | sidecar-deployment | 2024-09-12 07:02:49 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-84d458b68c SuccessfulCreate Created pod: vertx-create-span-sidecar-84d458b68c-n4xbx replicaset-controller logger.go:42: 07:03:34 | sidecar-deployment | 2024-09-12 07:02:49 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-84d458b68c to 1 deployment-controller logger.go:42: 07:03:34 | sidecar-deployment | 2024-09-12 07:02:52 +0000 UTC Normal Pod vertx-create-span-sidecar-54b59f9c8b-ktlf2 Binding Scheduled Successfully assigned kuttl-test-nice-ray/vertx-create-span-sidecar-54b59f9c8b-ktlf2 to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:03:34 | sidecar-deployment | 2024-09-12 07:02:52 +0000 UTC Normal Pod vertx-create-span-sidecar-54b59f9c8b-ktlf2 AddedInterface Add eth0 [10.130.0.24/23] from ovn-kubernetes logger.go:42: 07:03:34 | sidecar-deployment | 2024-09-12 07:02:52 +0000 UTC Normal Pod vertx-create-span-sidecar-54b59f9c8b-ktlf2.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:03:34 | sidecar-deployment | 2024-09-12 07:02:52 +0000 UTC Normal Pod vertx-create-span-sidecar-54b59f9c8b-ktlf2.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:03:34 | sidecar-deployment | 2024-09-12 07:02:52 +0000 UTC Normal Pod vertx-create-span-sidecar-54b59f9c8b-ktlf2.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:03:34 | sidecar-deployment | 2024-09-12 07:02:52 +0000 UTC Normal Pod vertx-create-span-sidecar-54b59f9c8b-ktlf2.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" already present on machine kubelet logger.go:42: 07:03:34 | sidecar-deployment | 2024-09-12 07:02:52 +0000 UTC Normal Pod vertx-create-span-sidecar-54b59f9c8b-ktlf2.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:03:34 | sidecar-deployment | 2024-09-12 07:02:52 +0000 UTC Normal Pod vertx-create-span-sidecar-54b59f9c8b-ktlf2.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:03:34 | sidecar-deployment | 2024-09-12 07:02:52 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-54b59f9c8b SuccessfulCreate Created pod: vertx-create-span-sidecar-54b59f9c8b-ktlf2 replicaset-controller logger.go:42: 07:03:34 | sidecar-deployment | 2024-09-12 07:02:52 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-54b59f9c8b to 1 deployment-controller logger.go:42: 07:03:34 | sidecar-deployment | 2024-09-12 07:02:53 +0000 UTC Normal Pod 00-find-service-ppfcn Binding Scheduled Successfully assigned kuttl-test-nice-ray/00-find-service-ppfcn to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:03:34 | sidecar-deployment | 2024-09-12 07:02:53 +0000 UTC Normal Job.batch 00-find-service SuccessfulCreate Created pod: 00-find-service-ppfcn job-controller logger.go:42: 07:03:34 | sidecar-deployment | 2024-09-12 07:02:54 +0000 UTC Normal Pod 00-find-service-ppfcn AddedInterface Add eth0 [10.130.0.25/23] from ovn-kubernetes logger.go:42: 07:03:34 | sidecar-deployment | 2024-09-12 07:02:54 +0000 UTC Normal Pod 00-find-service-ppfcn.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:03:34 | sidecar-deployment | 2024-09-12 07:02:55 +0000 UTC Normal Pod 00-find-service-ppfcn.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 801ms (801ms including waiting) kubelet logger.go:42: 07:03:34 | sidecar-deployment | 2024-09-12 07:02:55 +0000 UTC Normal Pod 00-find-service-ppfcn.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:03:34 | sidecar-deployment | 2024-09-12 07:02:55 +0000 UTC Normal Pod 00-find-service-ppfcn.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:03:34 | sidecar-deployment | 2024-09-12 07:02:57 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-n4xbx.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.130.0.23:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:03:34 | sidecar-deployment | 2024-09-12 07:02:57 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-n4xbx.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.130.0.23:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:03:34 | sidecar-deployment | 2024-09-12 07:02:59 +0000 UTC Normal Pod vertx-create-span-sidecar-84d458b68c-n4xbx.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 07:03:34 | sidecar-deployment | 2024-09-12 07:03:00 +0000 UTC Warning Pod vertx-create-span-sidecar-54b59f9c8b-ktlf2.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.130.0.24:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:03:34 | sidecar-deployment | 2024-09-12 07:03:00 +0000 UTC Warning Pod vertx-create-span-sidecar-54b59f9c8b-ktlf2.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.130.0.24:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:03:34 | sidecar-deployment | 2024-09-12 07:03:00 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-n4xbx.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.130.0.23:8080/": read tcp 10.130.0.2:39658->10.130.0.23:8080: read: connection reset by peer kubelet logger.go:42: 07:03:34 | sidecar-deployment | 2024-09-12 07:03:00 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-n4xbx.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.130.0.23:8080/": dial tcp 10.130.0.23:8080: connect: connection refused kubelet logger.go:42: 07:03:34 | sidecar-deployment | 2024-09-12 07:03:02 +0000 UTC Normal Pod vertx-create-span-sidecar-54b59f9c8b-ktlf2.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 07:03:34 | sidecar-deployment | 2024-09-12 07:03:02 +0000 UTC Warning Pod vertx-create-span-sidecar-54b59f9c8b-ktlf2.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.130.0.24:8080/": read tcp 10.130.0.2:58834->10.130.0.24:8080: read: connection reset by peer kubelet logger.go:42: 07:03:34 | sidecar-deployment | 2024-09-12 07:03:02 +0000 UTC Warning Pod vertx-create-span-sidecar-54b59f9c8b-ktlf2.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.130.0.24:8080/": dial tcp 10.130.0.24:8080: connect: connection refused kubelet logger.go:42: 07:03:34 | sidecar-deployment | 2024-09-12 07:03:06 +0000 UTC Normal Job.batch 00-find-service Completed Job completed job-controller logger.go:42: 07:03:34 | sidecar-deployment | 2024-09-12 07:03:07 +0000 UTC Normal Pod agent-as-sidecar2-c46744b7d-qxcrc Binding Scheduled Successfully assigned kuttl-test-nice-ray/agent-as-sidecar2-c46744b7d-qxcrc to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:03:34 | sidecar-deployment | 2024-09-12 07:03:07 +0000 UTC Warning Pod agent-as-sidecar2-c46744b7d-qxcrc FailedMount MountVolume.SetUp failed for volume "agent-as-sidecar2-collector-tls-config-volume" : secret "agent-as-sidecar2-collector-headless-tls" not found kubelet logger.go:42: 07:03:34 | sidecar-deployment | 2024-09-12 07:03:07 +0000 UTC Normal ReplicaSet.apps agent-as-sidecar2-c46744b7d SuccessfulCreate Created pod: agent-as-sidecar2-c46744b7d-qxcrc replicaset-controller logger.go:42: 07:03:34 | sidecar-deployment | 2024-09-12 07:03:07 +0000 UTC Normal Deployment.apps agent-as-sidecar2 ScalingReplicaSet Scaled up replica set agent-as-sidecar2-c46744b7d to 1 deployment-controller logger.go:42: 07:03:34 | sidecar-deployment | 2024-09-12 07:03:08 +0000 UTC Normal Pod agent-as-sidecar2-c46744b7d-qxcrc AddedInterface Add eth0 [10.130.0.26/23] from ovn-kubernetes logger.go:42: 07:03:34 | sidecar-deployment | 2024-09-12 07:03:08 +0000 UTC Normal Pod agent-as-sidecar2-c46744b7d-qxcrc.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" already present on machine kubelet logger.go:42: 07:03:34 | sidecar-deployment | 2024-09-12 07:03:08 +0000 UTC Normal Pod agent-as-sidecar2-c46744b7d-qxcrc.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:03:34 | sidecar-deployment | 2024-09-12 07:03:08 +0000 UTC Normal Pod agent-as-sidecar2-c46744b7d-qxcrc.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:03:34 | sidecar-deployment | 2024-09-12 07:03:10 +0000 UTC Normal Pod agent-as-sidecar-78c9bf8b96-99vz9.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:03:34 | sidecar-deployment | 2024-09-12 07:03:11 +0000 UTC Normal Pod 01-find-service-jt6n5 Binding Scheduled Successfully assigned kuttl-test-nice-ray/01-find-service-jt6n5 to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:03:34 | sidecar-deployment | 2024-09-12 07:03:11 +0000 UTC Normal Pod 01-find-service-jt6n5 AddedInterface Add eth0 [10.130.0.27/23] from ovn-kubernetes logger.go:42: 07:03:34 | sidecar-deployment | 2024-09-12 07:03:11 +0000 UTC Normal Pod 01-find-service-jt6n5.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:03:34 | sidecar-deployment | 2024-09-12 07:03:11 +0000 UTC Normal Job.batch 01-find-service SuccessfulCreate Created pod: 01-find-service-jt6n5 job-controller logger.go:42: 07:03:34 | sidecar-deployment | 2024-09-12 07:03:11 +0000 UTC Warning Pod vertx-create-span-sidecar-84d458b68c-n4xbx.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.130.0.23:8080/": read tcp 10.130.0.2:54066->10.130.0.23:8080: read: connection reset by peer kubelet logger.go:42: 07:03:34 | sidecar-deployment | 2024-09-12 07:03:12 +0000 UTC Normal Pod 01-find-service-jt6n5.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 819ms (819ms including waiting) kubelet logger.go:42: 07:03:34 | sidecar-deployment | 2024-09-12 07:03:12 +0000 UTC Normal Pod 01-find-service-jt6n5.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:03:34 | sidecar-deployment | 2024-09-12 07:03:12 +0000 UTC Normal Pod 01-find-service-jt6n5.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:03:34 | sidecar-deployment | 2024-09-12 07:03:13 +0000 UTC Normal Pod vertx-create-span-sidecar-5d4747c8f4-zvndr Binding Scheduled Successfully assigned kuttl-test-nice-ray/vertx-create-span-sidecar-5d4747c8f4-zvndr to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:03:34 | sidecar-deployment | 2024-09-12 07:03:13 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-5d4747c8f4 SuccessfulCreate Created pod: vertx-create-span-sidecar-5d4747c8f4-zvndr replicaset-controller logger.go:42: 07:03:34 | sidecar-deployment | 2024-09-12 07:03:13 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-84d458b68c SuccessfulDelete Deleted pod: vertx-create-span-sidecar-84d458b68c-n4xbx replicaset-controller logger.go:42: 07:03:34 | sidecar-deployment | 2024-09-12 07:03:13 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled down replica set vertx-create-span-sidecar-84d458b68c to 0 from 1 deployment-controller logger.go:42: 07:03:34 | sidecar-deployment | 2024-09-12 07:03:13 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-5d4747c8f4 to 1 from 0 deployment-controller logger.go:42: 07:03:34 | sidecar-deployment | 2024-09-12 07:03:14 +0000 UTC Normal Pod vertx-create-span-sidecar-5d4747c8f4-zvndr AddedInterface Add eth0 [10.130.0.28/23] from ovn-kubernetes logger.go:42: 07:03:34 | sidecar-deployment | 2024-09-12 07:03:14 +0000 UTC Normal Pod vertx-create-span-sidecar-5d4747c8f4-zvndr.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:03:34 | sidecar-deployment | 2024-09-12 07:03:14 +0000 UTC Normal Pod vertx-create-span-sidecar-5d4747c8f4-zvndr.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:03:34 | sidecar-deployment | 2024-09-12 07:03:14 +0000 UTC Normal Pod vertx-create-span-sidecar-5d4747c8f4-zvndr.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:03:34 | sidecar-deployment | 2024-09-12 07:03:14 +0000 UTC Normal Pod vertx-create-span-sidecar-5d4747c8f4-zvndr.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" already present on machine kubelet logger.go:42: 07:03:34 | sidecar-deployment | 2024-09-12 07:03:14 +0000 UTC Normal Pod vertx-create-span-sidecar-5d4747c8f4-zvndr.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:03:34 | sidecar-deployment | 2024-09-12 07:03:14 +0000 UTC Normal Pod vertx-create-span-sidecar-5d4747c8f4-zvndr.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:03:34 | sidecar-deployment | 2024-09-12 07:03:22 +0000 UTC Warning Pod vertx-create-span-sidecar-5d4747c8f4-zvndr.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.130.0.28:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:03:34 | sidecar-deployment | 2024-09-12 07:03:22 +0000 UTC Warning Pod vertx-create-span-sidecar-5d4747c8f4-zvndr.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.130.0.28:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:03:34 | sidecar-deployment | 2024-09-12 07:03:24 +0000 UTC Normal Pod vertx-create-span-sidecar-5d4747c8f4-zvndr.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 07:03:34 | sidecar-deployment | 2024-09-12 07:03:24 +0000 UTC Warning Pod vertx-create-span-sidecar-5d4747c8f4-zvndr.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.130.0.28:8080/": read tcp 10.130.0.2:42618->10.130.0.28:8080: read: connection reset by peer kubelet logger.go:42: 07:03:34 | sidecar-deployment | 2024-09-12 07:03:24 +0000 UTC Warning Pod vertx-create-span-sidecar-5d4747c8f4-zvndr.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.130.0.28:8080/": dial tcp 10.130.0.28:8080: connect: connection refused kubelet logger.go:42: 07:03:34 | sidecar-deployment | 2024-09-12 07:03:31 +0000 UTC Normal Job.batch 01-find-service Completed Job completed job-controller logger.go:42: 07:03:34 | sidecar-deployment | 2024-09-12 07:03:31 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-54b59f9c8b SuccessfulDelete Deleted pod: vertx-create-span-sidecar-54b59f9c8b-ktlf2 replicaset-controller logger.go:42: 07:03:34 | sidecar-deployment | 2024-09-12 07:03:31 +0000 UTC Normal Pod vertx-create-span-sidecar-6b4f4cd58-5pxrr Binding Scheduled Successfully assigned kuttl-test-nice-ray/vertx-create-span-sidecar-6b4f4cd58-5pxrr to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:03:34 | sidecar-deployment | 2024-09-12 07:03:31 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-6b4f4cd58 SuccessfulCreate Created pod: vertx-create-span-sidecar-6b4f4cd58-5pxrr replicaset-controller logger.go:42: 07:03:34 | sidecar-deployment | 2024-09-12 07:03:31 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled down replica set vertx-create-span-sidecar-54b59f9c8b to 0 from 1 deployment-controller logger.go:42: 07:03:34 | sidecar-deployment | 2024-09-12 07:03:31 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-6b4f4cd58 to 1 from 0 deployment-controller logger.go:42: 07:03:34 | sidecar-deployment | 2024-09-12 07:03:32 +0000 UTC Normal Pod vertx-create-span-sidecar-6b4f4cd58-5pxrr AddedInterface Add eth0 [10.130.0.29/23] from ovn-kubernetes logger.go:42: 07:03:34 | sidecar-deployment | 2024-09-12 07:03:32 +0000 UTC Normal Pod vertx-create-span-sidecar-6b4f4cd58-5pxrr.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:03:34 | sidecar-deployment | 2024-09-12 07:03:32 +0000 UTC Normal Pod vertx-create-span-sidecar-6b4f4cd58-5pxrr.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:03:34 | sidecar-deployment | 2024-09-12 07:03:32 +0000 UTC Normal Pod vertx-create-span-sidecar-6b4f4cd58-5pxrr.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:03:34 | sidecar-deployment | Deleting namespace: kuttl-test-nice-ray === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (139.68s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (6.10s) --- PASS: kuttl/harness/sidecar-namespace (62.27s) --- PASS: kuttl/harness/sidecar-skip-webhook (14.50s) --- PASS: kuttl/harness/sidecar-deployment (56.21s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name sidecar --report --output /logs/artifacts/sidecar.xml ./artifacts/kuttl-report.xml time="2024-09-12T07:03:42Z" level=debug msg="Setting a new name for the test suites" time="2024-09-12T07:03:42Z" level=debug msg="Removing 'artifacts' TestCase" time="2024-09-12T07:03:42Z" level=debug msg="normalizing test case names" time="2024-09-12T07:03:42Z" level=debug msg="sidecar/artifacts -> sidecar_artifacts" time="2024-09-12T07:03:42Z" level=debug msg="sidecar/sidecar-namespace -> sidecar_sidecar_namespace" time="2024-09-12T07:03:42Z" level=debug msg="sidecar/sidecar-skip-webhook -> sidecar_sidecar_skip_webhook" time="2024-09-12T07:03:42Z" level=debug msg="sidecar/sidecar-deployment -> sidecar_sidecar_deployment" +------------------------------+--------+ | NAME | RESULT | +------------------------------+--------+ | sidecar_artifacts | passed | | sidecar_sidecar_namespace | passed | | sidecar_sidecar_skip_webhook | passed | | sidecar_sidecar_deployment | passed | +------------------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh streaming false true + '[' 3 -ne 3 ']' + test_suite_name=streaming + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/streaming.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-streaming make[2]: Entering directory '/tmp/jaeger-tests' >>>> Elasticsearch image not loaded because SKIP_ES_EXTERNAL is true KAFKA_VERSION=3.6.0 \ SKIP_KAFKA=false \ SKIP_ES_EXTERNAL=true \ ./tests/e2e/streaming/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.29 True False 15m Cluster version is 4.15.29' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.29 True False 15m Cluster version is 4.15.29' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/streaming/render.sh ++ export SUITE_DIR=./tests/e2e/streaming ++ SUITE_DIR=./tests/e2e/streaming ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/streaming ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + is_secured=false + '[' true = true ']' + is_secured=true + '[' false = true ']' + start_test streaming-simple + '[' 1 -ne 1 ']' + test_name=streaming-simple + echo =========================================================================== =========================================================================== + info 'Rendering files for test streaming-simple' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test streaming-simple\e[0m' Rendering files for test streaming-simple + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/streaming/_build + '[' _build '!=' _build ']' + mkdir -p streaming-simple + cd streaming-simple + render_install_kafka my-cluster 00 + '[' 2 -ne 2 ']' + cluster_name=my-cluster + test_step=00 + CLUSTER_NAME=my-cluster + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/kafka-install.yaml.template -o ./00-install.yaml + render_assert_kafka false my-cluster 00 + '[' 3 -ne 3 ']' + autoprovisioned=false + cluster_name=my-cluster + test_step=00 + '[' false = true ']' + '[' false = true ']' + '[' false = false ']' + replicas=1 + CLUSTER_NAME=my-cluster + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-zookeeper-cluster.yaml.template -o ./00-assert.yaml ++ expr 00 + 1 + CLUSTER_NAME=my-cluster + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-kafka-cluster.yaml.template -o ./01-assert.yaml ++ expr 00 + 2 + CLUSTER_NAME=my-cluster + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-entity-operator.yaml.template -o ./02-assert.yaml + render_install_elasticsearch upstream 03 + '[' 2 -ne 2 ']' + deploy_mode=upstream + test_step=03 + '[' upstream = upstream ']' + '[' true = true ']' + template=/tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template + /tmp/jaeger-tests/bin/yq eval -s '"elasticsearch_" + $index' /tmp/jaeger-tests/tests/elasticsearch.yml + /tmp/jaeger-tests/bin/yq eval -i '.spec.template.spec.serviceAccountName="deploy-elasticsearch"' ./elasticsearch_0.yml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template -o ./03-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/elasticsearch-assert.yaml.template -o ./03-assert.yaml + JAEGER_NAME=simple-streaming + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/streaming-jaeger-assert.yaml.template -o ./04-assert.yaml + render_smoke_test simple-streaming true 05 + '[' 3 -ne 3 ']' + jaeger=simple-streaming + is_secured=true + test_step=05 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simple-streaming-query:443 + JAEGER_QUERY_ENDPOINT=https://simple-streaming-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simple-streaming-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simple-streaming-collector-headless:14268 + export JAEGER_NAME=simple-streaming + JAEGER_NAME=simple-streaming + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./05-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./05-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + '[' false = true ']' + start_test streaming-with-tls + '[' 1 -ne 1 ']' + test_name=streaming-with-tls + echo =========================================================================== =========================================================================== + info 'Rendering files for test streaming-with-tls' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test streaming-with-tls\e[0m' Rendering files for test streaming-with-tls + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/streaming/_build/streaming-simple + '[' streaming-simple '!=' _build ']' + cd .. + mkdir -p streaming-with-tls + cd streaming-with-tls + render_install_kafka my-cluster 00 + '[' 2 -ne 2 ']' + cluster_name=my-cluster + test_step=00 + CLUSTER_NAME=my-cluster + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/kafka-install.yaml.template -o ./00-install.yaml + render_assert_kafka false my-cluster 00 + '[' 3 -ne 3 ']' + autoprovisioned=false + cluster_name=my-cluster + test_step=00 + '[' false = true ']' + '[' false = true ']' + '[' false = false ']' + replicas=1 + CLUSTER_NAME=my-cluster + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-zookeeper-cluster.yaml.template -o ./00-assert.yaml ++ expr 00 + 1 + CLUSTER_NAME=my-cluster + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-kafka-cluster.yaml.template -o ./01-assert.yaml ++ expr 00 + 2 + CLUSTER_NAME=my-cluster + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-entity-operator.yaml.template -o ./02-assert.yaml + render_install_elasticsearch upstream 03 + '[' 2 -ne 2 ']' + deploy_mode=upstream + test_step=03 + '[' upstream = upstream ']' + '[' true = true ']' + template=/tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template + /tmp/jaeger-tests/bin/yq eval -s '"elasticsearch_" + $index' /tmp/jaeger-tests/tests/elasticsearch.yml + /tmp/jaeger-tests/bin/yq eval -i '.spec.template.spec.serviceAccountName="deploy-elasticsearch"' ./elasticsearch_0.yml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template -o ./03-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/elasticsearch-assert.yaml.template -o ./03-assert.yaml + render_smoke_test tls-streaming true 05 + '[' 3 -ne 3 ']' + jaeger=tls-streaming + is_secured=true + test_step=05 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://tls-streaming-query:443 + JAEGER_QUERY_ENDPOINT=https://tls-streaming-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://tls-streaming-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://tls-streaming-collector-headless:14268 + export JAEGER_NAME=tls-streaming + JAEGER_NAME=tls-streaming + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./05-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./05-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + '[' false = true ']' + start_test streaming-with-autoprovisioning-autoscale + '[' 1 -ne 1 ']' + test_name=streaming-with-autoprovisioning-autoscale + echo =========================================================================== =========================================================================== + info 'Rendering files for test streaming-with-autoprovisioning-autoscale' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test streaming-with-autoprovisioning-autoscale\e[0m' Rendering files for test streaming-with-autoprovisioning-autoscale + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/streaming/_build/streaming-with-tls + '[' streaming-with-tls '!=' _build ']' + cd .. + mkdir -p streaming-with-autoprovisioning-autoscale + cd streaming-with-autoprovisioning-autoscale + '[' true = true ']' + rm ./00-install.yaml ./00-assert.yaml + render_install_elasticsearch upstream 01 + '[' 2 -ne 2 ']' + deploy_mode=upstream + test_step=01 + '[' upstream = upstream ']' + '[' true = true ']' + template=/tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template + /tmp/jaeger-tests/bin/yq eval -s '"elasticsearch_" + $index' /tmp/jaeger-tests/tests/elasticsearch.yml + /tmp/jaeger-tests/bin/yq eval -i '.spec.template.spec.serviceAccountName="deploy-elasticsearch"' ./elasticsearch_0.yml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/elasticsearch-assert.yaml.template -o ./01-assert.yaml + jaeger_name=auto-provisioned + /tmp/jaeger-tests/bin/yq e -i '.spec.ingester.resources.requests.memory="20Mi"' ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.ingester.resources.requests.memory="500m"' ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.ingester.autoscale=true ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.ingester.minReplicas=1 ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.ingester.maxReplicas=2 ./02-install.yaml + render_assert_kafka true auto-provisioned 03 + '[' 3 -ne 3 ']' + autoprovisioned=true + cluster_name=auto-provisioned + test_step=03 + '[' true = true ']' + is_kafka_minimal_enabled + namespaces=(observability openshift-operators openshift-distributed-tracing) + for i in "${namespaces[@]}" ++ kubectl get pods -n observability -l name=jaeger-operator -o yaml ++ /tmp/jaeger-tests/bin/yq e '.items[0].spec.containers[0].env[] | select(.name=="KAFKA-PROVISIONING-MINIMAL").value' + enabled= + '[' '' == true ']' + for i in "${namespaces[@]}" ++ kubectl get pods -n openshift-operators -l name=jaeger-operator -o yaml ++ /tmp/jaeger-tests/bin/yq e '.items[0].spec.containers[0].env[] | select(.name=="KAFKA-PROVISIONING-MINIMAL").value' + enabled= + '[' '' == true ']' + for i in "${namespaces[@]}" ++ kubectl get pods -n openshift-distributed-tracing -l name=jaeger-operator -o yaml ++ /tmp/jaeger-tests/bin/yq e '.items[0].spec.containers[0].env[] | select(.name=="KAFKA-PROVISIONING-MINIMAL").value' + enabled=true + '[' true == true ']' + return 0 + replicas=1 + CLUSTER_NAME=auto-provisioned + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-zookeeper-cluster.yaml.template -o ./03-assert.yaml ++ expr 03 + 1 + CLUSTER_NAME=auto-provisioned + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-kafka-cluster.yaml.template -o ./04-assert.yaml ++ expr 03 + 2 + CLUSTER_NAME=auto-provisioned + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-entity-operator.yaml.template -o ./05-assert.yaml + version_lt 1.30 1.23 ++ echo 1.30 1.23 ++ tr ' ' '\n' ++ sort -rV ++ head -n 1 + test 1.30 '!=' 1.30 + rm ./08-assert.yaml + skip_test streaming-with-tls 'This test is flaky in Prow CI' + '[' 2 -ne 2 ']' + test_name=streaming-with-tls + message='This test is flaky in Prow CI' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/streaming/_build/streaming-with-autoprovisioning-autoscale + '[' streaming-with-autoprovisioning-autoscale '!=' _build ']' + cd .. + rm -rf streaming-with-tls + warning 'streaming-with-tls: This test is flaky in Prow CI' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: streaming-with-tls: This test is flaky in Prow CI\e[0m' WAR: streaming-with-tls: This test is flaky in Prow CI + skip_test streaming-simple 'This test is flaky in Prow CI' + '[' 2 -ne 2 ']' + test_name=streaming-simple + message='This test is flaky in Prow CI' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/streaming/_build + '[' _build '!=' _build ']' + rm -rf streaming-simple + warning 'streaming-simple: This test is flaky in Prow CI' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: streaming-simple: This test is flaky in Prow CI\e[0m' WAR: streaming-simple: This test is flaky in Prow CI make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running streaming E2E tests' Running streaming E2E tests + cd tests/e2e/streaming/_build + set +e + KUBECONFIG=/tmp/kubeconfig-2486720052 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-rosa-h-u0tm.f61o.s3.devshift.org:443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 2 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/streaming-with-autoprovisioning-autoscale === PAUSE kuttl/harness/streaming-with-autoprovisioning-autoscale === CONT kuttl/harness/artifacts logger.go:42: 07:03:56 | artifacts | Creating namespace: kuttl-test-tidy-redbird logger.go:42: 07:03:56 | artifacts | artifacts events from ns kuttl-test-tidy-redbird: logger.go:42: 07:03:56 | artifacts | Deleting namespace: kuttl-test-tidy-redbird === CONT kuttl/harness/streaming-with-autoprovisioning-autoscale logger.go:42: 07:04:02 | streaming-with-autoprovisioning-autoscale | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:04:02 | streaming-with-autoprovisioning-autoscale | Ignoring elasticsearch_0.yml as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:04:02 | streaming-with-autoprovisioning-autoscale | Ignoring elasticsearch_1.yml as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:04:02 | streaming-with-autoprovisioning-autoscale | Creating namespace: kuttl-test-bursting-ewe logger.go:42: 07:04:02 | streaming-with-autoprovisioning-autoscale/1-install | starting test step 1-install logger.go:42: 07:04:02 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c oc create sa deploy-elasticsearch -n $NAMESPACE 2>&1 | grep -v "already exists" || true] logger.go:42: 07:04:02 | streaming-with-autoprovisioning-autoscale/1-install | serviceaccount/deploy-elasticsearch created logger.go:42: 07:04:02 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c oc adm policy add-scc-to-user privileged -z deploy-elasticsearch -n $NAMESPACE 2>&1 | grep -v "already exists" || true] logger.go:42: 07:04:02 | streaming-with-autoprovisioning-autoscale/1-install | clusterrole.rbac.authorization.k8s.io/system:openshift:scc:privileged added: "deploy-elasticsearch" logger.go:42: 07:04:02 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c sleep 6] logger.go:42: 07:04:08 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c kubectl apply -f elasticsearch_0.yml -n $NAMESPACE] logger.go:42: 07:04:09 | streaming-with-autoprovisioning-autoscale/1-install | statefulset.apps/elasticsearch created logger.go:42: 07:04:09 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c sleep 3] logger.go:42: 07:04:12 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c kubectl apply -f elasticsearch_1.yml -n $NAMESPACE] logger.go:42: 07:04:13 | streaming-with-autoprovisioning-autoscale/1-install | service/elasticsearch created logger.go:42: 07:04:30 | streaming-with-autoprovisioning-autoscale/1-install | test step completed 1-install logger.go:42: 07:04:30 | streaming-with-autoprovisioning-autoscale/2-install | starting test step 2-install logger.go:42: 07:04:30 | streaming-with-autoprovisioning-autoscale/2-install | Jaeger:kuttl-test-bursting-ewe/auto-provisioned created logger.go:42: 07:04:30 | streaming-with-autoprovisioning-autoscale/2-install | test step completed 2-install logger.go:42: 07:04:30 | streaming-with-autoprovisioning-autoscale/3- | starting test step 3- logger.go:42: 07:06:50 | streaming-with-autoprovisioning-autoscale/3- | test step completed 3- logger.go:42: 07:06:50 | streaming-with-autoprovisioning-autoscale/4- | starting test step 4- logger.go:42: 07:07:21 | streaming-with-autoprovisioning-autoscale/4- | test step completed 4- logger.go:42: 07:07:21 | streaming-with-autoprovisioning-autoscale/5- | starting test step 5- logger.go:42: 07:07:45 | streaming-with-autoprovisioning-autoscale/5- | test step completed 5- logger.go:42: 07:07:45 | streaming-with-autoprovisioning-autoscale/6- | starting test step 6- logger.go:42: 07:08:01 | streaming-with-autoprovisioning-autoscale/6- | test step completed 6- logger.go:42: 07:08:01 | streaming-with-autoprovisioning-autoscale/7- | starting test step 7- logger.go:42: 07:08:02 | streaming-with-autoprovisioning-autoscale/7- | test step completed 7- logger.go:42: 07:08:02 | streaming-with-autoprovisioning-autoscale | streaming-with-autoprovisioning-autoscale events from ns kuttl-test-bursting-ewe: logger.go:42: 07:08:02 | streaming-with-autoprovisioning-autoscale | 2024-09-12 07:04:09 +0000 UTC Normal Pod elasticsearch-0 Binding Scheduled Successfully assigned kuttl-test-bursting-ewe/elasticsearch-0 to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:08:02 | streaming-with-autoprovisioning-autoscale | 2024-09-12 07:04:09 +0000 UTC Normal StatefulSet.apps elasticsearch SuccessfulCreate create Pod elasticsearch-0 in StatefulSet elasticsearch successful statefulset-controller logger.go:42: 07:08:02 | streaming-with-autoprovisioning-autoscale | 2024-09-12 07:04:10 +0000 UTC Normal Pod elasticsearch-0 AddedInterface Add eth0 [10.130.0.30/23] from ovn-kubernetes logger.go:42: 07:08:02 | streaming-with-autoprovisioning-autoscale | 2024-09-12 07:04:10 +0000 UTC Normal Pod elasticsearch-0.spec.containers{elasticsearch} Pulling Pulling image "docker.elastic.co/elasticsearch/elasticsearch-oss:6.8.6" kubelet logger.go:42: 07:08:02 | streaming-with-autoprovisioning-autoscale | 2024-09-12 07:04:19 +0000 UTC Normal Pod elasticsearch-0.spec.containers{elasticsearch} Pulled Successfully pulled image "docker.elastic.co/elasticsearch/elasticsearch-oss:6.8.6" in 9.504s (9.504s including waiting) kubelet logger.go:42: 07:08:02 | streaming-with-autoprovisioning-autoscale | 2024-09-12 07:04:19 +0000 UTC Normal Pod elasticsearch-0.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:08:02 | streaming-with-autoprovisioning-autoscale | 2024-09-12 07:04:19 +0000 UTC Normal Pod elasticsearch-0.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:08:02 | streaming-with-autoprovisioning-autoscale | 2024-09-12 07:04:24 +0000 UTC Warning Pod elasticsearch-0.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Get "http://10.130.0.30:9200/": dial tcp 10.130.0.30:9200: connect: connection refused kubelet logger.go:42: 07:08:02 | streaming-with-autoprovisioning-autoscale | 2024-09-12 07:06:09 +0000 UTC Normal PersistentVolumeClaim data-auto-provisioned-zookeeper-0 WaitForFirstConsumer waiting for first consumer to be created before binding persistentvolume-controller logger.go:42: 07:08:02 | streaming-with-autoprovisioning-autoscale | 2024-09-12 07:06:10 +0000 UTC Normal PodDisruptionBudget.policy auto-provisioned-zookeeper NoPods No matching pods found controllermanager logger.go:42: 07:08:02 | streaming-with-autoprovisioning-autoscale | 2024-09-12 07:06:10 +0000 UTC Normal PersistentVolumeClaim data-auto-provisioned-zookeeper-0 Provisioning External provisioner is provisioning volume for claim "kuttl-test-bursting-ewe/data-auto-provisioned-zookeeper-0" ebs.csi.aws.com_aws-ebs-csi-driver-controller-5664c47c89-rmlpw_d292fa2a-4ddd-48f3-92b9-8ab46e71f3c8 logger.go:42: 07:08:02 | streaming-with-autoprovisioning-autoscale | 2024-09-12 07:06:10 +0000 UTC Normal PersistentVolumeClaim data-auto-provisioned-zookeeper-0 ExternalProvisioning Waiting for a volume to be created either by the external provisioner 'ebs.csi.aws.com' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered. persistentvolume-controller logger.go:42: 07:08:02 | streaming-with-autoprovisioning-autoscale | 2024-09-12 07:06:14 +0000 UTC Normal PersistentVolumeClaim data-auto-provisioned-zookeeper-0 ProvisioningSucceeded Successfully provisioned volume pvc-6c7cf6e6-5296-4258-8ae2-c1374dc05df3 ebs.csi.aws.com_aws-ebs-csi-driver-controller-5664c47c89-rmlpw_d292fa2a-4ddd-48f3-92b9-8ab46e71f3c8 logger.go:42: 07:08:02 | streaming-with-autoprovisioning-autoscale | 2024-09-12 07:06:15 +0000 UTC Normal Pod auto-provisioned-zookeeper-0 Binding Scheduled Successfully assigned kuttl-test-bursting-ewe/auto-provisioned-zookeeper-0 to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:08:02 | streaming-with-autoprovisioning-autoscale | 2024-09-12 07:06:17 +0000 UTC Normal Pod auto-provisioned-zookeeper-0 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-6c7cf6e6-5296-4258-8ae2-c1374dc05df3" attachdetach-controller logger.go:42: 07:08:02 | streaming-with-autoprovisioning-autoscale | 2024-09-12 07:06:19 +0000 UTC Normal Pod auto-provisioned-zookeeper-0 AddedInterface Add eth0 [10.130.0.31/23] from ovn-kubernetes logger.go:42: 07:08:02 | streaming-with-autoprovisioning-autoscale | 2024-09-12 07:06:19 +0000 UTC Normal Pod auto-provisioned-zookeeper-0.spec.containers{zookeeper} Pulling Pulling image "registry.redhat.io/amq-streams/kafka-37-rhel9@sha256:42bf60ce31540dd61fab2c9886d791e41f063ea6f86628694b9e60e49bc8951b" kubelet logger.go:42: 07:08:02 | streaming-with-autoprovisioning-autoscale | 2024-09-12 07:06:31 +0000 UTC Normal Pod auto-provisioned-zookeeper-0.spec.containers{zookeeper} Pulled Successfully pulled image "registry.redhat.io/amq-streams/kafka-37-rhel9@sha256:42bf60ce31540dd61fab2c9886d791e41f063ea6f86628694b9e60e49bc8951b" in 11.978s (11.978s including waiting) kubelet logger.go:42: 07:08:02 | streaming-with-autoprovisioning-autoscale | 2024-09-12 07:06:31 +0000 UTC Normal Pod auto-provisioned-zookeeper-0.spec.containers{zookeeper} Created Created container zookeeper kubelet logger.go:42: 07:08:02 | streaming-with-autoprovisioning-autoscale | 2024-09-12 07:06:31 +0000 UTC Normal Pod auto-provisioned-zookeeper-0.spec.containers{zookeeper} Started Started container zookeeper kubelet logger.go:42: 07:08:02 | streaming-with-autoprovisioning-autoscale | 2024-09-12 07:06:51 +0000 UTC Normal PodDisruptionBudget.policy auto-provisioned-kafka NoPods No matching pods found controllermanager logger.go:42: 07:08:02 | streaming-with-autoprovisioning-autoscale | 2024-09-12 07:06:51 +0000 UTC Normal PersistentVolumeClaim data-0-auto-provisioned-kafka-0 WaitForFirstConsumer waiting for first consumer to be created before binding persistentvolume-controller logger.go:42: 07:08:02 | streaming-with-autoprovisioning-autoscale | 2024-09-12 07:06:51 +0000 UTC Normal PersistentVolumeClaim data-0-auto-provisioned-kafka-0 Provisioning External provisioner is provisioning volume for claim "kuttl-test-bursting-ewe/data-0-auto-provisioned-kafka-0" ebs.csi.aws.com_aws-ebs-csi-driver-controller-5664c47c89-rmlpw_d292fa2a-4ddd-48f3-92b9-8ab46e71f3c8 logger.go:42: 07:08:02 | streaming-with-autoprovisioning-autoscale | 2024-09-12 07:06:51 +0000 UTC Normal PersistentVolumeClaim data-0-auto-provisioned-kafka-0 ExternalProvisioning Waiting for a volume to be created either by the external provisioner 'ebs.csi.aws.com' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered. persistentvolume-controller logger.go:42: 07:08:02 | streaming-with-autoprovisioning-autoscale | 2024-09-12 07:06:56 +0000 UTC Normal Pod auto-provisioned-kafka-0 Binding Scheduled Successfully assigned kuttl-test-bursting-ewe/auto-provisioned-kafka-0 to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:08:02 | streaming-with-autoprovisioning-autoscale | 2024-09-12 07:06:56 +0000 UTC Normal PersistentVolumeClaim data-0-auto-provisioned-kafka-0 ProvisioningSucceeded Successfully provisioned volume pvc-cd896722-1b14-4225-b772-b8b7921bbe14 ebs.csi.aws.com_aws-ebs-csi-driver-controller-5664c47c89-rmlpw_d292fa2a-4ddd-48f3-92b9-8ab46e71f3c8 logger.go:42: 07:08:02 | streaming-with-autoprovisioning-autoscale | 2024-09-12 07:06:59 +0000 UTC Normal Pod auto-provisioned-kafka-0 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-cd896722-1b14-4225-b772-b8b7921bbe14" attachdetach-controller logger.go:42: 07:08:02 | streaming-with-autoprovisioning-autoscale | 2024-09-12 07:07:00 +0000 UTC Normal Pod auto-provisioned-kafka-0 AddedInterface Add eth0 [10.130.0.32/23] from ovn-kubernetes logger.go:42: 07:08:02 | streaming-with-autoprovisioning-autoscale | 2024-09-12 07:07:00 +0000 UTC Normal Pod auto-provisioned-kafka-0.spec.containers{kafka} Pulled Container image "registry.redhat.io/amq-streams/kafka-37-rhel9@sha256:42bf60ce31540dd61fab2c9886d791e41f063ea6f86628694b9e60e49bc8951b" already present on machine kubelet logger.go:42: 07:08:02 | streaming-with-autoprovisioning-autoscale | 2024-09-12 07:07:01 +0000 UTC Normal Pod auto-provisioned-kafka-0.spec.containers{kafka} Created Created container kafka kubelet logger.go:42: 07:08:02 | streaming-with-autoprovisioning-autoscale | 2024-09-12 07:07:01 +0000 UTC Normal Pod auto-provisioned-kafka-0.spec.containers{kafka} Started Started container kafka kubelet logger.go:42: 07:08:02 | streaming-with-autoprovisioning-autoscale | 2024-09-12 07:07:24 +0000 UTC Normal Pod auto-provisioned-entity-operator-6f8bc8cddb-hkfpc Binding Scheduled Successfully assigned kuttl-test-bursting-ewe/auto-provisioned-entity-operator-6f8bc8cddb-hkfpc to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:08:02 | streaming-with-autoprovisioning-autoscale | 2024-09-12 07:07:24 +0000 UTC Normal Pod auto-provisioned-entity-operator-6f8bc8cddb-hkfpc AddedInterface Add eth0 [10.130.0.33/23] from ovn-kubernetes logger.go:42: 07:08:02 | streaming-with-autoprovisioning-autoscale | 2024-09-12 07:07:24 +0000 UTC Normal Pod auto-provisioned-entity-operator-6f8bc8cddb-hkfpc.spec.containers{topic-operator} Pulled Container image "registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:95f5aa75cd1f7228e78fd4d88d786713fba4cf828dc22bc2dd1d0380909c1aef" already present on machine kubelet logger.go:42: 07:08:02 | streaming-with-autoprovisioning-autoscale | 2024-09-12 07:07:24 +0000 UTC Normal Pod auto-provisioned-entity-operator-6f8bc8cddb-hkfpc.spec.containers{topic-operator} Created Created container topic-operator kubelet logger.go:42: 07:08:02 | streaming-with-autoprovisioning-autoscale | 2024-09-12 07:07:24 +0000 UTC Normal Pod auto-provisioned-entity-operator-6f8bc8cddb-hkfpc.spec.containers{topic-operator} Started Started container topic-operator kubelet logger.go:42: 07:08:02 | streaming-with-autoprovisioning-autoscale | 2024-09-12 07:07:24 +0000 UTC Normal Pod auto-provisioned-entity-operator-6f8bc8cddb-hkfpc.spec.containers{user-operator} Pulled Container image "registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:95f5aa75cd1f7228e78fd4d88d786713fba4cf828dc22bc2dd1d0380909c1aef" already present on machine kubelet logger.go:42: 07:08:02 | streaming-with-autoprovisioning-autoscale | 2024-09-12 07:07:24 +0000 UTC Normal Pod auto-provisioned-entity-operator-6f8bc8cddb-hkfpc.spec.containers{user-operator} Created Created container user-operator kubelet logger.go:42: 07:08:02 | streaming-with-autoprovisioning-autoscale | 2024-09-12 07:07:24 +0000 UTC Normal Pod auto-provisioned-entity-operator-6f8bc8cddb-hkfpc.spec.containers{user-operator} Started Started container user-operator kubelet logger.go:42: 07:08:02 | streaming-with-autoprovisioning-autoscale | 2024-09-12 07:07:24 +0000 UTC Normal ReplicaSet.apps auto-provisioned-entity-operator-6f8bc8cddb SuccessfulCreate Created pod: auto-provisioned-entity-operator-6f8bc8cddb-hkfpc replicaset-controller logger.go:42: 07:08:02 | streaming-with-autoprovisioning-autoscale | 2024-09-12 07:07:24 +0000 UTC Normal Deployment.apps auto-provisioned-entity-operator ScalingReplicaSet Scaled up replica set auto-provisioned-entity-operator-6f8bc8cddb to 1 deployment-controller logger.go:42: 07:08:02 | streaming-with-autoprovisioning-autoscale | 2024-09-12 07:07:47 +0000 UTC Normal Pod auto-provisioned-collector-5f9c7f9995-b9xlx Binding Scheduled Successfully assigned kuttl-test-bursting-ewe/auto-provisioned-collector-5f9c7f9995-b9xlx to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:08:02 | streaming-with-autoprovisioning-autoscale | 2024-09-12 07:07:47 +0000 UTC Normal ReplicaSet.apps auto-provisioned-collector-5f9c7f9995 SuccessfulCreate Created pod: auto-provisioned-collector-5f9c7f9995-b9xlx replicaset-controller logger.go:42: 07:08:02 | streaming-with-autoprovisioning-autoscale | 2024-09-12 07:07:47 +0000 UTC Normal Deployment.apps auto-provisioned-collector ScalingReplicaSet Scaled up replica set auto-provisioned-collector-5f9c7f9995 to 1 deployment-controller logger.go:42: 07:08:02 | streaming-with-autoprovisioning-autoscale | 2024-09-12 07:07:47 +0000 UTC Normal Pod auto-provisioned-ingester-fc869fd74-t5g85 Binding Scheduled Successfully assigned kuttl-test-bursting-ewe/auto-provisioned-ingester-fc869fd74-t5g85 to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:08:02 | streaming-with-autoprovisioning-autoscale | 2024-09-12 07:07:47 +0000 UTC Normal Pod auto-provisioned-ingester-fc869fd74-t5g85 AddedInterface Add eth0 [10.130.0.35/23] from ovn-kubernetes logger.go:42: 07:08:02 | streaming-with-autoprovisioning-autoscale | 2024-09-12 07:07:47 +0000 UTC Normal Pod auto-provisioned-ingester-fc869fd74-t5g85.spec.containers{jaeger-ingester} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-ingester-rhel8@sha256:7f435686c86e24ca85e6ebfa2f3eb5a456a81a8c2d4df56e3f7b68fff46eaf85" kubelet logger.go:42: 07:08:02 | streaming-with-autoprovisioning-autoscale | 2024-09-12 07:07:47 +0000 UTC Normal ReplicaSet.apps auto-provisioned-ingester-fc869fd74 SuccessfulCreate Created pod: auto-provisioned-ingester-fc869fd74-t5g85 replicaset-controller logger.go:42: 07:08:02 | streaming-with-autoprovisioning-autoscale | 2024-09-12 07:07:47 +0000 UTC Normal Deployment.apps auto-provisioned-ingester ScalingReplicaSet Scaled up replica set auto-provisioned-ingester-fc869fd74 to 1 deployment-controller logger.go:42: 07:08:02 | streaming-with-autoprovisioning-autoscale | 2024-09-12 07:07:47 +0000 UTC Normal Pod auto-provisioned-query-8c5945856-pjn9b Binding Scheduled Successfully assigned kuttl-test-bursting-ewe/auto-provisioned-query-8c5945856-pjn9b to ip-10-0-52-42.us-west-2.compute.internal default-scheduler logger.go:42: 07:08:02 | streaming-with-autoprovisioning-autoscale | 2024-09-12 07:07:47 +0000 UTC Normal Pod auto-provisioned-query-8c5945856-pjn9b AddedInterface Add eth0 [10.128.0.43/23] from ovn-kubernetes logger.go:42: 07:08:02 | streaming-with-autoprovisioning-autoscale | 2024-09-12 07:07:47 +0000 UTC Normal Pod auto-provisioned-query-8c5945856-pjn9b.spec.containers{jaeger-query} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a9186dcd910256c0f464b0a3928844a01de166a10c186c97ef4581bf288c23cb" kubelet logger.go:42: 07:08:02 | streaming-with-autoprovisioning-autoscale | 2024-09-12 07:07:47 +0000 UTC Normal ReplicaSet.apps auto-provisioned-query-8c5945856 SuccessfulCreate Created pod: auto-provisioned-query-8c5945856-pjn9b replicaset-controller logger.go:42: 07:08:02 | streaming-with-autoprovisioning-autoscale | 2024-09-12 07:07:47 +0000 UTC Normal Deployment.apps auto-provisioned-query ScalingReplicaSet Scaled up replica set auto-provisioned-query-8c5945856 to 1 deployment-controller logger.go:42: 07:08:02 | streaming-with-autoprovisioning-autoscale | 2024-09-12 07:07:48 +0000 UTC Normal Pod auto-provisioned-collector-5f9c7f9995-b9xlx AddedInterface Add eth0 [10.130.0.34/23] from ovn-kubernetes logger.go:42: 07:08:02 | streaming-with-autoprovisioning-autoscale | 2024-09-12 07:07:48 +0000 UTC Normal Pod auto-provisioned-collector-5f9c7f9995-b9xlx.spec.containers{jaeger-collector} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:4bab63290ff52e4f6328009f9c8b2c4496b230e9f8a98eac01736a66a291ff6c" kubelet logger.go:42: 07:08:02 | streaming-with-autoprovisioning-autoscale | 2024-09-12 07:07:51 +0000 UTC Normal Pod auto-provisioned-collector-5f9c7f9995-b9xlx.spec.containers{jaeger-collector} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:4bab63290ff52e4f6328009f9c8b2c4496b230e9f8a98eac01736a66a291ff6c" in 2.795s (2.795s including waiting) kubelet logger.go:42: 07:08:02 | streaming-with-autoprovisioning-autoscale | 2024-09-12 07:07:51 +0000 UTC Normal Pod auto-provisioned-collector-5f9c7f9995-b9xlx.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:08:02 | streaming-with-autoprovisioning-autoscale | 2024-09-12 07:07:51 +0000 UTC Normal Pod auto-provisioned-collector-5f9c7f9995-b9xlx.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:08:02 | streaming-with-autoprovisioning-autoscale | 2024-09-12 07:07:52 +0000 UTC Normal Pod auto-provisioned-query-8c5945856-pjn9b.spec.containers{jaeger-query} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a9186dcd910256c0f464b0a3928844a01de166a10c186c97ef4581bf288c23cb" in 4.767s (4.767s including waiting) kubelet logger.go:42: 07:08:02 | streaming-with-autoprovisioning-autoscale | 2024-09-12 07:07:52 +0000 UTC Normal Pod auto-provisioned-query-8c5945856-pjn9b.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:08:02 | streaming-with-autoprovisioning-autoscale | 2024-09-12 07:07:52 +0000 UTC Normal Pod auto-provisioned-query-8c5945856-pjn9b.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:08:02 | streaming-with-autoprovisioning-autoscale | 2024-09-12 07:07:52 +0000 UTC Normal Pod auto-provisioned-query-8c5945856-pjn9b.spec.containers{oauth-proxy} Pulling Pulling image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" kubelet logger.go:42: 07:08:02 | streaming-with-autoprovisioning-autoscale | 2024-09-12 07:07:53 +0000 UTC Normal Pod auto-provisioned-ingester-fc869fd74-t5g85.spec.containers{jaeger-ingester} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-ingester-rhel8@sha256:7f435686c86e24ca85e6ebfa2f3eb5a456a81a8c2d4df56e3f7b68fff46eaf85" in 5.744s (5.744s including waiting) kubelet logger.go:42: 07:08:02 | streaming-with-autoprovisioning-autoscale | 2024-09-12 07:07:53 +0000 UTC Normal Pod auto-provisioned-ingester-fc869fd74-t5g85.spec.containers{jaeger-ingester} Created Created container jaeger-ingester kubelet logger.go:42: 07:08:02 | streaming-with-autoprovisioning-autoscale | 2024-09-12 07:07:53 +0000 UTC Normal Pod auto-provisioned-ingester-fc869fd74-t5g85.spec.containers{jaeger-ingester} Started Started container jaeger-ingester kubelet logger.go:42: 07:08:02 | streaming-with-autoprovisioning-autoscale | 2024-09-12 07:07:54 +0000 UTC Warning Pod auto-provisioned-ingester-fc869fd74-t5g85.spec.containers{jaeger-ingester} Unhealthy Readiness probe failed: HTTP probe failed with statuscode: 503 kubelet logger.go:42: 07:08:02 | streaming-with-autoprovisioning-autoscale | 2024-09-12 07:07:56 +0000 UTC Normal Pod auto-provisioned-query-8c5945856-pjn9b.spec.containers{oauth-proxy} Pulled Successfully pulled image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" in 3.431s (3.431s including waiting) kubelet logger.go:42: 07:08:02 | streaming-with-autoprovisioning-autoscale | 2024-09-12 07:07:56 +0000 UTC Normal Pod auto-provisioned-query-8c5945856-pjn9b.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:08:02 | streaming-with-autoprovisioning-autoscale | 2024-09-12 07:07:56 +0000 UTC Normal Pod auto-provisioned-query-8c5945856-pjn9b.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:08:02 | streaming-with-autoprovisioning-autoscale | 2024-09-12 07:07:56 +0000 UTC Normal Pod auto-provisioned-query-8c5945856-pjn9b.spec.containers{jaeger-agent} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" kubelet logger.go:42: 07:08:02 | streaming-with-autoprovisioning-autoscale | 2024-09-12 07:07:59 +0000 UTC Normal Pod auto-provisioned-query-8c5945856-pjn9b.spec.containers{jaeger-agent} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" in 2.907s (2.907s including waiting) kubelet logger.go:42: 07:08:02 | streaming-with-autoprovisioning-autoscale | 2024-09-12 07:07:59 +0000 UTC Normal Pod auto-provisioned-query-8c5945856-pjn9b.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:08:02 | streaming-with-autoprovisioning-autoscale | 2024-09-12 07:07:59 +0000 UTC Normal Pod auto-provisioned-query-8c5945856-pjn9b.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:08:02 | streaming-with-autoprovisioning-autoscale | Deleting namespace: kuttl-test-bursting-ewe === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (259.62s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (6.08s) --- PASS: kuttl/harness/streaming-with-autoprovisioning-autoscale (252.90s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name streaming --report --output /logs/artifacts/streaming.xml ./artifacts/kuttl-report.xml time="2024-09-12T07:08:15Z" level=debug msg="Setting a new name for the test suites" time="2024-09-12T07:08:15Z" level=debug msg="Removing 'artifacts' TestCase" time="2024-09-12T07:08:15Z" level=debug msg="normalizing test case names" time="2024-09-12T07:08:15Z" level=debug msg="streaming/artifacts -> streaming_artifacts" time="2024-09-12T07:08:15Z" level=debug msg="streaming/streaming-with-autoprovisioning-autoscale -> streaming_streaming_with_autoprovisioning_autoscale" +-----------------------------------------------------+--------+ | NAME | RESULT | +-----------------------------------------------------+--------+ | streaming_artifacts | passed | | streaming_streaming_with_autoprovisioning_autoscale | passed | +-----------------------------------------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh upgrade false true + '[' 3 -ne 3 ']' + test_suite_name=upgrade + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/upgrade.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-upgrade make[2]: Entering directory '/tmp/jaeger-tests' make docker JAEGER_VERSION=1.60.1 IMG="quay.io//jaeger-operator:next" make[3]: Entering directory '/tmp/jaeger-tests' [ ! -z "true" ] || docker build --build-arg=GOPROXY= --build-arg=VERSION="1.60.1" --build-arg=JAEGER_VERSION=1.60.1 --build-arg=TARGETARCH= --build-arg VERSION_DATE=2024-09-12T07:08:15Z --build-arg VERSION_PKG="github.com/jaegertracing/jaeger-operator/pkg/version" -t "quay.io//jaeger-operator:next" . make[3]: Leaving directory '/tmp/jaeger-tests' touch build-e2e-upgrade-image SKIP_ES_EXTERNAL=true IMG=quay.io//jaeger-operator:"1.60.1" JAEGER_OPERATOR_VERSION="1.60.1" JAEGER_VERSION="1.60.0" ./tests/e2e/upgrade/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.29 True False 19m Cluster version is 4.15.29' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.29 True False 19m Cluster version is 4.15.29' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/upgrade/render.sh ++ export SUITE_DIR=./tests/e2e/upgrade ++ SUITE_DIR=./tests/e2e/upgrade ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/upgrade ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + export JAEGER_NAME + '[' true = true ']' + skip_test upgrade 'Test not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=upgrade + message='Test not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/upgrade/_build + '[' _build '!=' _build ']' + rm -rf upgrade + warning 'upgrade: Test not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: upgrade: Test not supported in OpenShift\e[0m' WAR: upgrade: Test not supported in OpenShift + '[' true = true ']' + skip_test upgrade-from-latest-release 'Test not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=upgrade-from-latest-release + message='Test not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/upgrade/_build + '[' _build '!=' _build ']' + rm -rf upgrade-from-latest-release + warning 'upgrade-from-latest-release: Test not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: upgrade-from-latest-release: Test not supported in OpenShift\e[0m' WAR: upgrade-from-latest-release: Test not supported in OpenShift make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running upgrade E2E tests' Running upgrade E2E tests + cd tests/e2e/upgrade/_build + set +e + KUBECONFIG=/tmp/kubeconfig-2486720052 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-rosa-h-u0tm.f61o.s3.devshift.org:443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 1 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === CONT kuttl/harness/artifacts logger.go:42: 07:08:17 | artifacts | Creating namespace: kuttl-test-expert-ghoul logger.go:42: 07:08:17 | artifacts | artifacts events from ns kuttl-test-expert-ghoul: logger.go:42: 07:08:17 | artifacts | Deleting namespace: kuttl-test-expert-ghoul === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (6.71s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (6.11s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name upgrade --report --output /logs/artifacts/upgrade.xml ./artifacts/kuttl-report.xml time="2024-09-12T07:08:23Z" level=debug msg="Setting a new name for the test suites" time="2024-09-12T07:08:23Z" level=debug msg="Removing 'artifacts' TestCase" time="2024-09-12T07:08:23Z" level=debug msg="normalizing test case names" time="2024-09-12T07:08:23Z" level=debug msg="upgrade/artifacts -> upgrade_artifacts" +-------------------+--------+ | NAME | RESULT | +-------------------+--------+ | upgrade_artifacts | passed | +-------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh miscellaneous false true + '[' 3 -ne 3 ']' + test_suite_name=miscellaneous + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/miscellaneous.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-miscellaneous make[2]: Entering directory '/tmp/jaeger-tests' SKIP_ES_EXTERNAL=true ./tests/e2e/miscellaneous/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.29 True False 19m Cluster version is 4.15.29' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.29 True False 19m Cluster version is 4.15.29' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/miscellaneous/render.sh ++ export SUITE_DIR=./tests/e2e/miscellaneous ++ SUITE_DIR=./tests/e2e/miscellaneous ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/miscellaneous ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + start_test collector-autoscale + '[' 1 -ne 1 ']' + test_name=collector-autoscale + echo =========================================================================== =========================================================================== + info 'Rendering files for test collector-autoscale' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test collector-autoscale\e[0m' Rendering files for test collector-autoscale + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build + '[' _build '!=' _build ']' + mkdir -p collector-autoscale + cd collector-autoscale + jaeger_name=simple-prod + jaeger_deploy_mode=production + [[ true = true ]] + [[ true = true ]] + jaeger_deploy_mode=production_autoprovisioned + ELASTICSEARCH_NODECOUNT=1 + render_install_jaeger simple-prod production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=simple-prod + JAEGER_NAME=simple-prod + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.collector.resources.requests.memory="200m"' 01-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.collector.autoscale=true 01-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.collector.minReplicas=1 01-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.collector.maxReplicas=2 01-install.yaml + version_lt 1.30 1.23 ++ echo 1.30 1.23 ++ tr ' ' '\n' ++ sort -rV ++ head -n 1 + test 1.30 '!=' 1.30 + rm ./03-assert.yaml + generate_otlp_e2e_tests http + test_protocol=http + is_secured=false + '[' true = true ']' + is_secured=true + start_test collector-otlp-allinone-http + '[' 1 -ne 1 ']' + test_name=collector-otlp-allinone-http + echo =========================================================================== =========================================================================== + info 'Rendering files for test collector-otlp-allinone-http' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test collector-otlp-allinone-http\e[0m' Rendering files for test collector-otlp-allinone-http + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/collector-autoscale + '[' collector-autoscale '!=' _build ']' + cd .. + mkdir -p collector-otlp-allinone-http + cd collector-otlp-allinone-http + render_install_jaeger my-jaeger allInOne 00 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=allInOne + test_step=00 + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_otlp_smoke_test my-jaeger http true 01 + '[' 4 -ne 4 ']' + jaeger=my-jaeger + reporting_protocol=http + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template + '[' http = grpc ']' + reporting_port=:4318 + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 + OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + REPORTING_PROTOCOL=http + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset OTEL_EXPORTER_OTLP_ENDPOINT + start_test collector-otlp-production-http + '[' 1 -ne 1 ']' + test_name=collector-otlp-production-http + echo =========================================================================== =========================================================================== + info 'Rendering files for test collector-otlp-production-http' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test collector-otlp-production-http\e[0m' Rendering files for test collector-otlp-production-http + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/collector-otlp-allinone-http + '[' collector-otlp-allinone-http '!=' _build ']' + cd .. + mkdir -p collector-otlp-production-http + cd collector-otlp-production-http + jaeger_deploy_mode=production + [[ true = true ]] + [[ true = true ]] + jaeger_deploy_mode=production_autoprovisioned + render_install_jaeger my-jaeger production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + render_otlp_smoke_test my-jaeger http true 02 + '[' 4 -ne 4 ']' + jaeger=my-jaeger + reporting_protocol=http + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template + '[' http = grpc ']' + reporting_port=:4318 + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 + OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + REPORTING_PROTOCOL=http + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset OTEL_EXPORTER_OTLP_ENDPOINT + generate_otlp_e2e_tests grpc + test_protocol=grpc + is_secured=false + '[' true = true ']' + is_secured=true + start_test collector-otlp-allinone-grpc + '[' 1 -ne 1 ']' + test_name=collector-otlp-allinone-grpc + echo =========================================================================== =========================================================================== + info 'Rendering files for test collector-otlp-allinone-grpc' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test collector-otlp-allinone-grpc\e[0m' Rendering files for test collector-otlp-allinone-grpc + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/collector-otlp-production-http + '[' collector-otlp-production-http '!=' _build ']' + cd .. + mkdir -p collector-otlp-allinone-grpc + cd collector-otlp-allinone-grpc + render_install_jaeger my-jaeger allInOne 00 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=allInOne + test_step=00 + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_otlp_smoke_test my-jaeger grpc true 01 + '[' 4 -ne 4 ']' + jaeger=my-jaeger + reporting_protocol=grpc + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template + '[' grpc = grpc ']' + reporting_port=:4317 + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 + OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + REPORTING_PROTOCOL=grpc + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset OTEL_EXPORTER_OTLP_ENDPOINT + start_test collector-otlp-production-grpc + '[' 1 -ne 1 ']' + test_name=collector-otlp-production-grpc + echo =========================================================================== =========================================================================== + info 'Rendering files for test collector-otlp-production-grpc' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test collector-otlp-production-grpc\e[0m' Rendering files for test collector-otlp-production-grpc + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/collector-otlp-allinone-grpc + '[' collector-otlp-allinone-grpc '!=' _build ']' + cd .. + mkdir -p collector-otlp-production-grpc + cd collector-otlp-production-grpc + jaeger_deploy_mode=production + [[ true = true ]] + [[ true = true ]] + jaeger_deploy_mode=production_autoprovisioned + render_install_jaeger my-jaeger production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + render_otlp_smoke_test my-jaeger grpc true 02 + '[' 4 -ne 4 ']' + jaeger=my-jaeger + reporting_protocol=grpc + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template + '[' grpc = grpc ']' + reporting_port=:4317 + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 + OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + REPORTING_PROTOCOL=grpc + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset OTEL_EXPORTER_OTLP_ENDPOINT + '[' true = true ']' + skip_test istio 'Test not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=istio + message='Test not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/collector-otlp-production-grpc + '[' collector-otlp-production-grpc '!=' _build ']' + cd .. + rm -rf istio + warning 'istio: Test not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: istio: Test not supported in OpenShift\e[0m' WAR: istio: Test not supported in OpenShift + '[' true = true ']' + skip_test outside-cluster 'Test not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=outside-cluster + message='Test not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build + '[' _build '!=' _build ']' + rm -rf outside-cluster + warning 'outside-cluster: Test not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: outside-cluster: Test not supported in OpenShift\e[0m' WAR: outside-cluster: Test not supported in OpenShift + start_test set-custom-img + '[' 1 -ne 1 ']' + test_name=set-custom-img + echo =========================================================================== =========================================================================== + info 'Rendering files for test set-custom-img' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test set-custom-img\e[0m' Rendering files for test set-custom-img + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build + '[' _build '!=' _build ']' + mkdir -p set-custom-img + cd set-custom-img + jaeger_name=my-jaeger + jaeger_deploy_mode=production + [[ true = true ]] + [[ true = true ]] + jaeger_deploy_mode=production_autoprovisioned + render_install_jaeger my-jaeger production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + cp ./01-install.yaml ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.collector.image="test"' ./02-install.yaml + '[' true = true ']' + skip_test non-cluster-wide 'Test not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=non-cluster-wide + message='Test not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/set-custom-img + '[' set-custom-img '!=' _build ']' + cd .. + rm -rf non-cluster-wide + warning 'non-cluster-wide: Test not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: non-cluster-wide: Test not supported in OpenShift\e[0m' WAR: non-cluster-wide: Test not supported in OpenShift make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running miscellaneous E2E tests' Running miscellaneous E2E tests + cd tests/e2e/miscellaneous/_build + set +e + KUBECONFIG=/tmp/kubeconfig-2486720052 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-rosa-h-u0tm.f61o.s3.devshift.org:443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 8 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/cassandra-spark === PAUSE kuttl/harness/cassandra-spark === RUN kuttl/harness/collector-autoscale === PAUSE kuttl/harness/collector-autoscale === RUN kuttl/harness/collector-otlp-allinone-grpc === PAUSE kuttl/harness/collector-otlp-allinone-grpc === RUN kuttl/harness/collector-otlp-allinone-http === PAUSE kuttl/harness/collector-otlp-allinone-http === RUN kuttl/harness/collector-otlp-production-grpc === PAUSE kuttl/harness/collector-otlp-production-grpc === RUN kuttl/harness/collector-otlp-production-http === PAUSE kuttl/harness/collector-otlp-production-http === RUN kuttl/harness/set-custom-img === PAUSE kuttl/harness/set-custom-img === CONT kuttl/harness/artifacts logger.go:42: 07:08:35 | artifacts | Creating namespace: kuttl-test-major-perch logger.go:42: 07:08:35 | artifacts | artifacts events from ns kuttl-test-major-perch: logger.go:42: 07:08:35 | artifacts | Deleting namespace: kuttl-test-major-perch === CONT kuttl/harness/collector-otlp-allinone-http logger.go:42: 07:08:41 | collector-otlp-allinone-http | Creating namespace: kuttl-test-needed-finch logger.go:42: 07:08:41 | collector-otlp-allinone-http/0-install | starting test step 0-install logger.go:42: 07:08:41 | collector-otlp-allinone-http/0-install | Jaeger:kuttl-test-needed-finch/my-jaeger created logger.go:42: 07:08:47 | collector-otlp-allinone-http/0-install | test step completed 0-install logger.go:42: 07:08:47 | collector-otlp-allinone-http/1-smoke-test | starting test step 1-smoke-test logger.go:42: 07:08:47 | collector-otlp-allinone-http/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:08:49 | collector-otlp-allinone-http/1-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:08:59 | collector-otlp-allinone-http/1-smoke-test | running command: [sh -c REPORTING_PROTOCOL=http ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/otlp-smoke-test.yaml.template -o otlp-smoke-test-job.yaml] logger.go:42: 07:09:00 | collector-otlp-allinone-http/1-smoke-test | running command: [sh -c kubectl create -f otlp-smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:09:00 | collector-otlp-allinone-http/1-smoke-test | job.batch/report-span created logger.go:42: 07:09:00 | collector-otlp-allinone-http/1-smoke-test | job.batch/check-span created logger.go:42: 07:09:13 | collector-otlp-allinone-http/1-smoke-test | test step completed 1-smoke-test logger.go:42: 07:09:14 | collector-otlp-allinone-http | collector-otlp-allinone-http events from ns kuttl-test-needed-finch: logger.go:42: 07:09:14 | collector-otlp-allinone-http | 2024-09-12 07:08:42 +0000 UTC Normal Pod my-jaeger-89f9868bb-9fhsw Binding Scheduled Successfully assigned kuttl-test-needed-finch/my-jaeger-89f9868bb-9fhsw to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:09:14 | collector-otlp-allinone-http | 2024-09-12 07:08:42 +0000 UTC Normal Pod my-jaeger-89f9868bb-9fhsw AddedInterface Add eth0 [10.130.0.36/23] from ovn-kubernetes logger.go:42: 07:09:14 | collector-otlp-allinone-http | 2024-09-12 07:08:42 +0000 UTC Normal Pod my-jaeger-89f9868bb-9fhsw.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" already present on machine kubelet logger.go:42: 07:09:14 | collector-otlp-allinone-http | 2024-09-12 07:08:42 +0000 UTC Normal ReplicaSet.apps my-jaeger-89f9868bb SuccessfulCreate Created pod: my-jaeger-89f9868bb-9fhsw replicaset-controller logger.go:42: 07:09:14 | collector-otlp-allinone-http | 2024-09-12 07:08:42 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-89f9868bb to 1 deployment-controller logger.go:42: 07:09:14 | collector-otlp-allinone-http | 2024-09-12 07:08:43 +0000 UTC Normal Pod my-jaeger-89f9868bb-9fhsw.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:09:14 | collector-otlp-allinone-http | 2024-09-12 07:08:43 +0000 UTC Normal Pod my-jaeger-89f9868bb-9fhsw.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:09:14 | collector-otlp-allinone-http | 2024-09-12 07:08:43 +0000 UTC Normal Pod my-jaeger-89f9868bb-9fhsw.spec.containers{oauth-proxy} Pulling Pulling image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" kubelet logger.go:42: 07:09:14 | collector-otlp-allinone-http | 2024-09-12 07:08:46 +0000 UTC Normal Pod my-jaeger-89f9868bb-9fhsw.spec.containers{oauth-proxy} Pulled Successfully pulled image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" in 3.062s (3.062s including waiting) kubelet logger.go:42: 07:09:14 | collector-otlp-allinone-http | 2024-09-12 07:08:46 +0000 UTC Normal Pod my-jaeger-89f9868bb-9fhsw.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:09:14 | collector-otlp-allinone-http | 2024-09-12 07:08:46 +0000 UTC Normal Pod my-jaeger-89f9868bb-9fhsw.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:09:14 | collector-otlp-allinone-http | 2024-09-12 07:08:52 +0000 UTC Normal Pod my-jaeger-89f9868bb-9fhsw.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:09:14 | collector-otlp-allinone-http | 2024-09-12 07:08:52 +0000 UTC Normal Pod my-jaeger-89f9868bb-9fhsw.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:09:14 | collector-otlp-allinone-http | 2024-09-12 07:08:52 +0000 UTC Normal ReplicaSet.apps my-jaeger-89f9868bb SuccessfulDelete Deleted pod: my-jaeger-89f9868bb-9fhsw replicaset-controller logger.go:42: 07:09:14 | collector-otlp-allinone-http | 2024-09-12 07:08:52 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled down replica set my-jaeger-89f9868bb to 0 from 1 deployment-controller logger.go:42: 07:09:14 | collector-otlp-allinone-http | 2024-09-12 07:08:53 +0000 UTC Normal ReplicaSet.apps my-jaeger-5b8444b5bf SuccessfulCreate Created pod: my-jaeger-5b8444b5bf-r2nsj replicaset-controller logger.go:42: 07:09:14 | collector-otlp-allinone-http | 2024-09-12 07:08:53 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-5b8444b5bf to 1 deployment-controller logger.go:42: 07:09:14 | collector-otlp-allinone-http | 2024-09-12 07:08:54 +0000 UTC Normal Pod my-jaeger-5b8444b5bf-r2nsj Binding Scheduled Successfully assigned kuttl-test-needed-finch/my-jaeger-5b8444b5bf-r2nsj to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:09:14 | collector-otlp-allinone-http | 2024-09-12 07:08:54 +0000 UTC Normal Pod my-jaeger-5b8444b5bf-r2nsj AddedInterface Add eth0 [10.130.0.37/23] from ovn-kubernetes logger.go:42: 07:09:14 | collector-otlp-allinone-http | 2024-09-12 07:08:54 +0000 UTC Normal Pod my-jaeger-5b8444b5bf-r2nsj.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" already present on machine kubelet logger.go:42: 07:09:14 | collector-otlp-allinone-http | 2024-09-12 07:08:54 +0000 UTC Normal Pod my-jaeger-5b8444b5bf-r2nsj.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:09:14 | collector-otlp-allinone-http | 2024-09-12 07:08:54 +0000 UTC Normal Pod my-jaeger-5b8444b5bf-r2nsj.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:09:14 | collector-otlp-allinone-http | 2024-09-12 07:08:54 +0000 UTC Normal Pod my-jaeger-5b8444b5bf-r2nsj.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:09:14 | collector-otlp-allinone-http | 2024-09-12 07:08:54 +0000 UTC Normal Pod my-jaeger-5b8444b5bf-r2nsj.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:09:14 | collector-otlp-allinone-http | 2024-09-12 07:08:54 +0000 UTC Normal Pod my-jaeger-5b8444b5bf-r2nsj.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:09:14 | collector-otlp-allinone-http | 2024-09-12 07:09:00 +0000 UTC Normal Pod check-span-ddl67 Binding Scheduled Successfully assigned kuttl-test-needed-finch/check-span-ddl67 to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:09:14 | collector-otlp-allinone-http | 2024-09-12 07:09:00 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-ddl67 job-controller logger.go:42: 07:09:14 | collector-otlp-allinone-http | 2024-09-12 07:09:00 +0000 UTC Normal Pod report-span-8nvrk Binding Scheduled Successfully assigned kuttl-test-needed-finch/report-span-8nvrk to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:09:14 | collector-otlp-allinone-http | 2024-09-12 07:09:00 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-8nvrk job-controller logger.go:42: 07:09:14 | collector-otlp-allinone-http | 2024-09-12 07:09:01 +0000 UTC Normal Pod check-span-ddl67 AddedInterface Add eth0 [10.130.0.39/23] from ovn-kubernetes logger.go:42: 07:09:14 | collector-otlp-allinone-http | 2024-09-12 07:09:01 +0000 UTC Normal Pod check-span-ddl67.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:09:14 | collector-otlp-allinone-http | 2024-09-12 07:09:01 +0000 UTC Normal Pod report-span-8nvrk AddedInterface Add eth0 [10.130.0.38/23] from ovn-kubernetes logger.go:42: 07:09:14 | collector-otlp-allinone-http | 2024-09-12 07:09:01 +0000 UTC Normal Pod report-span-8nvrk.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:09:14 | collector-otlp-allinone-http | 2024-09-12 07:09:02 +0000 UTC Normal Pod check-span-ddl67.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 844ms (844ms including waiting) kubelet logger.go:42: 07:09:14 | collector-otlp-allinone-http | 2024-09-12 07:09:02 +0000 UTC Normal Pod check-span-ddl67.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:09:14 | collector-otlp-allinone-http | 2024-09-12 07:09:02 +0000 UTC Normal Pod check-span-ddl67.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:09:14 | collector-otlp-allinone-http | 2024-09-12 07:09:02 +0000 UTC Normal Pod report-span-8nvrk.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 906ms (906ms including waiting) kubelet logger.go:42: 07:09:14 | collector-otlp-allinone-http | 2024-09-12 07:09:02 +0000 UTC Normal Pod report-span-8nvrk.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:09:14 | collector-otlp-allinone-http | 2024-09-12 07:09:02 +0000 UTC Normal Pod report-span-8nvrk.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:09:14 | collector-otlp-allinone-http | 2024-09-12 07:09:13 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:09:14 | collector-otlp-allinone-http | Deleting namespace: kuttl-test-needed-finch === CONT kuttl/harness/set-custom-img logger.go:42: 07:09:26 | set-custom-img | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:09:26 | set-custom-img | Ignoring check-collector-img.sh as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:09:26 | set-custom-img | Creating namespace: kuttl-test-apparent-pony logger.go:42: 07:09:26 | set-custom-img/1-install | starting test step 1-install logger.go:42: 07:09:26 | set-custom-img/1-install | Jaeger:kuttl-test-apparent-pony/my-jaeger created logger.go:42: 07:10:33 | set-custom-img/1-install | test step completed 1-install logger.go:42: 07:10:33 | set-custom-img/2-install | starting test step 2-install logger.go:42: 07:10:34 | set-custom-img/2-install | Jaeger:kuttl-test-apparent-pony/my-jaeger updated logger.go:42: 07:10:34 | set-custom-img/2-install | test step completed 2-install logger.go:42: 07:10:34 | set-custom-img/3-check-image | starting test step 3-check-image logger.go:42: 07:10:34 | set-custom-img/3-check-image | running command: [sh -c ./check-collector-img.sh] logger.go:42: 07:10:34 | set-custom-img/3-check-image | Collector image missmatch. Expected: test. Has: registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:4bab63290ff52e4f6328009f9c8b2c4496b230e9f8a98eac01736a66a291ff6c logger.go:42: 07:10:39 | set-custom-img/3-check-image | Collector image asserted properly! logger.go:42: 07:10:40 | set-custom-img/3-check-image | test step completed 3-check-image logger.go:42: 07:10:40 | set-custom-img | set-custom-img events from ns kuttl-test-apparent-pony: logger.go:42: 07:10:40 | set-custom-img | 2024-09-12 07:09:32 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestapparentponymyjaeger-1-c59f7b85 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestapparentponymyjaeger-1-c59f7b85nlhvd replicaset-controller logger.go:42: 07:10:40 | set-custom-img | 2024-09-12 07:09:32 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestapparentponymyjaeger-1-c59f7b85nlhvd Binding Scheduled Successfully assigned kuttl-test-apparent-pony/elasticsearch-cdm-kuttltestapparentponymyjaeger-1-c59f7b85nlhvd to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:10:40 | set-custom-img | 2024-09-12 07:09:32 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestapparentponymyjaeger-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestapparentponymyjaeger-1-c59f7b85 to 1 deployment-controller logger.go:42: 07:10:40 | set-custom-img | 2024-09-12 07:09:33 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestapparentponymyjaeger-1-c59f7b85nlhvd AddedInterface Add eth0 [10.130.0.40/23] from ovn-kubernetes logger.go:42: 07:10:40 | set-custom-img | 2024-09-12 07:09:33 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestapparentponymyjaeger-1-c59f7b85nlhvd.spec.containers{elasticsearch} Pulling Pulling image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:bc97b8e13087050dfb1a75b02c5b780cbb8fb12a849a655169d072ba8bbf42b4" kubelet logger.go:42: 07:10:40 | set-custom-img | 2024-09-12 07:09:33 +0000 UTC Normal Pod my-jaeger-collector-7dcc79cbb8-9d2x4 Binding Scheduled Successfully assigned kuttl-test-apparent-pony/my-jaeger-collector-7dcc79cbb8-9d2x4 to ip-10-0-54-110.us-west-2.compute.internal default-scheduler logger.go:42: 07:10:40 | set-custom-img | 2024-09-12 07:09:33 +0000 UTC Warning Pod my-jaeger-collector-7dcc79cbb8-9d2x4 FailedMount MountVolume.SetUp failed for volume "my-jaeger-collector-tls-config-volume" : secret "my-jaeger-collector-headless-tls" not found kubelet logger.go:42: 07:10:40 | set-custom-img | 2024-09-12 07:09:33 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-7dcc79cbb8 SuccessfulCreate Created pod: my-jaeger-collector-7dcc79cbb8-9d2x4 replicaset-controller logger.go:42: 07:10:40 | set-custom-img | 2024-09-12 07:09:33 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-7dcc79cbb8 to 1 deployment-controller logger.go:42: 07:10:40 | set-custom-img | 2024-09-12 07:09:33 +0000 UTC Normal Pod my-jaeger-query-557db9f866-5kqkx Binding Scheduled Successfully assigned kuttl-test-apparent-pony/my-jaeger-query-557db9f866-5kqkx to ip-10-0-52-42.us-west-2.compute.internal default-scheduler logger.go:42: 07:10:40 | set-custom-img | 2024-09-12 07:09:33 +0000 UTC Normal Pod my-jaeger-query-557db9f866-5kqkx AddedInterface Add eth0 [10.128.0.44/23] from ovn-kubernetes logger.go:42: 07:10:40 | set-custom-img | 2024-09-12 07:09:33 +0000 UTC Normal Pod my-jaeger-query-557db9f866-5kqkx.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a9186dcd910256c0f464b0a3928844a01de166a10c186c97ef4581bf288c23cb" already present on machine kubelet logger.go:42: 07:10:40 | set-custom-img | 2024-09-12 07:09:33 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-557db9f866 SuccessfulCreate Created pod: my-jaeger-query-557db9f866-5kqkx replicaset-controller logger.go:42: 07:10:40 | set-custom-img | 2024-09-12 07:09:33 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-557db9f866 to 1 deployment-controller logger.go:42: 07:10:40 | set-custom-img | 2024-09-12 07:09:34 +0000 UTC Normal Pod my-jaeger-collector-7dcc79cbb8-9d2x4 AddedInterface Add eth0 [10.129.0.32/23] from ovn-kubernetes logger.go:42: 07:10:40 | set-custom-img | 2024-09-12 07:09:34 +0000 UTC Normal Pod my-jaeger-collector-7dcc79cbb8-9d2x4.spec.containers{jaeger-collector} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:4bab63290ff52e4f6328009f9c8b2c4496b230e9f8a98eac01736a66a291ff6c" kubelet logger.go:42: 07:10:40 | set-custom-img | 2024-09-12 07:09:34 +0000 UTC Normal Pod my-jaeger-query-557db9f866-5kqkx.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:10:40 | set-custom-img | 2024-09-12 07:09:34 +0000 UTC Normal Pod my-jaeger-query-557db9f866-5kqkx.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:10:40 | set-custom-img | 2024-09-12 07:09:34 +0000 UTC Normal Pod my-jaeger-query-557db9f866-5kqkx.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:10:40 | set-custom-img | 2024-09-12 07:09:34 +0000 UTC Normal Pod my-jaeger-query-557db9f866-5kqkx.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:10:40 | set-custom-img | 2024-09-12 07:09:34 +0000 UTC Normal Pod my-jaeger-query-557db9f866-5kqkx.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:10:40 | set-custom-img | 2024-09-12 07:09:34 +0000 UTC Normal Pod my-jaeger-query-557db9f866-5kqkx.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" already present on machine kubelet logger.go:42: 07:10:40 | set-custom-img | 2024-09-12 07:09:34 +0000 UTC Normal Pod my-jaeger-query-557db9f866-5kqkx.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:10:40 | set-custom-img | 2024-09-12 07:09:34 +0000 UTC Normal Pod my-jaeger-query-557db9f866-5kqkx.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:10:40 | set-custom-img | 2024-09-12 07:09:35 +0000 UTC Warning Pod my-jaeger-query-557db9f866-5kqkx.spec.containers{jaeger-query} Unhealthy Readiness probe failed: HTTP probe failed with statuscode: 503 kubelet logger.go:42: 07:10:40 | set-custom-img | 2024-09-12 07:09:37 +0000 UTC Normal Pod my-jaeger-collector-7dcc79cbb8-9d2x4.spec.containers{jaeger-collector} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:4bab63290ff52e4f6328009f9c8b2c4496b230e9f8a98eac01736a66a291ff6c" in 3.255s (3.255s including waiting) kubelet logger.go:42: 07:10:40 | set-custom-img | 2024-09-12 07:09:37 +0000 UTC Normal Pod my-jaeger-collector-7dcc79cbb8-9d2x4.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:10:40 | set-custom-img | 2024-09-12 07:09:37 +0000 UTC Normal Pod my-jaeger-collector-7dcc79cbb8-9d2x4.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:10:40 | set-custom-img | 2024-09-12 07:09:38 +0000 UTC Warning Pod my-jaeger-collector-7dcc79cbb8-9d2x4.spec.containers{jaeger-collector} Unhealthy Readiness probe failed: HTTP probe failed with statuscode: 503 kubelet logger.go:42: 07:10:40 | set-custom-img | 2024-09-12 07:09:43 +0000 UTC Normal Pod my-jaeger-collector-7dcc79cbb8-9d2x4.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:4bab63290ff52e4f6328009f9c8b2c4496b230e9f8a98eac01736a66a291ff6c" already present on machine kubelet logger.go:42: 07:10:40 | set-custom-img | 2024-09-12 07:09:44 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestapparentponymyjaeger-1-c59f7b85nlhvd.spec.containers{elasticsearch} Pulled Successfully pulled image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:bc97b8e13087050dfb1a75b02c5b780cbb8fb12a849a655169d072ba8bbf42b4" in 11.111s (11.111s including waiting) kubelet logger.go:42: 07:10:40 | set-custom-img | 2024-09-12 07:09:44 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestapparentponymyjaeger-1-c59f7b85nlhvd.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:10:40 | set-custom-img | 2024-09-12 07:09:44 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestapparentponymyjaeger-1-c59f7b85nlhvd.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:10:40 | set-custom-img | 2024-09-12 07:09:44 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestapparentponymyjaeger-1-c59f7b85nlhvd.spec.containers{proxy} Pulling Pulling image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:1d1d7d47b616995d18692f3a2d6232d7e1a0f41bc1503d87100f31939e080a78" kubelet logger.go:42: 07:10:40 | set-custom-img | 2024-09-12 07:09:46 +0000 UTC Warning Pod my-jaeger-query-557db9f866-5kqkx.spec.containers{jaeger-query} BackOff Back-off restarting failed container jaeger-query in pod my-jaeger-query-557db9f866-5kqkx_kuttl-test-apparent-pony(e911430f-a19c-41d3-937e-385f0bd61082) kubelet logger.go:42: 07:10:40 | set-custom-img | 2024-09-12 07:09:48 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:10:40 | set-custom-img | 2024-09-12 07:09:48 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:10:40 | set-custom-img | 2024-09-12 07:09:48 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:10:40 | set-custom-img | 2024-09-12 07:09:49 +0000 UTC Warning Pod my-jaeger-collector-7dcc79cbb8-9d2x4.spec.containers{jaeger-collector} Unhealthy Liveness probe failed: Get "http://10.129.0.32:14269/": dial tcp 10.129.0.32:14269: connect: connection refused kubelet logger.go:42: 07:10:40 | set-custom-img | 2024-09-12 07:09:49 +0000 UTC Warning Pod my-jaeger-collector-7dcc79cbb8-9d2x4.spec.containers{jaeger-collector} BackOff Back-off restarting failed container jaeger-collector in pod my-jaeger-collector-7dcc79cbb8-9d2x4_kuttl-test-apparent-pony(5fd81a15-caf5-4192-82eb-468292b4cf6a) kubelet logger.go:42: 07:10:40 | set-custom-img | 2024-09-12 07:09:50 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestapparentponymyjaeger-1-c59f7b85nlhvd.spec.containers{proxy} Pulled Successfully pulled image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:1d1d7d47b616995d18692f3a2d6232d7e1a0f41bc1503d87100f31939e080a78" in 5.58s (5.58s including waiting) kubelet logger.go:42: 07:10:40 | set-custom-img | 2024-09-12 07:09:50 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestapparentponymyjaeger-1-c59f7b85nlhvd.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:10:40 | set-custom-img | 2024-09-12 07:09:50 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestapparentponymyjaeger-1-c59f7b85nlhvd.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:10:40 | set-custom-img | 2024-09-12 07:09:58 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestapparentponymyjaeger-1-c59f7b85nlhvd.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:10:40 | set-custom-img | 2024-09-12 07:10:35 +0000 UTC Normal Pod my-jaeger-collector-7dcc79cbb8-9d2x4.spec.containers{jaeger-collector} Killing Stopping container jaeger-collector kubelet logger.go:42: 07:10:40 | set-custom-img | 2024-09-12 07:10:35 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-7dcc79cbb8 SuccessfulDelete Deleted pod: my-jaeger-collector-7dcc79cbb8-9d2x4 replicaset-controller logger.go:42: 07:10:40 | set-custom-img | 2024-09-12 07:10:35 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled down replica set my-jaeger-collector-7dcc79cbb8 to 0 from 1 deployment-controller logger.go:42: 07:10:40 | set-custom-img | 2024-09-12 07:10:36 +0000 UTC Normal Pod my-jaeger-collector-cff86d95-swlmd Binding Scheduled Successfully assigned kuttl-test-apparent-pony/my-jaeger-collector-cff86d95-swlmd to ip-10-0-54-110.us-west-2.compute.internal default-scheduler logger.go:42: 07:10:40 | set-custom-img | 2024-09-12 07:10:36 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-cff86d95 SuccessfulCreate Created pod: my-jaeger-collector-cff86d95-swlmd replicaset-controller logger.go:42: 07:10:40 | set-custom-img | 2024-09-12 07:10:36 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-cff86d95 to 1 deployment-controller logger.go:42: 07:10:40 | set-custom-img | 2024-09-12 07:10:37 +0000 UTC Normal Pod my-jaeger-collector-cff86d95-swlmd AddedInterface Add eth0 [10.129.0.34/23] from ovn-kubernetes logger.go:42: 07:10:40 | set-custom-img | 2024-09-12 07:10:37 +0000 UTC Normal Pod my-jaeger-collector-cff86d95-swlmd.spec.containers{jaeger-collector} Pulling Pulling image "test" kubelet logger.go:42: 07:10:40 | set-custom-img | 2024-09-12 07:10:38 +0000 UTC Warning Pod my-jaeger-collector-cff86d95-swlmd.spec.containers{jaeger-collector} Failed Failed to pull image "test": reading manifest latest in docker.io/library/test: requested access to the resource is denied kubelet logger.go:42: 07:10:40 | set-custom-img | 2024-09-12 07:10:38 +0000 UTC Warning Pod my-jaeger-collector-cff86d95-swlmd.spec.containers{jaeger-collector} Failed Error: ErrImagePull kubelet logger.go:42: 07:10:40 | set-custom-img | 2024-09-12 07:10:38 +0000 UTC Normal Pod my-jaeger-collector-cff86d95-swlmd.spec.containers{jaeger-collector} BackOff Back-off pulling image "test" kubelet logger.go:42: 07:10:40 | set-custom-img | 2024-09-12 07:10:38 +0000 UTC Warning Pod my-jaeger-collector-cff86d95-swlmd.spec.containers{jaeger-collector} Failed Error: ImagePullBackOff kubelet logger.go:42: 07:10:40 | set-custom-img | Deleting namespace: kuttl-test-apparent-pony === CONT kuttl/harness/collector-otlp-production-http logger.go:42: 07:10:46 | collector-otlp-production-http | Creating namespace: kuttl-test-cheerful-grouse logger.go:42: 07:10:46 | collector-otlp-production-http/1-install | starting test step 1-install logger.go:42: 07:10:46 | collector-otlp-production-http/1-install | Jaeger:kuttl-test-cheerful-grouse/my-jaeger created logger.go:42: 07:11:21 | collector-otlp-production-http/1-install | test step completed 1-install logger.go:42: 07:11:21 | collector-otlp-production-http/2-smoke-test | starting test step 2-smoke-test logger.go:42: 07:11:21 | collector-otlp-production-http/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:11:23 | collector-otlp-production-http/2-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:11:30 | collector-otlp-production-http/2-smoke-test | running command: [sh -c REPORTING_PROTOCOL=http ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/otlp-smoke-test.yaml.template -o otlp-smoke-test-job.yaml] logger.go:42: 07:11:30 | collector-otlp-production-http/2-smoke-test | running command: [sh -c kubectl create -f otlp-smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:11:31 | collector-otlp-production-http/2-smoke-test | job.batch/report-span created logger.go:42: 07:11:31 | collector-otlp-production-http/2-smoke-test | job.batch/check-span created logger.go:42: 07:11:46 | collector-otlp-production-http/2-smoke-test | test step completed 2-smoke-test logger.go:42: 07:11:46 | collector-otlp-production-http | collector-otlp-production-http events from ns kuttl-test-cheerful-grouse: logger.go:42: 07:11:46 | collector-otlp-production-http | 2024-09-12 07:10:51 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestcheerfulgrousemyjaeger-1-65467fd5c SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestcheerfulgrousemyjaeger-1-65467fvwd2d replicaset-controller logger.go:42: 07:11:46 | collector-otlp-production-http | 2024-09-12 07:10:51 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestcheerfulgrousemyjaeger-1-65467fvwd2d Binding Scheduled Successfully assigned kuttl-test-cheerful-grouse/elasticsearch-cdm-kuttltestcheerfulgrousemyjaeger-1-65467fvwd2d to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:11:46 | collector-otlp-production-http | 2024-09-12 07:10:51 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestcheerfulgrousemyjaeger-1-65467fvwd2d AddedInterface Add eth0 [10.130.0.41/23] from ovn-kubernetes logger.go:42: 07:11:46 | collector-otlp-production-http | 2024-09-12 07:10:51 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestcheerfulgrousemyjaeger-1-65467fvwd2d.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:bc97b8e13087050dfb1a75b02c5b780cbb8fb12a849a655169d072ba8bbf42b4" already present on machine kubelet logger.go:42: 07:11:46 | collector-otlp-production-http | 2024-09-12 07:10:51 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestcheerfulgrousemyjaeger-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestcheerfulgrousemyjaeger-1-65467fd5c to 1 deployment-controller logger.go:42: 07:11:46 | collector-otlp-production-http | 2024-09-12 07:10:52 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestcheerfulgrousemyjaeger-1-65467fvwd2d.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:11:46 | collector-otlp-production-http | 2024-09-12 07:10:52 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestcheerfulgrousemyjaeger-1-65467fvwd2d.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:11:46 | collector-otlp-production-http | 2024-09-12 07:10:52 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestcheerfulgrousemyjaeger-1-65467fvwd2d.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:1d1d7d47b616995d18692f3a2d6232d7e1a0f41bc1503d87100f31939e080a78" already present on machine kubelet logger.go:42: 07:11:46 | collector-otlp-production-http | 2024-09-12 07:10:52 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestcheerfulgrousemyjaeger-1-65467fvwd2d.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:11:46 | collector-otlp-production-http | 2024-09-12 07:10:52 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestcheerfulgrousemyjaeger-1-65467fvwd2d.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:11:46 | collector-otlp-production-http | 2024-09-12 07:11:07 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestcheerfulgrousemyjaeger-1-65467fvwd2d.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:11:46 | collector-otlp-production-http | 2024-09-12 07:11:18 +0000 UTC Normal Pod my-jaeger-collector-57cbd56974-t9vl4 Binding Scheduled Successfully assigned kuttl-test-cheerful-grouse/my-jaeger-collector-57cbd56974-t9vl4 to ip-10-0-54-110.us-west-2.compute.internal default-scheduler logger.go:42: 07:11:46 | collector-otlp-production-http | 2024-09-12 07:11:18 +0000 UTC Normal Pod my-jaeger-collector-57cbd56974-t9vl4 AddedInterface Add eth0 [10.129.0.35/23] from ovn-kubernetes logger.go:42: 07:11:46 | collector-otlp-production-http | 2024-09-12 07:11:18 +0000 UTC Normal Pod my-jaeger-collector-57cbd56974-t9vl4.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:4bab63290ff52e4f6328009f9c8b2c4496b230e9f8a98eac01736a66a291ff6c" already present on machine kubelet logger.go:42: 07:11:46 | collector-otlp-production-http | 2024-09-12 07:11:18 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-57cbd56974 SuccessfulCreate Created pod: my-jaeger-collector-57cbd56974-t9vl4 replicaset-controller logger.go:42: 07:11:46 | collector-otlp-production-http | 2024-09-12 07:11:18 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-57cbd56974 to 1 deployment-controller logger.go:42: 07:11:46 | collector-otlp-production-http | 2024-09-12 07:11:18 +0000 UTC Normal Pod my-jaeger-query-59cd5ddd7c-bfh56 Binding Scheduled Successfully assigned kuttl-test-cheerful-grouse/my-jaeger-query-59cd5ddd7c-bfh56 to ip-10-0-52-42.us-west-2.compute.internal default-scheduler logger.go:42: 07:11:46 | collector-otlp-production-http | 2024-09-12 07:11:18 +0000 UTC Normal Pod my-jaeger-query-59cd5ddd7c-bfh56 AddedInterface Add eth0 [10.128.0.45/23] from ovn-kubernetes logger.go:42: 07:11:46 | collector-otlp-production-http | 2024-09-12 07:11:18 +0000 UTC Normal Pod my-jaeger-query-59cd5ddd7c-bfh56.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a9186dcd910256c0f464b0a3928844a01de166a10c186c97ef4581bf288c23cb" already present on machine kubelet logger.go:42: 07:11:46 | collector-otlp-production-http | 2024-09-12 07:11:18 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-59cd5ddd7c SuccessfulCreate Created pod: my-jaeger-query-59cd5ddd7c-bfh56 replicaset-controller logger.go:42: 07:11:46 | collector-otlp-production-http | 2024-09-12 07:11:18 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-59cd5ddd7c to 1 deployment-controller logger.go:42: 07:11:46 | collector-otlp-production-http | 2024-09-12 07:11:19 +0000 UTC Normal Pod my-jaeger-collector-57cbd56974-t9vl4.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:11:46 | collector-otlp-production-http | 2024-09-12 07:11:19 +0000 UTC Normal Pod my-jaeger-collector-57cbd56974-t9vl4.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:11:46 | collector-otlp-production-http | 2024-09-12 07:11:19 +0000 UTC Normal Pod my-jaeger-query-59cd5ddd7c-bfh56.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:11:46 | collector-otlp-production-http | 2024-09-12 07:11:19 +0000 UTC Normal Pod my-jaeger-query-59cd5ddd7c-bfh56.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:11:46 | collector-otlp-production-http | 2024-09-12 07:11:19 +0000 UTC Normal Pod my-jaeger-query-59cd5ddd7c-bfh56.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:11:46 | collector-otlp-production-http | 2024-09-12 07:11:19 +0000 UTC Normal Pod my-jaeger-query-59cd5ddd7c-bfh56.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:11:46 | collector-otlp-production-http | 2024-09-12 07:11:19 +0000 UTC Normal Pod my-jaeger-query-59cd5ddd7c-bfh56.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:11:46 | collector-otlp-production-http | 2024-09-12 07:11:19 +0000 UTC Normal Pod my-jaeger-query-59cd5ddd7c-bfh56.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" already present on machine kubelet logger.go:42: 07:11:46 | collector-otlp-production-http | 2024-09-12 07:11:19 +0000 UTC Normal Pod my-jaeger-query-59cd5ddd7c-bfh56.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:11:46 | collector-otlp-production-http | 2024-09-12 07:11:19 +0000 UTC Normal Pod my-jaeger-query-59cd5ddd7c-bfh56.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:11:46 | collector-otlp-production-http | 2024-09-12 07:11:23 +0000 UTC Normal Pod my-jaeger-query-59cd5ddd7c-bfh56.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:11:46 | collector-otlp-production-http | 2024-09-12 07:11:23 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-59cd5ddd7c SuccessfulDelete Deleted pod: my-jaeger-query-59cd5ddd7c-bfh56 replicaset-controller logger.go:42: 07:11:46 | collector-otlp-production-http | 2024-09-12 07:11:23 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled down replica set my-jaeger-query-59cd5ddd7c to 0 from 1 deployment-controller logger.go:42: 07:11:46 | collector-otlp-production-http | 2024-09-12 07:11:24 +0000 UTC Normal Pod my-jaeger-query-59cd5ddd7c-bfh56.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:11:46 | collector-otlp-production-http | 2024-09-12 07:11:24 +0000 UTC Normal Pod my-jaeger-query-59cd5ddd7c-bfh56.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:11:46 | collector-otlp-production-http | 2024-09-12 07:11:24 +0000 UTC Normal Pod my-jaeger-query-d98fc7cf-sbvdb Binding Scheduled Successfully assigned kuttl-test-cheerful-grouse/my-jaeger-query-d98fc7cf-sbvdb to ip-10-0-52-42.us-west-2.compute.internal default-scheduler logger.go:42: 07:11:46 | collector-otlp-production-http | 2024-09-12 07:11:24 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-d98fc7cf SuccessfulCreate Created pod: my-jaeger-query-d98fc7cf-sbvdb replicaset-controller logger.go:42: 07:11:46 | collector-otlp-production-http | 2024-09-12 07:11:24 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-d98fc7cf to 1 deployment-controller logger.go:42: 07:11:46 | collector-otlp-production-http | 2024-09-12 07:11:25 +0000 UTC Normal Pod my-jaeger-query-d98fc7cf-sbvdb AddedInterface Add eth0 [10.128.0.46/23] from ovn-kubernetes logger.go:42: 07:11:46 | collector-otlp-production-http | 2024-09-12 07:11:25 +0000 UTC Normal Pod my-jaeger-query-d98fc7cf-sbvdb.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a9186dcd910256c0f464b0a3928844a01de166a10c186c97ef4581bf288c23cb" already present on machine kubelet logger.go:42: 07:11:46 | collector-otlp-production-http | 2024-09-12 07:11:25 +0000 UTC Normal Pod my-jaeger-query-d98fc7cf-sbvdb.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:11:46 | collector-otlp-production-http | 2024-09-12 07:11:25 +0000 UTC Normal Pod my-jaeger-query-d98fc7cf-sbvdb.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:11:46 | collector-otlp-production-http | 2024-09-12 07:11:25 +0000 UTC Normal Pod my-jaeger-query-d98fc7cf-sbvdb.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:11:46 | collector-otlp-production-http | 2024-09-12 07:11:25 +0000 UTC Normal Pod my-jaeger-query-d98fc7cf-sbvdb.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:11:46 | collector-otlp-production-http | 2024-09-12 07:11:25 +0000 UTC Normal Pod my-jaeger-query-d98fc7cf-sbvdb.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:11:46 | collector-otlp-production-http | 2024-09-12 07:11:25 +0000 UTC Normal Pod my-jaeger-query-d98fc7cf-sbvdb.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" already present on machine kubelet logger.go:42: 07:11:46 | collector-otlp-production-http | 2024-09-12 07:11:25 +0000 UTC Normal Pod my-jaeger-query-d98fc7cf-sbvdb.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:11:46 | collector-otlp-production-http | 2024-09-12 07:11:25 +0000 UTC Normal Pod my-jaeger-query-d98fc7cf-sbvdb.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:11:46 | collector-otlp-production-http | 2024-09-12 07:11:31 +0000 UTC Normal Pod check-span-6f6p8 Binding Scheduled Successfully assigned kuttl-test-cheerful-grouse/check-span-6f6p8 to ip-10-0-54-110.us-west-2.compute.internal default-scheduler logger.go:42: 07:11:46 | collector-otlp-production-http | 2024-09-12 07:11:31 +0000 UTC Normal Pod check-span-6f6p8 AddedInterface Add eth0 [10.129.0.37/23] from ovn-kubernetes logger.go:42: 07:11:46 | collector-otlp-production-http | 2024-09-12 07:11:31 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-6f6p8 job-controller logger.go:42: 07:11:46 | collector-otlp-production-http | 2024-09-12 07:11:31 +0000 UTC Normal Pod report-span-gpdmc Binding Scheduled Successfully assigned kuttl-test-cheerful-grouse/report-span-gpdmc to ip-10-0-54-110.us-west-2.compute.internal default-scheduler logger.go:42: 07:11:46 | collector-otlp-production-http | 2024-09-12 07:11:31 +0000 UTC Normal Pod report-span-gpdmc AddedInterface Add eth0 [10.129.0.36/23] from ovn-kubernetes logger.go:42: 07:11:46 | collector-otlp-production-http | 2024-09-12 07:11:31 +0000 UTC Normal Pod report-span-gpdmc.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:11:46 | collector-otlp-production-http | 2024-09-12 07:11:31 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-gpdmc job-controller logger.go:42: 07:11:46 | collector-otlp-production-http | 2024-09-12 07:11:32 +0000 UTC Normal Pod check-span-6f6p8.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:11:46 | collector-otlp-production-http | 2024-09-12 07:11:33 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:11:46 | collector-otlp-production-http | 2024-09-12 07:11:33 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:11:46 | collector-otlp-production-http | 2024-09-12 07:11:33 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:11:46 | collector-otlp-production-http | 2024-09-12 07:11:34 +0000 UTC Normal Pod check-span-6f6p8.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 2.859s (2.859s including waiting) kubelet logger.go:42: 07:11:46 | collector-otlp-production-http | 2024-09-12 07:11:34 +0000 UTC Normal Pod report-span-gpdmc.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 3.217s (3.217s including waiting) kubelet logger.go:42: 07:11:46 | collector-otlp-production-http | 2024-09-12 07:11:34 +0000 UTC Normal Pod report-span-gpdmc.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:11:46 | collector-otlp-production-http | 2024-09-12 07:11:35 +0000 UTC Normal Pod check-span-6f6p8.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:11:46 | collector-otlp-production-http | 2024-09-12 07:11:35 +0000 UTC Normal Pod check-span-6f6p8.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:11:46 | collector-otlp-production-http | 2024-09-12 07:11:35 +0000 UTC Normal Pod report-span-gpdmc.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:11:46 | collector-otlp-production-http | 2024-09-12 07:11:45 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:11:46 | collector-otlp-production-http | Deleting namespace: kuttl-test-cheerful-grouse === CONT kuttl/harness/collector-otlp-production-grpc logger.go:42: 07:11:58 | collector-otlp-production-grpc | Creating namespace: kuttl-test-profound-wallaby logger.go:42: 07:11:59 | collector-otlp-production-grpc/1-install | starting test step 1-install logger.go:42: 07:11:59 | collector-otlp-production-grpc/1-install | Jaeger:kuttl-test-profound-wallaby/my-jaeger created logger.go:42: 07:13:07 | collector-otlp-production-grpc/1-install | test step completed 1-install logger.go:42: 07:13:07 | collector-otlp-production-grpc/2-smoke-test | starting test step 2-smoke-test logger.go:42: 07:13:07 | collector-otlp-production-grpc/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:13:08 | collector-otlp-production-grpc/2-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:13:17 | collector-otlp-production-grpc/2-smoke-test | running command: [sh -c REPORTING_PROTOCOL=grpc ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/otlp-smoke-test.yaml.template -o otlp-smoke-test-job.yaml] logger.go:42: 07:13:18 | collector-otlp-production-grpc/2-smoke-test | running command: [sh -c kubectl create -f otlp-smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:13:26 | collector-otlp-production-grpc/2-smoke-test | job.batch/report-span created logger.go:42: 07:13:26 | collector-otlp-production-grpc/2-smoke-test | job.batch/check-span created logger.go:42: 07:13:46 | collector-otlp-production-grpc/2-smoke-test | test step completed 2-smoke-test logger.go:42: 07:13:46 | collector-otlp-production-grpc | collector-otlp-production-grpc events from ns kuttl-test-profound-wallaby: logger.go:42: 07:13:46 | collector-otlp-production-grpc | 2024-09-12 07:12:02 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestprofoundwallabymyjaeger-1-6b44fcs949 Binding Scheduled Successfully assigned kuttl-test-profound-wallaby/elasticsearch-cdm-kuttltestprofoundwallabymyjaeger-1-6b44fcs949 to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:13:46 | collector-otlp-production-grpc | 2024-09-12 07:12:02 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestprofoundwallabymyjaeger-1-6b44fcs949 FailedMount MountVolume.SetUp failed for volume "elasticsearch-metrics" : secret "elasticsearch-metrics" not found kubelet logger.go:42: 07:13:46 | collector-otlp-production-grpc | 2024-09-12 07:12:02 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestprofoundwallabymyjaeger-1-6b44ffbc55 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestprofoundwallabymyjaeger-1-6b44fcs949 replicaset-controller logger.go:42: 07:13:46 | collector-otlp-production-grpc | 2024-09-12 07:12:02 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestprofoundwallabymyjaeger-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestprofoundwallabymyjaeger-1-6b44ffbc55 to 1 deployment-controller logger.go:42: 07:13:46 | collector-otlp-production-grpc | 2024-09-12 07:12:03 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestprofoundwallabymyjaeger-1-6b44fcs949 AddedInterface Add eth0 [10.130.0.42/23] from ovn-kubernetes logger.go:42: 07:13:46 | collector-otlp-production-grpc | 2024-09-12 07:12:03 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestprofoundwallabymyjaeger-1-6b44fcs949.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:bc97b8e13087050dfb1a75b02c5b780cbb8fb12a849a655169d072ba8bbf42b4" already present on machine kubelet logger.go:42: 07:13:46 | collector-otlp-production-grpc | 2024-09-12 07:12:03 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestprofoundwallabymyjaeger-1-6b44fcs949.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:13:46 | collector-otlp-production-grpc | 2024-09-12 07:12:03 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestprofoundwallabymyjaeger-1-6b44fcs949.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:13:46 | collector-otlp-production-grpc | 2024-09-12 07:12:03 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestprofoundwallabymyjaeger-1-6b44fcs949.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:1d1d7d47b616995d18692f3a2d6232d7e1a0f41bc1503d87100f31939e080a78" already present on machine kubelet logger.go:42: 07:13:46 | collector-otlp-production-grpc | 2024-09-12 07:12:03 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestprofoundwallabymyjaeger-1-6b44fcs949.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:13:46 | collector-otlp-production-grpc | 2024-09-12 07:12:03 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestprofoundwallabymyjaeger-1-6b44fcs949.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:13:46 | collector-otlp-production-grpc | 2024-09-12 07:12:13 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestprofoundwallabymyjaeger-1-6b44fcs949.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 07:13:46 | collector-otlp-production-grpc | 2024-09-12 07:12:18 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestprofoundwallabymyjaeger-1-6b44fcs949.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:13:46 | collector-otlp-production-grpc | 2024-09-12 07:12:29 +0000 UTC Normal Pod my-jaeger-collector-7b567b894d-rdzl4 Binding Scheduled Successfully assigned kuttl-test-profound-wallaby/my-jaeger-collector-7b567b894d-rdzl4 to ip-10-0-54-110.us-west-2.compute.internal default-scheduler logger.go:42: 07:13:46 | collector-otlp-production-grpc | 2024-09-12 07:12:29 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-7b567b894d SuccessfulCreate Created pod: my-jaeger-collector-7b567b894d-rdzl4 replicaset-controller logger.go:42: 07:13:46 | collector-otlp-production-grpc | 2024-09-12 07:12:29 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-7b567b894d to 1 deployment-controller logger.go:42: 07:13:46 | collector-otlp-production-grpc | 2024-09-12 07:12:29 +0000 UTC Normal Pod my-jaeger-query-6f4cdf8667-4xn8h Binding Scheduled Successfully assigned kuttl-test-profound-wallaby/my-jaeger-query-6f4cdf8667-4xn8h to ip-10-0-52-42.us-west-2.compute.internal default-scheduler logger.go:42: 07:13:46 | collector-otlp-production-grpc | 2024-09-12 07:12:29 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-6f4cdf8667 SuccessfulCreate Created pod: my-jaeger-query-6f4cdf8667-4xn8h replicaset-controller logger.go:42: 07:13:46 | collector-otlp-production-grpc | 2024-09-12 07:12:29 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-6f4cdf8667 to 1 deployment-controller logger.go:42: 07:13:46 | collector-otlp-production-grpc | 2024-09-12 07:12:30 +0000 UTC Normal Pod my-jaeger-collector-7b567b894d-rdzl4 AddedInterface Add eth0 [10.129.0.38/23] from ovn-kubernetes logger.go:42: 07:13:46 | collector-otlp-production-grpc | 2024-09-12 07:12:30 +0000 UTC Normal Pod my-jaeger-collector-7b567b894d-rdzl4.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:4bab63290ff52e4f6328009f9c8b2c4496b230e9f8a98eac01736a66a291ff6c" already present on machine kubelet logger.go:42: 07:13:46 | collector-otlp-production-grpc | 2024-09-12 07:12:30 +0000 UTC Normal Pod my-jaeger-collector-7b567b894d-rdzl4.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:13:46 | collector-otlp-production-grpc | 2024-09-12 07:12:30 +0000 UTC Normal Pod my-jaeger-collector-7b567b894d-rdzl4.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:13:46 | collector-otlp-production-grpc | 2024-09-12 07:12:30 +0000 UTC Normal Pod my-jaeger-query-6f4cdf8667-4xn8h AddedInterface Add eth0 [10.128.0.47/23] from ovn-kubernetes logger.go:42: 07:13:46 | collector-otlp-production-grpc | 2024-09-12 07:12:30 +0000 UTC Normal Pod my-jaeger-query-6f4cdf8667-4xn8h.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a9186dcd910256c0f464b0a3928844a01de166a10c186c97ef4581bf288c23cb" already present on machine kubelet logger.go:42: 07:13:46 | collector-otlp-production-grpc | 2024-09-12 07:12:30 +0000 UTC Normal Pod my-jaeger-query-6f4cdf8667-4xn8h.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:13:46 | collector-otlp-production-grpc | 2024-09-12 07:12:30 +0000 UTC Normal Pod my-jaeger-query-6f4cdf8667-4xn8h.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:13:46 | collector-otlp-production-grpc | 2024-09-12 07:12:30 +0000 UTC Normal Pod my-jaeger-query-6f4cdf8667-4xn8h.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:13:46 | collector-otlp-production-grpc | 2024-09-12 07:12:30 +0000 UTC Normal Pod my-jaeger-query-6f4cdf8667-4xn8h.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:13:46 | collector-otlp-production-grpc | 2024-09-12 07:12:30 +0000 UTC Normal Pod my-jaeger-query-6f4cdf8667-4xn8h.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:13:46 | collector-otlp-production-grpc | 2024-09-12 07:12:30 +0000 UTC Normal Pod my-jaeger-query-6f4cdf8667-4xn8h.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" already present on machine kubelet logger.go:42: 07:13:46 | collector-otlp-production-grpc | 2024-09-12 07:12:30 +0000 UTC Normal Pod my-jaeger-query-6f4cdf8667-4xn8h.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:13:46 | collector-otlp-production-grpc | 2024-09-12 07:12:30 +0000 UTC Normal Pod my-jaeger-query-6f4cdf8667-4xn8h.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:13:46 | collector-otlp-production-grpc | 2024-09-12 07:12:54 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: an error on the server ("Internal Server Error: \"/apis/metrics.k8s.io/v1beta1/namespaces/kuttl-test-profound-wallaby/pods?labelSelector=app%3Djaeger%2Capp.kubernetes.io%2Fcomponent%3Dcollector%2Capp.kubernetes.io%2Finstance%3Dmy-jaeger%2Capp.kubernetes.io%2Fmanaged-by%3Djaeger-operator%2Capp.kubernetes.io%2Fname%3Dmy-jaeger-collector%2Capp.kubernetes.io%2Fpart-of%3Djaeger\": Post \"https://172.30.0.1:443/apis/authorization.k8s.io/v1/subjectaccessreviews?timeout=10s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)") has prevented the request from succeeding (get pods.metrics.k8s.io) horizontal-pod-autoscaler logger.go:42: 07:13:46 | collector-otlp-production-grpc | 2024-09-12 07:13:02 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: unable to fetch metrics from resource metrics API: an error on the server ("Internal Server Error: \"/apis/metrics.k8s.io/v1beta1/namespaces/kuttl-test-profound-wallaby/pods?labelSelector=app%3Djaeger%2Capp.kubernetes.io%2Fcomponent%3Dcollector%2Capp.kubernetes.io%2Finstance%3Dmy-jaeger%2Capp.kubernetes.io%2Fmanaged-by%3Djaeger-operator%2Capp.kubernetes.io%2Fname%3Dmy-jaeger-collector%2Capp.kubernetes.io%2Fpart-of%3Djaeger\": Post \"https://172.30.0.1:443/apis/authorization.k8s.io/v1/subjectaccessreviews?timeout=10s\": http2: client connection lost") has prevented the request from succeeding (get pods.metrics.k8s.io) horizontal-pod-autoscaler logger.go:42: 07:13:46 | collector-otlp-production-grpc | 2024-09-12 07:13:02 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: unable to fetch metrics from resource metrics API: an error on the server ("Internal Server Error: \"/apis/metrics.k8s.io/v1beta1/namespaces/kuttl-test-profound-wallaby/pods?labelSelector=app%3Djaeger%2Capp.kubernetes.io%2Fcomponent%3Dcollector%2Capp.kubernetes.io%2Finstance%3Dmy-jaeger%2Capp.kubernetes.io%2Fmanaged-by%3Djaeger-operator%2Capp.kubernetes.io%2Fname%3Dmy-jaeger-collector%2Capp.kubernetes.io%2Fpart-of%3Djaeger\": Post \"https://172.30.0.1:443/apis/authorization.k8s.io/v1/subjectaccessreviews?timeout=10s\": net/http: request canceled (Client.Timeout exceeded while awaiting headers)") has prevented the request from succeeding (get pods.metrics.k8s.io) horizontal-pod-autoscaler logger.go:42: 07:13:46 | collector-otlp-production-grpc | 2024-09-12 07:13:10 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:13:46 | collector-otlp-production-grpc | 2024-09-12 07:13:10 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:13:46 | collector-otlp-production-grpc | 2024-09-12 07:13:10 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:13:46 | collector-otlp-production-grpc | 2024-09-12 07:13:11 +0000 UTC Normal Pod my-jaeger-query-6f4cdf8667-4xn8h.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:13:46 | collector-otlp-production-grpc | 2024-09-12 07:13:11 +0000 UTC Normal Pod my-jaeger-query-6f4cdf8667-4xn8h.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:13:46 | collector-otlp-production-grpc | 2024-09-12 07:13:11 +0000 UTC Normal Pod my-jaeger-query-6f4cdf8667-4xn8h.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:13:46 | collector-otlp-production-grpc | 2024-09-12 07:13:11 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-6f4cdf8667 SuccessfulDelete Deleted pod: my-jaeger-query-6f4cdf8667-4xn8h replicaset-controller logger.go:42: 07:13:46 | collector-otlp-production-grpc | 2024-09-12 07:13:11 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled down replica set my-jaeger-query-6f4cdf8667 to 0 from 1 deployment-controller logger.go:42: 07:13:46 | collector-otlp-production-grpc | 2024-09-12 07:13:12 +0000 UTC Normal Pod my-jaeger-query-688797cb59-j2t82 Binding Scheduled Successfully assigned kuttl-test-profound-wallaby/my-jaeger-query-688797cb59-j2t82 to ip-10-0-52-42.us-west-2.compute.internal default-scheduler logger.go:42: 07:13:46 | collector-otlp-production-grpc | 2024-09-12 07:13:12 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-688797cb59 SuccessfulCreate Created pod: my-jaeger-query-688797cb59-j2t82 replicaset-controller logger.go:42: 07:13:46 | collector-otlp-production-grpc | 2024-09-12 07:13:12 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-688797cb59 to 1 deployment-controller logger.go:42: 07:13:46 | collector-otlp-production-grpc | 2024-09-12 07:13:13 +0000 UTC Normal Pod my-jaeger-query-688797cb59-j2t82 AddedInterface Add eth0 [10.128.0.48/23] from ovn-kubernetes logger.go:42: 07:13:46 | collector-otlp-production-grpc | 2024-09-12 07:13:13 +0000 UTC Normal Pod my-jaeger-query-688797cb59-j2t82.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a9186dcd910256c0f464b0a3928844a01de166a10c186c97ef4581bf288c23cb" already present on machine kubelet logger.go:42: 07:13:46 | collector-otlp-production-grpc | 2024-09-12 07:13:13 +0000 UTC Normal Pod my-jaeger-query-688797cb59-j2t82.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:13:46 | collector-otlp-production-grpc | 2024-09-12 07:13:13 +0000 UTC Normal Pod my-jaeger-query-688797cb59-j2t82.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:13:46 | collector-otlp-production-grpc | 2024-09-12 07:13:13 +0000 UTC Normal Pod my-jaeger-query-688797cb59-j2t82.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:13:46 | collector-otlp-production-grpc | 2024-09-12 07:13:13 +0000 UTC Normal Pod my-jaeger-query-688797cb59-j2t82.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:13:46 | collector-otlp-production-grpc | 2024-09-12 07:13:13 +0000 UTC Normal Pod my-jaeger-query-688797cb59-j2t82.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:13:46 | collector-otlp-production-grpc | 2024-09-12 07:13:13 +0000 UTC Normal Pod my-jaeger-query-688797cb59-j2t82.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" already present on machine kubelet logger.go:42: 07:13:46 | collector-otlp-production-grpc | 2024-09-12 07:13:13 +0000 UTC Normal Pod my-jaeger-query-688797cb59-j2t82.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:13:46 | collector-otlp-production-grpc | 2024-09-12 07:13:13 +0000 UTC Normal Pod my-jaeger-query-688797cb59-j2t82.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:13:46 | collector-otlp-production-grpc | 2024-09-12 07:13:26 +0000 UTC Normal Pod check-span-9xlwc Binding Scheduled Successfully assigned kuttl-test-profound-wallaby/check-span-9xlwc to ip-10-0-54-110.us-west-2.compute.internal default-scheduler logger.go:42: 07:13:46 | collector-otlp-production-grpc | 2024-09-12 07:13:26 +0000 UTC Normal Pod check-span-9xlwc AddedInterface Add eth0 [10.129.0.40/23] from ovn-kubernetes logger.go:42: 07:13:46 | collector-otlp-production-grpc | 2024-09-12 07:13:26 +0000 UTC Normal Pod check-span-9xlwc.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:13:46 | collector-otlp-production-grpc | 2024-09-12 07:13:26 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-9xlwc job-controller logger.go:42: 07:13:46 | collector-otlp-production-grpc | 2024-09-12 07:13:26 +0000 UTC Normal Pod report-span-mzv9j Binding Scheduled Successfully assigned kuttl-test-profound-wallaby/report-span-mzv9j to ip-10-0-54-110.us-west-2.compute.internal default-scheduler logger.go:42: 07:13:46 | collector-otlp-production-grpc | 2024-09-12 07:13:26 +0000 UTC Normal Pod report-span-mzv9j AddedInterface Add eth0 [10.129.0.39/23] from ovn-kubernetes logger.go:42: 07:13:46 | collector-otlp-production-grpc | 2024-09-12 07:13:26 +0000 UTC Normal Pod report-span-mzv9j.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:13:46 | collector-otlp-production-grpc | 2024-09-12 07:13:26 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-mzv9j job-controller logger.go:42: 07:13:46 | collector-otlp-production-grpc | 2024-09-12 07:13:27 +0000 UTC Normal Pod check-span-9xlwc.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 847ms (848ms including waiting) kubelet logger.go:42: 07:13:46 | collector-otlp-production-grpc | 2024-09-12 07:13:27 +0000 UTC Normal Pod check-span-9xlwc.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:13:46 | collector-otlp-production-grpc | 2024-09-12 07:13:27 +0000 UTC Normal Pod check-span-9xlwc.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:13:46 | collector-otlp-production-grpc | 2024-09-12 07:13:27 +0000 UTC Normal Pod report-span-mzv9j.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 852ms (852ms including waiting) kubelet logger.go:42: 07:13:46 | collector-otlp-production-grpc | 2024-09-12 07:13:27 +0000 UTC Normal Pod report-span-mzv9j.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:13:46 | collector-otlp-production-grpc | 2024-09-12 07:13:27 +0000 UTC Normal Pod report-span-mzv9j.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:13:46 | collector-otlp-production-grpc | 2024-09-12 07:13:40 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 07:13:46 | collector-otlp-production-grpc | 2024-09-12 07:13:40 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get memory utilization: missing request for memory in container jaeger-collector of Pod my-jaeger-collector-7b567b894d-rdzl4 horizontal-pod-autoscaler logger.go:42: 07:13:46 | collector-otlp-production-grpc | 2024-09-12 07:13:40 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 07:13:46 | collector-otlp-production-grpc | 2024-09-12 07:13:46 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:13:46 | collector-otlp-production-grpc | Deleting namespace: kuttl-test-profound-wallaby === CONT kuttl/harness/collector-autoscale logger.go:42: 07:13:59 | collector-autoscale | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:13:59 | collector-autoscale | Creating namespace: kuttl-test-many-monkey logger.go:42: 07:13:59 | collector-autoscale/1-install | starting test step 1-install logger.go:42: 07:13:59 | collector-autoscale/1-install | Jaeger:kuttl-test-many-monkey/simple-prod created logger.go:42: 07:14:36 | collector-autoscale/1-install | test step completed 1-install logger.go:42: 07:14:36 | collector-autoscale/2- | starting test step 2- logger.go:42: 07:14:36 | collector-autoscale/2- | test step completed 2- logger.go:42: 07:14:36 | collector-autoscale | collector-autoscale events from ns kuttl-test-many-monkey: logger.go:42: 07:14:36 | collector-autoscale | 2024-09-12 07:14:05 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestmanymonkeysimpleprod-1-5d5948d6c7 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestmanymonkeysimpleprod-1-5d5948d6r2lnm replicaset-controller logger.go:42: 07:14:36 | collector-autoscale | 2024-09-12 07:14:05 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmanymonkeysimpleprod-1-5d5948d6r2lnm Binding Scheduled Successfully assigned kuttl-test-many-monkey/elasticsearch-cdm-kuttltestmanymonkeysimpleprod-1-5d5948d6r2lnm to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:14:36 | collector-autoscale | 2024-09-12 07:14:05 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestmanymonkeysimpleprod-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestmanymonkeysimpleprod-1-5d5948d6c7 to 1 deployment-controller logger.go:42: 07:14:36 | collector-autoscale | 2024-09-12 07:14:06 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmanymonkeysimpleprod-1-5d5948d6r2lnm AddedInterface Add eth0 [10.130.0.43/23] from ovn-kubernetes logger.go:42: 07:14:36 | collector-autoscale | 2024-09-12 07:14:06 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmanymonkeysimpleprod-1-5d5948d6r2lnm.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:bc97b8e13087050dfb1a75b02c5b780cbb8fb12a849a655169d072ba8bbf42b4" already present on machine kubelet logger.go:42: 07:14:36 | collector-autoscale | 2024-09-12 07:14:06 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmanymonkeysimpleprod-1-5d5948d6r2lnm.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:14:36 | collector-autoscale | 2024-09-12 07:14:06 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmanymonkeysimpleprod-1-5d5948d6r2lnm.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:14:36 | collector-autoscale | 2024-09-12 07:14:06 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmanymonkeysimpleprod-1-5d5948d6r2lnm.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:1d1d7d47b616995d18692f3a2d6232d7e1a0f41bc1503d87100f31939e080a78" already present on machine kubelet logger.go:42: 07:14:36 | collector-autoscale | 2024-09-12 07:14:06 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmanymonkeysimpleprod-1-5d5948d6r2lnm.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:14:36 | collector-autoscale | 2024-09-12 07:14:06 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmanymonkeysimpleprod-1-5d5948d6r2lnm.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:14:36 | collector-autoscale | 2024-09-12 07:14:16 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestmanymonkeysimpleprod-1-5d5948d6r2lnm.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 07:14:36 | collector-autoscale | 2024-09-12 07:14:21 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestmanymonkeysimpleprod-1-5d5948d6r2lnm.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:14:36 | collector-autoscale | 2024-09-12 07:14:32 +0000 UTC Normal Pod simple-prod-collector-5cb64c9675-qg2bq Binding Scheduled Successfully assigned kuttl-test-many-monkey/simple-prod-collector-5cb64c9675-qg2bq to ip-10-0-54-110.us-west-2.compute.internal default-scheduler logger.go:42: 07:14:36 | collector-autoscale | 2024-09-12 07:14:32 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-5cb64c9675 SuccessfulCreate Created pod: simple-prod-collector-5cb64c9675-qg2bq replicaset-controller logger.go:42: 07:14:36 | collector-autoscale | 2024-09-12 07:14:32 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-5cb64c9675 to 1 deployment-controller logger.go:42: 07:14:36 | collector-autoscale | 2024-09-12 07:14:32 +0000 UTC Normal Pod simple-prod-query-6c6b87b865-wjrgw Binding Scheduled Successfully assigned kuttl-test-many-monkey/simple-prod-query-6c6b87b865-wjrgw to ip-10-0-52-42.us-west-2.compute.internal default-scheduler logger.go:42: 07:14:36 | collector-autoscale | 2024-09-12 07:14:32 +0000 UTC Normal ReplicaSet.apps simple-prod-query-6c6b87b865 SuccessfulCreate Created pod: simple-prod-query-6c6b87b865-wjrgw replicaset-controller logger.go:42: 07:14:36 | collector-autoscale | 2024-09-12 07:14:32 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-6c6b87b865 to 1 deployment-controller logger.go:42: 07:14:36 | collector-autoscale | 2024-09-12 07:14:33 +0000 UTC Normal Pod simple-prod-collector-5cb64c9675-qg2bq AddedInterface Add eth0 [10.129.0.41/23] from ovn-kubernetes logger.go:42: 07:14:36 | collector-autoscale | 2024-09-12 07:14:33 +0000 UTC Normal Pod simple-prod-collector-5cb64c9675-qg2bq.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:4bab63290ff52e4f6328009f9c8b2c4496b230e9f8a98eac01736a66a291ff6c" already present on machine kubelet logger.go:42: 07:14:36 | collector-autoscale | 2024-09-12 07:14:33 +0000 UTC Normal Pod simple-prod-collector-5cb64c9675-qg2bq.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:14:36 | collector-autoscale | 2024-09-12 07:14:33 +0000 UTC Normal Pod simple-prod-collector-5cb64c9675-qg2bq.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:14:36 | collector-autoscale | 2024-09-12 07:14:33 +0000 UTC Normal Pod simple-prod-query-6c6b87b865-wjrgw AddedInterface Add eth0 [10.128.0.49/23] from ovn-kubernetes logger.go:42: 07:14:36 | collector-autoscale | 2024-09-12 07:14:33 +0000 UTC Normal Pod simple-prod-query-6c6b87b865-wjrgw.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a9186dcd910256c0f464b0a3928844a01de166a10c186c97ef4581bf288c23cb" already present on machine kubelet logger.go:42: 07:14:36 | collector-autoscale | 2024-09-12 07:14:33 +0000 UTC Normal Pod simple-prod-query-6c6b87b865-wjrgw.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:14:36 | collector-autoscale | 2024-09-12 07:14:33 +0000 UTC Normal Pod simple-prod-query-6c6b87b865-wjrgw.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:14:36 | collector-autoscale | 2024-09-12 07:14:33 +0000 UTC Normal Pod simple-prod-query-6c6b87b865-wjrgw.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:14:36 | collector-autoscale | 2024-09-12 07:14:33 +0000 UTC Normal Pod simple-prod-query-6c6b87b865-wjrgw.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:14:36 | collector-autoscale | 2024-09-12 07:14:33 +0000 UTC Normal Pod simple-prod-query-6c6b87b865-wjrgw.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:14:36 | collector-autoscale | 2024-09-12 07:14:33 +0000 UTC Normal Pod simple-prod-query-6c6b87b865-wjrgw.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" already present on machine kubelet logger.go:42: 07:14:36 | collector-autoscale | 2024-09-12 07:14:33 +0000 UTC Normal Pod simple-prod-query-6c6b87b865-wjrgw.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:14:36 | collector-autoscale | 2024-09-12 07:14:33 +0000 UTC Normal Pod simple-prod-query-6c6b87b865-wjrgw.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:14:36 | collector-autoscale | Deleting namespace: kuttl-test-many-monkey === CONT kuttl/harness/collector-otlp-allinone-grpc logger.go:42: 07:14:43 | collector-otlp-allinone-grpc | Creating namespace: kuttl-test-neat-horse logger.go:42: 07:14:43 | collector-otlp-allinone-grpc/0-install | starting test step 0-install logger.go:42: 07:14:43 | collector-otlp-allinone-grpc/0-install | Jaeger:kuttl-test-neat-horse/my-jaeger created logger.go:42: 07:14:46 | collector-otlp-allinone-grpc/0-install | test step completed 0-install logger.go:42: 07:14:46 | collector-otlp-allinone-grpc/1-smoke-test | starting test step 1-smoke-test logger.go:42: 07:14:46 | collector-otlp-allinone-grpc/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:14:47 | collector-otlp-allinone-grpc/1-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:14:54 | collector-otlp-allinone-grpc/1-smoke-test | running command: [sh -c REPORTING_PROTOCOL=grpc ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/otlp-smoke-test.yaml.template -o otlp-smoke-test-job.yaml] logger.go:42: 07:14:55 | collector-otlp-allinone-grpc/1-smoke-test | running command: [sh -c kubectl create -f otlp-smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:14:55 | collector-otlp-allinone-grpc/1-smoke-test | job.batch/report-span created logger.go:42: 07:14:55 | collector-otlp-allinone-grpc/1-smoke-test | job.batch/check-span created logger.go:42: 07:15:16 | collector-otlp-allinone-grpc/1-smoke-test | test step completed 1-smoke-test logger.go:42: 07:15:16 | collector-otlp-allinone-grpc | collector-otlp-allinone-grpc events from ns kuttl-test-neat-horse: logger.go:42: 07:15:16 | collector-otlp-allinone-grpc | 2024-09-12 07:14:44 +0000 UTC Normal Pod my-jaeger-8769b8d75-wsscm Binding Scheduled Successfully assigned kuttl-test-neat-horse/my-jaeger-8769b8d75-wsscm to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:15:16 | collector-otlp-allinone-grpc | 2024-09-12 07:14:44 +0000 UTC Normal Pod my-jaeger-8769b8d75-wsscm AddedInterface Add eth0 [10.130.0.44/23] from ovn-kubernetes logger.go:42: 07:15:16 | collector-otlp-allinone-grpc | 2024-09-12 07:14:44 +0000 UTC Normal Pod my-jaeger-8769b8d75-wsscm.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" already present on machine kubelet logger.go:42: 07:15:16 | collector-otlp-allinone-grpc | 2024-09-12 07:14:44 +0000 UTC Normal Pod my-jaeger-8769b8d75-wsscm.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:15:16 | collector-otlp-allinone-grpc | 2024-09-12 07:14:44 +0000 UTC Normal Pod my-jaeger-8769b8d75-wsscm.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:15:16 | collector-otlp-allinone-grpc | 2024-09-12 07:14:44 +0000 UTC Normal Pod my-jaeger-8769b8d75-wsscm.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:15:16 | collector-otlp-allinone-grpc | 2024-09-12 07:14:44 +0000 UTC Normal Pod my-jaeger-8769b8d75-wsscm.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:15:16 | collector-otlp-allinone-grpc | 2024-09-12 07:14:44 +0000 UTC Normal Pod my-jaeger-8769b8d75-wsscm.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:15:16 | collector-otlp-allinone-grpc | 2024-09-12 07:14:44 +0000 UTC Normal ReplicaSet.apps my-jaeger-8769b8d75 SuccessfulCreate Created pod: my-jaeger-8769b8d75-wsscm replicaset-controller logger.go:42: 07:15:16 | collector-otlp-allinone-grpc | 2024-09-12 07:14:44 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-8769b8d75 to 1 deployment-controller logger.go:42: 07:15:16 | collector-otlp-allinone-grpc | 2024-09-12 07:14:48 +0000 UTC Normal Pod my-jaeger-8769b8d75-wsscm.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:15:16 | collector-otlp-allinone-grpc | 2024-09-12 07:14:48 +0000 UTC Normal Pod my-jaeger-8769b8d75-wsscm.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:15:16 | collector-otlp-allinone-grpc | 2024-09-12 07:14:48 +0000 UTC Normal ReplicaSet.apps my-jaeger-8769b8d75 SuccessfulDelete Deleted pod: my-jaeger-8769b8d75-wsscm replicaset-controller logger.go:42: 07:15:16 | collector-otlp-allinone-grpc | 2024-09-12 07:14:48 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled down replica set my-jaeger-8769b8d75 to 0 from 1 deployment-controller logger.go:42: 07:15:16 | collector-otlp-allinone-grpc | 2024-09-12 07:14:49 +0000 UTC Normal Pod my-jaeger-6664bb8fb8-qrbx8 Binding Scheduled Successfully assigned kuttl-test-neat-horse/my-jaeger-6664bb8fb8-qrbx8 to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:15:16 | collector-otlp-allinone-grpc | 2024-09-12 07:14:49 +0000 UTC Normal ReplicaSet.apps my-jaeger-6664bb8fb8 SuccessfulCreate Created pod: my-jaeger-6664bb8fb8-qrbx8 replicaset-controller logger.go:42: 07:15:16 | collector-otlp-allinone-grpc | 2024-09-12 07:14:49 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-6664bb8fb8 to 1 deployment-controller logger.go:42: 07:15:16 | collector-otlp-allinone-grpc | 2024-09-12 07:14:50 +0000 UTC Normal Pod my-jaeger-6664bb8fb8-qrbx8 AddedInterface Add eth0 [10.130.0.45/23] from ovn-kubernetes logger.go:42: 07:15:16 | collector-otlp-allinone-grpc | 2024-09-12 07:14:50 +0000 UTC Normal Pod my-jaeger-6664bb8fb8-qrbx8.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" already present on machine kubelet logger.go:42: 07:15:16 | collector-otlp-allinone-grpc | 2024-09-12 07:14:50 +0000 UTC Normal Pod my-jaeger-6664bb8fb8-qrbx8.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:15:16 | collector-otlp-allinone-grpc | 2024-09-12 07:14:50 +0000 UTC Normal Pod my-jaeger-6664bb8fb8-qrbx8.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:15:16 | collector-otlp-allinone-grpc | 2024-09-12 07:14:50 +0000 UTC Normal Pod my-jaeger-6664bb8fb8-qrbx8.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:15:16 | collector-otlp-allinone-grpc | 2024-09-12 07:14:50 +0000 UTC Normal Pod my-jaeger-6664bb8fb8-qrbx8.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:15:16 | collector-otlp-allinone-grpc | 2024-09-12 07:14:50 +0000 UTC Normal Pod my-jaeger-6664bb8fb8-qrbx8.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:15:16 | collector-otlp-allinone-grpc | 2024-09-12 07:14:55 +0000 UTC Normal Pod check-span-f5zb9 Binding Scheduled Successfully assigned kuttl-test-neat-horse/check-span-f5zb9 to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:15:16 | collector-otlp-allinone-grpc | 2024-09-12 07:14:55 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-f5zb9 job-controller logger.go:42: 07:15:16 | collector-otlp-allinone-grpc | 2024-09-12 07:14:55 +0000 UTC Normal Pod report-span-hwn5l Binding Scheduled Successfully assigned kuttl-test-neat-horse/report-span-hwn5l to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:15:16 | collector-otlp-allinone-grpc | 2024-09-12 07:14:55 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-hwn5l job-controller logger.go:42: 07:15:16 | collector-otlp-allinone-grpc | 2024-09-12 07:14:56 +0000 UTC Normal Pod check-span-f5zb9 AddedInterface Add eth0 [10.130.0.47/23] from ovn-kubernetes logger.go:42: 07:15:16 | collector-otlp-allinone-grpc | 2024-09-12 07:14:56 +0000 UTC Normal Pod check-span-f5zb9.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:15:16 | collector-otlp-allinone-grpc | 2024-09-12 07:14:56 +0000 UTC Normal Pod report-span-hwn5l AddedInterface Add eth0 [10.130.0.46/23] from ovn-kubernetes logger.go:42: 07:15:16 | collector-otlp-allinone-grpc | 2024-09-12 07:14:56 +0000 UTC Normal Pod report-span-hwn5l.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:15:16 | collector-otlp-allinone-grpc | 2024-09-12 07:14:57 +0000 UTC Normal Pod check-span-f5zb9.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 818ms (818ms including waiting) kubelet logger.go:42: 07:15:16 | collector-otlp-allinone-grpc | 2024-09-12 07:14:57 +0000 UTC Normal Pod check-span-f5zb9.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:15:16 | collector-otlp-allinone-grpc | 2024-09-12 07:14:57 +0000 UTC Normal Pod check-span-f5zb9.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:15:16 | collector-otlp-allinone-grpc | 2024-09-12 07:14:57 +0000 UTC Normal Pod report-span-hwn5l.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 846ms (846ms including waiting) kubelet logger.go:42: 07:15:16 | collector-otlp-allinone-grpc | 2024-09-12 07:14:57 +0000 UTC Normal Pod report-span-hwn5l.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:15:16 | collector-otlp-allinone-grpc | 2024-09-12 07:14:57 +0000 UTC Normal Pod report-span-hwn5l.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:15:16 | collector-otlp-allinone-grpc | 2024-09-12 07:15:15 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:15:16 | collector-otlp-allinone-grpc | Deleting namespace: kuttl-test-neat-horse === CONT kuttl/harness/cassandra-spark logger.go:42: 07:15:28 | cassandra-spark | Ignoring 01-assert.yaml.template as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:15:28 | cassandra-spark | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:15:28 | cassandra-spark | Creating namespace: kuttl-test-pet-pegasus logger.go:42: 07:15:28 | cassandra-spark | cassandra-spark events from ns kuttl-test-pet-pegasus: logger.go:42: 07:15:28 | cassandra-spark | Deleting namespace: kuttl-test-pet-pegasus === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (419.69s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (6.10s) --- PASS: kuttl/harness/collector-otlp-allinone-http (44.57s) --- PASS: kuttl/harness/set-custom-img (80.38s) --- PASS: kuttl/harness/collector-otlp-production-http (72.43s) --- PASS: kuttl/harness/collector-otlp-production-grpc (120.30s) --- PASS: kuttl/harness/collector-autoscale (44.13s) --- PASS: kuttl/harness/collector-otlp-allinone-grpc (45.02s) --- PASS: kuttl/harness/cassandra-spark (6.11s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name miscellaneous --report --output /logs/artifacts/miscellaneous.xml ./artifacts/kuttl-report.xml time="2024-09-12T07:15:34Z" level=debug msg="Setting a new name for the test suites" time="2024-09-12T07:15:34Z" level=debug msg="Removing 'artifacts' TestCase" time="2024-09-12T07:15:34Z" level=debug msg="normalizing test case names" time="2024-09-12T07:15:34Z" level=debug msg="miscellaneous/artifacts -> miscellaneous_artifacts" time="2024-09-12T07:15:34Z" level=debug msg="miscellaneous/collector-otlp-allinone-http -> miscellaneous_collector_otlp_allinone_http" time="2024-09-12T07:15:34Z" level=debug msg="miscellaneous/set-custom-img -> miscellaneous_set_custom_img" time="2024-09-12T07:15:34Z" level=debug msg="miscellaneous/collector-otlp-production-http -> miscellaneous_collector_otlp_production_http" time="2024-09-12T07:15:34Z" level=debug msg="miscellaneous/collector-otlp-production-grpc -> miscellaneous_collector_otlp_production_grpc" time="2024-09-12T07:15:34Z" level=debug msg="miscellaneous/collector-autoscale -> miscellaneous_collector_autoscale" time="2024-09-12T07:15:34Z" level=debug msg="miscellaneous/collector-otlp-allinone-grpc -> miscellaneous_collector_otlp_allinone_grpc" time="2024-09-12T07:15:34Z" level=debug msg="miscellaneous/cassandra-spark -> miscellaneous_cassandra_spark" +----------------------------------------------+--------+ | NAME | RESULT | +----------------------------------------------+--------+ | miscellaneous_artifacts | passed | | miscellaneous_collector_otlp_allinone_http | passed | | miscellaneous_set_custom_img | passed | | miscellaneous_collector_otlp_production_http | passed | | miscellaneous_collector_otlp_production_grpc | passed | | miscellaneous_collector_autoscale | passed | | miscellaneous_collector_otlp_allinone_grpc | passed | | miscellaneous_cassandra_spark | passed | +----------------------------------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh ui false true + '[' 3 -ne 3 ']' + test_suite_name=ui + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/ui.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-ui make[2]: Entering directory '/tmp/jaeger-tests' >>>> Elasticsearch image not loaded because SKIP_ES_EXTERNAL is true SKIP_ES_EXTERNAL=true ./tests/e2e/ui/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.29 True False 26m Cluster version is 4.15.29' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.29 True False 26m Cluster version is 4.15.29' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/ui/render.sh ++ export SUITE_DIR=./tests/e2e/ui ++ SUITE_DIR=./tests/e2e/ui ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/ui ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + start_test allinone + '[' 1 -ne 1 ']' + test_name=allinone + echo =========================================================================== =========================================================================== + info 'Rendering files for test allinone' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test allinone\e[0m' Rendering files for test allinone + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/ui/_build + '[' _build '!=' _build ']' + mkdir -p allinone + cd allinone + export GET_URL_COMMAND + export URL + export JAEGER_NAME=all-in-one-ui + JAEGER_NAME=all-in-one-ui + '[' true = true ']' + GET_URL_COMMAND='kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE' + URL='https://$(kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE)/search' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/ensure-ingress-host.sh.template -o ./ensure-ingress-host.sh + chmod +x ./ensure-ingress-host.sh + EXPECTED_CODE=200 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./01-curl.yaml + ASSERT_PRESENT=true + TRACKING_ID=MyTrackingId + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/test-ui-config.yaml.template -o ./04-test-ui-config.yaml + start_test production + '[' 1 -ne 1 ']' + test_name=production + echo =========================================================================== =========================================================================== + info 'Rendering files for test production' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test production\e[0m' Rendering files for test production + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/ui/_build/allinone + '[' allinone '!=' _build ']' + cd .. + mkdir -p production + cd production + export JAEGER_NAME=production-ui + JAEGER_NAME=production-ui + [[ true = true ]] + [[ true = true ]] + render_install_jaeger production-ui production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=production-ui + JAEGER_NAME=production-ui + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/ensure-ingress-host.sh.template -o ./ensure-ingress-host.sh + chmod +x ./ensure-ingress-host.sh + '[' true = true ']' + INSECURE=true + EXPECTED_CODE=403 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./02-check-forbbiden-access.yaml + EXPECTED_CODE=200 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./03-curl.yaml + INSECURE=true + EXPECTED_CODE=200 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./05-check-disabled-security.yaml + ASSERT_PRESENT=false + TRACKING_ID=MyTrackingId + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/test-ui-config.yaml.template -o ./06-check-NO-gaID.yaml + ASSERT_PRESENT=true + TRACKING_ID=MyTrackingId + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/test-ui-config.yaml.template -o ./08-check-gaID.yaml make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running ui E2E tests' Running ui E2E tests + cd tests/e2e/ui/_build + set +e + KUBECONFIG=/tmp/kubeconfig-2486720052 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-rosa-h-u0tm.f61o.s3.devshift.org:443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 3 tests === RUN kuttl/harness === RUN kuttl/harness/allinone === PAUSE kuttl/harness/allinone === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/production === PAUSE kuttl/harness/production === CONT kuttl/harness/allinone logger.go:42: 07:15:42 | allinone | Ignoring ensure-ingress-host.sh as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:15:42 | allinone | Creating namespace: kuttl-test-promoted-rat logger.go:42: 07:15:42 | allinone/0-install | starting test step 0-install logger.go:42: 07:15:42 | allinone/0-install | Jaeger:kuttl-test-promoted-rat/all-in-one-ui created logger.go:42: 07:15:43 | allinone/0-install | test step completed 0-install logger.go:42: 07:15:43 | allinone/1-curl | starting test step 1-curl logger.go:42: 07:15:43 | allinone/1-curl | running command: [./ensure-ingress-host.sh] logger.go:42: 07:15:43 | allinone/1-curl | Checking the Ingress host value was populated logger.go:42: 07:15:43 | allinone/1-curl | Try number 0 logger.go:42: 07:15:43 | allinone/1-curl | Hostname is all-in-one-ui-kuttl-test-promoted-rat.apps.rosa.ci-rosa-h-u0tm.f61o.s3.devshift.org logger.go:42: 07:15:43 | allinone/1-curl | running command: [sh -c ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 200 true $NAMESPACE all-in-one-ui] logger.go:42: 07:15:44 | allinone/1-curl | Checking an expected HTTP response logger.go:42: 07:15:44 | allinone/1-curl | Running in OpenShift logger.go:42: 07:15:44 | allinone/1-curl | User not provided. Getting the token... logger.go:42: 07:15:45 | allinone/1-curl | Warning: resource jaegers/all-in-one-ui is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:15:52 | allinone/1-curl | Try number 1/30 the https://all-in-one-ui-kuttl-test-promoted-rat.apps.rosa.ci-rosa-h-u0tm.f61o.s3.devshift.org/search logger.go:42: 07:15:53 | allinone/1-curl | curl response asserted properly logger.go:42: 07:15:53 | allinone/1-curl | test step completed 1-curl logger.go:42: 07:15:53 | allinone/2-delete | starting test step 2-delete logger.go:42: 07:15:53 | allinone/2-delete | Jaeger:kuttl-test-promoted-rat/all-in-one-ui created logger.go:42: 07:15:53 | allinone/2-delete | test step completed 2-delete logger.go:42: 07:15:53 | allinone/3-install | starting test step 3-install logger.go:42: 07:15:54 | allinone/3-install | Jaeger:kuttl-test-promoted-rat/all-in-one-ui updated logger.go:42: 07:15:54 | allinone/3-install | test step completed 3-install logger.go:42: 07:15:54 | allinone/4-test-ui-config | starting test step 4-test-ui-config logger.go:42: 07:15:54 | allinone/4-test-ui-config | running command: [./ensure-ingress-host.sh] logger.go:42: 07:15:54 | allinone/4-test-ui-config | Checking the Ingress host value was populated logger.go:42: 07:15:54 | allinone/4-test-ui-config | Try number 0 logger.go:42: 07:16:04 | allinone/4-test-ui-config | Try number 1 logger.go:42: 07:16:04 | allinone/4-test-ui-config | Hostname is all-in-one-ui-kuttl-test-promoted-rat.apps.rosa.ci-rosa-h-u0tm.f61o.s3.devshift.org logger.go:42: 07:16:04 | allinone/4-test-ui-config | running command: [sh -c ASSERT_PRESENT=true EXPECTED_CONTENT=MyTrackingId QUERY_HOSTNAME=https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search go run ../../../../cmd-utils/uiconfig/main.go] logger.go:42: 07:16:05 | allinone/4-test-ui-config | time="2024-09-12T07:16:05Z" level=info msg="Querying https://all-in-one-ui-kuttl-test-promoted-rat.apps.rosa.ci-rosa-h-u0tm.f61o.s3.devshift.org/search..." logger.go:42: 07:16:05 | allinone/4-test-ui-config | time="2024-09-12T07:16:05Z" level=info msg="No secret provided for the Authorization header" logger.go:42: 07:16:05 | allinone/4-test-ui-config | time="2024-09-12T07:16:05Z" level=info msg="Polling to https://all-in-one-ui-kuttl-test-promoted-rat.apps.rosa.ci-rosa-h-u0tm.f61o.s3.devshift.org/search" logger.go:42: 07:16:05 | allinone/4-test-ui-config | time="2024-09-12T07:16:05Z" level=info msg="Doing request number 0" logger.go:42: 07:16:06 | allinone/4-test-ui-config | time="2024-09-12T07:16:06Z" level=info msg="Content found and asserted!" logger.go:42: 07:16:06 | allinone/4-test-ui-config | time="2024-09-12T07:16:06Z" level=info msg="Success!" logger.go:42: 07:16:06 | allinone/4-test-ui-config | test step completed 4-test-ui-config logger.go:42: 07:16:06 | allinone | allinone events from ns kuttl-test-promoted-rat: logger.go:42: 07:16:06 | allinone | 2024-09-12 07:15:42 +0000 UTC Normal Pod all-in-one-ui-cd554ccb6-cqwbk Binding Scheduled Successfully assigned kuttl-test-promoted-rat/all-in-one-ui-cd554ccb6-cqwbk to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:16:06 | allinone | 2024-09-12 07:15:42 +0000 UTC Normal ReplicaSet.apps all-in-one-ui-cd554ccb6 SuccessfulCreate Created pod: all-in-one-ui-cd554ccb6-cqwbk replicaset-controller logger.go:42: 07:16:06 | allinone | 2024-09-12 07:15:42 +0000 UTC Normal Deployment.apps all-in-one-ui ScalingReplicaSet Scaled up replica set all-in-one-ui-cd554ccb6 to 1 deployment-controller logger.go:42: 07:16:06 | allinone | 2024-09-12 07:15:43 +0000 UTC Normal Pod all-in-one-ui-cd554ccb6-cqwbk AddedInterface Add eth0 [10.130.0.48/23] from ovn-kubernetes logger.go:42: 07:16:06 | allinone | 2024-09-12 07:15:43 +0000 UTC Normal Pod all-in-one-ui-cd554ccb6-cqwbk.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" already present on machine kubelet logger.go:42: 07:16:06 | allinone | 2024-09-12 07:15:43 +0000 UTC Normal Pod all-in-one-ui-cd554ccb6-cqwbk.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:16:06 | allinone | 2024-09-12 07:15:43 +0000 UTC Normal Pod all-in-one-ui-cd554ccb6-cqwbk.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:16:06 | allinone | 2024-09-12 07:15:43 +0000 UTC Normal Pod all-in-one-ui-cd554ccb6-cqwbk.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:16:06 | allinone | 2024-09-12 07:15:43 +0000 UTC Normal Pod all-in-one-ui-cd554ccb6-cqwbk.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:16:06 | allinone | 2024-09-12 07:15:43 +0000 UTC Normal Pod all-in-one-ui-cd554ccb6-cqwbk.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:16:06 | allinone | 2024-09-12 07:15:46 +0000 UTC Normal ReplicaSet.apps all-in-one-ui-cd554ccb6 SuccessfulDelete Deleted pod: all-in-one-ui-cd554ccb6-cqwbk replicaset-controller logger.go:42: 07:16:06 | allinone | 2024-09-12 07:15:46 +0000 UTC Normal Deployment.apps all-in-one-ui ScalingReplicaSet Scaled down replica set all-in-one-ui-cd554ccb6 to 0 from 1 deployment-controller logger.go:42: 07:16:06 | allinone | 2024-09-12 07:15:47 +0000 UTC Normal Pod all-in-one-ui-cd554ccb6-cqwbk.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:16:06 | allinone | 2024-09-12 07:15:47 +0000 UTC Normal Pod all-in-one-ui-cd554ccb6-cqwbk.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:16:06 | allinone | 2024-09-12 07:15:49 +0000 UTC Normal Pod all-in-one-ui-77cffc798f-c9whx Binding Scheduled Successfully assigned kuttl-test-promoted-rat/all-in-one-ui-77cffc798f-c9whx to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:16:06 | allinone | 2024-09-12 07:15:49 +0000 UTC Normal Pod all-in-one-ui-77cffc798f-c9whx AddedInterface Add eth0 [10.130.0.49/23] from ovn-kubernetes logger.go:42: 07:16:06 | allinone | 2024-09-12 07:15:49 +0000 UTC Normal Pod all-in-one-ui-77cffc798f-c9whx.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" already present on machine kubelet logger.go:42: 07:16:06 | allinone | 2024-09-12 07:15:49 +0000 UTC Normal Pod all-in-one-ui-77cffc798f-c9whx.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:16:06 | allinone | 2024-09-12 07:15:49 +0000 UTC Normal Pod all-in-one-ui-77cffc798f-c9whx.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:16:06 | allinone | 2024-09-12 07:15:49 +0000 UTC Normal Pod all-in-one-ui-77cffc798f-c9whx.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:16:06 | allinone | 2024-09-12 07:15:49 +0000 UTC Normal Pod all-in-one-ui-77cffc798f-c9whx.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:16:06 | allinone | 2024-09-12 07:15:49 +0000 UTC Normal Pod all-in-one-ui-77cffc798f-c9whx.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:16:06 | allinone | 2024-09-12 07:15:49 +0000 UTC Normal ReplicaSet.apps all-in-one-ui-77cffc798f SuccessfulCreate Created pod: all-in-one-ui-77cffc798f-c9whx replicaset-controller logger.go:42: 07:16:06 | allinone | 2024-09-12 07:15:49 +0000 UTC Normal Deployment.apps all-in-one-ui ScalingReplicaSet Scaled up replica set all-in-one-ui-77cffc798f to 1 deployment-controller logger.go:42: 07:16:06 | allinone | 2024-09-12 07:15:53 +0000 UTC Normal Pod all-in-one-ui-77cffc798f-c9whx.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:16:06 | allinone | 2024-09-12 07:15:53 +0000 UTC Normal Pod all-in-one-ui-77cffc798f-c9whx.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:16:06 | allinone | 2024-09-12 07:15:54 +0000 UTC Normal Pod all-in-one-ui-776c8b47b6-dj7sg Binding Scheduled Successfully assigned kuttl-test-promoted-rat/all-in-one-ui-776c8b47b6-dj7sg to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:16:06 | allinone | 2024-09-12 07:15:54 +0000 UTC Normal ReplicaSet.apps all-in-one-ui-776c8b47b6 SuccessfulCreate Created pod: all-in-one-ui-776c8b47b6-dj7sg replicaset-controller logger.go:42: 07:16:06 | allinone | 2024-09-12 07:15:54 +0000 UTC Normal Deployment.apps all-in-one-ui ScalingReplicaSet Scaled up replica set all-in-one-ui-776c8b47b6 to 1 deployment-controller logger.go:42: 07:16:06 | allinone | 2024-09-12 07:15:55 +0000 UTC Normal Pod all-in-one-ui-776c8b47b6-dj7sg AddedInterface Add eth0 [10.130.0.50/23] from ovn-kubernetes logger.go:42: 07:16:06 | allinone | 2024-09-12 07:15:55 +0000 UTC Normal Pod all-in-one-ui-776c8b47b6-dj7sg.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" already present on machine kubelet logger.go:42: 07:16:06 | allinone | 2024-09-12 07:15:55 +0000 UTC Normal Pod all-in-one-ui-776c8b47b6-dj7sg.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:16:06 | allinone | 2024-09-12 07:15:55 +0000 UTC Normal Pod all-in-one-ui-776c8b47b6-dj7sg.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:16:06 | allinone | 2024-09-12 07:15:55 +0000 UTC Normal Pod all-in-one-ui-776c8b47b6-dj7sg.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:16:06 | allinone | 2024-09-12 07:15:55 +0000 UTC Normal Pod all-in-one-ui-776c8b47b6-dj7sg.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:16:06 | allinone | 2024-09-12 07:15:55 +0000 UTC Normal Pod all-in-one-ui-776c8b47b6-dj7sg.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:16:06 | allinone | 2024-09-12 07:15:55 +0000 UTC Normal Pod all-in-one-ui-776c8b47b6-dj7sg.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:16:06 | allinone | 2024-09-12 07:15:55 +0000 UTC Normal Pod all-in-one-ui-776c8b47b6-dj7sg.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:16:06 | allinone | 2024-09-12 07:15:55 +0000 UTC Normal ReplicaSet.apps all-in-one-ui-776c8b47b6 SuccessfulDelete Deleted pod: all-in-one-ui-776c8b47b6-dj7sg replicaset-controller logger.go:42: 07:16:06 | allinone | 2024-09-12 07:15:55 +0000 UTC Normal Deployment.apps all-in-one-ui ScalingReplicaSet Scaled down replica set all-in-one-ui-776c8b47b6 to 0 from 1 deployment-controller logger.go:42: 07:16:06 | allinone | 2024-09-12 07:15:57 +0000 UTC Normal Pod all-in-one-ui-7d888f75df-gjmrv Binding Scheduled Successfully assigned kuttl-test-promoted-rat/all-in-one-ui-7d888f75df-gjmrv to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:16:06 | allinone | 2024-09-12 07:15:57 +0000 UTC Normal Pod all-in-one-ui-7d888f75df-gjmrv AddedInterface Add eth0 [10.130.0.51/23] from ovn-kubernetes logger.go:42: 07:16:06 | allinone | 2024-09-12 07:15:57 +0000 UTC Normal Pod all-in-one-ui-7d888f75df-gjmrv.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" already present on machine kubelet logger.go:42: 07:16:06 | allinone | 2024-09-12 07:15:57 +0000 UTC Normal Pod all-in-one-ui-7d888f75df-gjmrv.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:16:06 | allinone | 2024-09-12 07:15:57 +0000 UTC Normal Pod all-in-one-ui-7d888f75df-gjmrv.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:16:06 | allinone | 2024-09-12 07:15:57 +0000 UTC Normal ReplicaSet.apps all-in-one-ui-7d888f75df SuccessfulCreate Created pod: all-in-one-ui-7d888f75df-gjmrv replicaset-controller logger.go:42: 07:16:06 | allinone | 2024-09-12 07:15:57 +0000 UTC Normal Deployment.apps all-in-one-ui ScalingReplicaSet Scaled up replica set all-in-one-ui-7d888f75df to 1 deployment-controller logger.go:42: 07:16:06 | allinone | Deleting namespace: kuttl-test-promoted-rat === CONT kuttl/harness/production logger.go:42: 07:16:13 | production | Ignoring add-tracking-id.yaml as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:16:13 | production | Ignoring ensure-ingress-host.sh as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:16:13 | production | Creating namespace: kuttl-test-dashing-elephant logger.go:42: 07:16:13 | production/1-install | starting test step 1-install logger.go:42: 07:16:13 | production/1-install | Jaeger:kuttl-test-dashing-elephant/production-ui created logger.go:42: 07:16:47 | production/1-install | test step completed 1-install logger.go:42: 07:16:47 | production/2-check-forbbiden-access | starting test step 2-check-forbbiden-access logger.go:42: 07:16:47 | production/2-check-forbbiden-access | running command: [./ensure-ingress-host.sh] logger.go:42: 07:16:47 | production/2-check-forbbiden-access | Checking the Ingress host value was populated logger.go:42: 07:16:47 | production/2-check-forbbiden-access | Try number 0 logger.go:42: 07:16:48 | production/2-check-forbbiden-access | Hostname is production-ui-kuttl-test-dashing-elephant.apps.rosa.ci-rosa-h-u0tm.f61o.s3.devshift.org logger.go:42: 07:16:48 | production/2-check-forbbiden-access | running command: [sh -c INSECURE=true ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 403 true $NAMESPACE production-ui] logger.go:42: 07:16:48 | production/2-check-forbbiden-access | Checking an expected HTTP response logger.go:42: 07:16:48 | production/2-check-forbbiden-access | Running in OpenShift logger.go:42: 07:16:48 | production/2-check-forbbiden-access | Not using any secret logger.go:42: 07:16:48 | production/2-check-forbbiden-access | Try number 1/30 the https://production-ui-kuttl-test-dashing-elephant.apps.rosa.ci-rosa-h-u0tm.f61o.s3.devshift.org/search logger.go:42: 07:16:48 | production/2-check-forbbiden-access | HTTP response is 503. 403 expected. Waiting 10 s logger.go:42: 07:16:58 | production/2-check-forbbiden-access | Try number 2/30 the https://production-ui-kuttl-test-dashing-elephant.apps.rosa.ci-rosa-h-u0tm.f61o.s3.devshift.org/search logger.go:42: 07:16:58 | production/2-check-forbbiden-access | curl response asserted properly logger.go:42: 07:16:59 | production/2-check-forbbiden-access | test step completed 2-check-forbbiden-access logger.go:42: 07:16:59 | production/3-curl | starting test step 3-curl logger.go:42: 07:16:59 | production/3-curl | running command: [./ensure-ingress-host.sh] logger.go:42: 07:16:59 | production/3-curl | Checking the Ingress host value was populated logger.go:42: 07:16:59 | production/3-curl | Try number 0 logger.go:42: 07:16:59 | production/3-curl | Hostname is production-ui-kuttl-test-dashing-elephant.apps.rosa.ci-rosa-h-u0tm.f61o.s3.devshift.org logger.go:42: 07:16:59 | production/3-curl | running command: [sh -c ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 200 true $NAMESPACE production-ui] logger.go:42: 07:16:59 | production/3-curl | Checking an expected HTTP response logger.go:42: 07:16:59 | production/3-curl | Running in OpenShift logger.go:42: 07:16:59 | production/3-curl | User not provided. Getting the token... logger.go:42: 07:17:01 | production/3-curl | Warning: resource jaegers/production-ui is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:17:08 | production/3-curl | Try number 1/30 the https://production-ui-kuttl-test-dashing-elephant.apps.rosa.ci-rosa-h-u0tm.f61o.s3.devshift.org/search logger.go:42: 07:17:08 | production/3-curl | curl response asserted properly logger.go:42: 07:17:08 | production/3-curl | test step completed 3-curl logger.go:42: 07:17:08 | production/4-install | starting test step 4-install logger.go:42: 07:17:09 | production/4-install | Jaeger:kuttl-test-dashing-elephant/production-ui updated logger.go:42: 07:17:09 | production/4-install | test step completed 4-install logger.go:42: 07:17:09 | production/5-check-disabled-security | starting test step 5-check-disabled-security logger.go:42: 07:17:09 | production/5-check-disabled-security | running command: [./ensure-ingress-host.sh] logger.go:42: 07:17:09 | production/5-check-disabled-security | Checking the Ingress host value was populated logger.go:42: 07:17:09 | production/5-check-disabled-security | Try number 0 logger.go:42: 07:17:09 | production/5-check-disabled-security | Hostname is production-ui-kuttl-test-dashing-elephant.apps.rosa.ci-rosa-h-u0tm.f61o.s3.devshift.org logger.go:42: 07:17:09 | production/5-check-disabled-security | running command: [sh -c INSECURE=true ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 200 true $NAMESPACE production-ui] logger.go:42: 07:17:09 | production/5-check-disabled-security | Checking an expected HTTP response logger.go:42: 07:17:09 | production/5-check-disabled-security | Running in OpenShift logger.go:42: 07:17:09 | production/5-check-disabled-security | Not using any secret logger.go:42: 07:17:09 | production/5-check-disabled-security | Try number 1/30 the https://production-ui-kuttl-test-dashing-elephant.apps.rosa.ci-rosa-h-u0tm.f61o.s3.devshift.org/search logger.go:42: 07:17:25 | production/5-check-disabled-security | HTTP response is 503. 200 expected. Waiting 10 s logger.go:42: 07:17:35 | production/5-check-disabled-security | Try number 2/30 the https://production-ui-kuttl-test-dashing-elephant.apps.rosa.ci-rosa-h-u0tm.f61o.s3.devshift.org/search logger.go:42: 07:17:36 | production/5-check-disabled-security | curl response asserted properly logger.go:42: 07:17:36 | production/5-check-disabled-security | test step completed 5-check-disabled-security logger.go:42: 07:17:36 | production/6-check-NO-gaID | starting test step 6-check-NO-gaID logger.go:42: 07:17:36 | production/6-check-NO-gaID | running command: [./ensure-ingress-host.sh] logger.go:42: 07:17:36 | production/6-check-NO-gaID | Checking the Ingress host value was populated logger.go:42: 07:17:36 | production/6-check-NO-gaID | Try number 0 logger.go:42: 07:17:36 | production/6-check-NO-gaID | Hostname is production-ui-kuttl-test-dashing-elephant.apps.rosa.ci-rosa-h-u0tm.f61o.s3.devshift.org logger.go:42: 07:17:36 | production/6-check-NO-gaID | running command: [sh -c ASSERT_PRESENT=false EXPECTED_CONTENT=MyTrackingId QUERY_HOSTNAME=https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search go run ../../../../cmd-utils/uiconfig/main.go] logger.go:42: 07:17:36 | production/6-check-NO-gaID | time="2024-09-12T07:17:36Z" level=info msg="Querying https://production-ui-kuttl-test-dashing-elephant.apps.rosa.ci-rosa-h-u0tm.f61o.s3.devshift.org/search..." logger.go:42: 07:17:36 | production/6-check-NO-gaID | time="2024-09-12T07:17:36Z" level=info msg="No secret provided for the Authorization header" logger.go:42: 07:17:36 | production/6-check-NO-gaID | time="2024-09-12T07:17:36Z" level=info msg="Polling to https://production-ui-kuttl-test-dashing-elephant.apps.rosa.ci-rosa-h-u0tm.f61o.s3.devshift.org/search" logger.go:42: 07:17:36 | production/6-check-NO-gaID | time="2024-09-12T07:17:36Z" level=info msg="Doing request number 0" logger.go:42: 07:17:37 | production/6-check-NO-gaID | time="2024-09-12T07:17:37Z" level=info msg="Content not found and asserted it was not found!" logger.go:42: 07:17:37 | production/6-check-NO-gaID | time="2024-09-12T07:17:37Z" level=info msg="Success!" logger.go:42: 07:17:37 | production/6-check-NO-gaID | test step completed 6-check-NO-gaID logger.go:42: 07:17:37 | production/7-add-tracking-id | starting test step 7-add-tracking-id logger.go:42: 07:17:37 | production/7-add-tracking-id | running command: [sh -c kubectl apply -f add-tracking-id.yaml -n $NAMESPACE] logger.go:42: 07:17:38 | production/7-add-tracking-id | jaeger.jaegertracing.io/production-ui configured logger.go:42: 07:17:38 | production/7-add-tracking-id | test step completed 7-add-tracking-id logger.go:42: 07:17:38 | production/8-check-gaID | starting test step 8-check-gaID logger.go:42: 07:17:38 | production/8-check-gaID | running command: [./ensure-ingress-host.sh] logger.go:42: 07:17:38 | production/8-check-gaID | Checking the Ingress host value was populated logger.go:42: 07:17:38 | production/8-check-gaID | Try number 0 logger.go:42: 07:17:38 | production/8-check-gaID | Hostname is production-ui-kuttl-test-dashing-elephant.apps.rosa.ci-rosa-h-u0tm.f61o.s3.devshift.org logger.go:42: 07:17:38 | production/8-check-gaID | running command: [sh -c ASSERT_PRESENT=true EXPECTED_CONTENT=MyTrackingId QUERY_HOSTNAME=https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search go run ../../../../cmd-utils/uiconfig/main.go] logger.go:42: 07:17:38 | production/8-check-gaID | time="2024-09-12T07:17:38Z" level=info msg="Querying https://production-ui-kuttl-test-dashing-elephant.apps.rosa.ci-rosa-h-u0tm.f61o.s3.devshift.org/search..." logger.go:42: 07:17:38 | production/8-check-gaID | time="2024-09-12T07:17:38Z" level=info msg="No secret provided for the Authorization header" logger.go:42: 07:17:38 | production/8-check-gaID | time="2024-09-12T07:17:38Z" level=info msg="Polling to https://production-ui-kuttl-test-dashing-elephant.apps.rosa.ci-rosa-h-u0tm.f61o.s3.devshift.org/search" logger.go:42: 07:17:38 | production/8-check-gaID | time="2024-09-12T07:17:38Z" level=info msg="Doing request number 0" logger.go:42: 07:17:39 | production/8-check-gaID | time="2024-09-12T07:17:39Z" level=warning msg="Status code: 503" logger.go:42: 07:17:47 | production/8-check-gaID | time="2024-09-12T07:17:47Z" level=info msg="Doing request number 1" logger.go:42: 07:17:47 | production/8-check-gaID | time="2024-09-12T07:17:47Z" level=info msg="Content found and asserted!" logger.go:42: 07:17:47 | production/8-check-gaID | time="2024-09-12T07:17:47Z" level=info msg="Success!" logger.go:42: 07:17:47 | production/8-check-gaID | test step completed 8-check-gaID logger.go:42: 07:17:47 | production | production events from ns kuttl-test-dashing-elephant: logger.go:42: 07:17:47 | production | 2024-09-12 07:16:17 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestdashingelephantproductionui-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestdashingelephantproductionui-1-c86878fc5 to 1 deployment-controller logger.go:42: 07:17:47 | production | 2024-09-12 07:16:18 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestdashingelephantproductionui-1-c86878fc5 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestdashingelephantproductionui-1-c8gzf4 replicaset-controller logger.go:42: 07:17:47 | production | 2024-09-12 07:16:18 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestdashingelephantproductionui-1-c8gzf4 Binding Scheduled Successfully assigned kuttl-test-dashing-elephant/elasticsearch-cdm-kuttltestdashingelephantproductionui-1-c8gzf4 to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:17:47 | production | 2024-09-12 07:16:18 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestdashingelephantproductionui-1-c8gzf4 FailedMount MountVolume.SetUp failed for volume "elasticsearch-metrics" : secret "elasticsearch-metrics" not found kubelet logger.go:42: 07:17:47 | production | 2024-09-12 07:16:19 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestdashingelephantproductionui-1-c8gzf4 AddedInterface Add eth0 [10.130.0.52/23] from ovn-kubernetes logger.go:42: 07:17:47 | production | 2024-09-12 07:16:19 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestdashingelephantproductionui-1-c8gzf4.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:bc97b8e13087050dfb1a75b02c5b780cbb8fb12a849a655169d072ba8bbf42b4" already present on machine kubelet logger.go:42: 07:17:47 | production | 2024-09-12 07:16:19 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestdashingelephantproductionui-1-c8gzf4.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:17:47 | production | 2024-09-12 07:16:19 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestdashingelephantproductionui-1-c8gzf4.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:17:47 | production | 2024-09-12 07:16:19 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestdashingelephantproductionui-1-c8gzf4.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:1d1d7d47b616995d18692f3a2d6232d7e1a0f41bc1503d87100f31939e080a78" already present on machine kubelet logger.go:42: 07:17:47 | production | 2024-09-12 07:16:19 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestdashingelephantproductionui-1-c8gzf4.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:17:47 | production | 2024-09-12 07:16:19 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestdashingelephantproductionui-1-c8gzf4.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:17:47 | production | 2024-09-12 07:16:34 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestdashingelephantproductionui-1-c8gzf4.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:17:47 | production | 2024-09-12 07:16:44 +0000 UTC Normal Pod production-ui-collector-559745849d-v9x6t Binding Scheduled Successfully assigned kuttl-test-dashing-elephant/production-ui-collector-559745849d-v9x6t to ip-10-0-54-110.us-west-2.compute.internal default-scheduler logger.go:42: 07:17:47 | production | 2024-09-12 07:16:44 +0000 UTC Normal ReplicaSet.apps production-ui-collector-559745849d SuccessfulCreate Created pod: production-ui-collector-559745849d-v9x6t replicaset-controller logger.go:42: 07:17:47 | production | 2024-09-12 07:16:44 +0000 UTC Normal Deployment.apps production-ui-collector ScalingReplicaSet Scaled up replica set production-ui-collector-559745849d to 1 deployment-controller logger.go:42: 07:17:47 | production | 2024-09-12 07:16:44 +0000 UTC Normal Pod production-ui-query-596789575f-cq2ql Binding Scheduled Successfully assigned kuttl-test-dashing-elephant/production-ui-query-596789575f-cq2ql to ip-10-0-52-42.us-west-2.compute.internal default-scheduler logger.go:42: 07:17:47 | production | 2024-09-12 07:16:44 +0000 UTC Normal ReplicaSet.apps production-ui-query-596789575f SuccessfulCreate Created pod: production-ui-query-596789575f-cq2ql replicaset-controller logger.go:42: 07:17:47 | production | 2024-09-12 07:16:44 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled up replica set production-ui-query-596789575f to 1 deployment-controller logger.go:42: 07:17:47 | production | 2024-09-12 07:16:45 +0000 UTC Normal Pod production-ui-collector-559745849d-v9x6t AddedInterface Add eth0 [10.129.0.42/23] from ovn-kubernetes logger.go:42: 07:17:47 | production | 2024-09-12 07:16:45 +0000 UTC Normal Pod production-ui-collector-559745849d-v9x6t.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:4bab63290ff52e4f6328009f9c8b2c4496b230e9f8a98eac01736a66a291ff6c" already present on machine kubelet logger.go:42: 07:17:47 | production | 2024-09-12 07:16:45 +0000 UTC Normal Pod production-ui-collector-559745849d-v9x6t.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:17:47 | production | 2024-09-12 07:16:45 +0000 UTC Normal Pod production-ui-collector-559745849d-v9x6t.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:17:47 | production | 2024-09-12 07:16:45 +0000 UTC Normal Pod production-ui-query-596789575f-cq2ql AddedInterface Add eth0 [10.128.0.50/23] from ovn-kubernetes logger.go:42: 07:17:47 | production | 2024-09-12 07:16:45 +0000 UTC Normal Pod production-ui-query-596789575f-cq2ql.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a9186dcd910256c0f464b0a3928844a01de166a10c186c97ef4581bf288c23cb" already present on machine kubelet logger.go:42: 07:17:47 | production | 2024-09-12 07:16:45 +0000 UTC Normal Pod production-ui-query-596789575f-cq2ql.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:17:47 | production | 2024-09-12 07:16:45 +0000 UTC Normal Pod production-ui-query-596789575f-cq2ql.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:17:47 | production | 2024-09-12 07:16:45 +0000 UTC Normal Pod production-ui-query-596789575f-cq2ql.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:17:47 | production | 2024-09-12 07:16:45 +0000 UTC Normal Pod production-ui-query-596789575f-cq2ql.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:17:47 | production | 2024-09-12 07:16:45 +0000 UTC Normal Pod production-ui-query-596789575f-cq2ql.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:17:47 | production | 2024-09-12 07:16:45 +0000 UTC Normal Pod production-ui-query-596789575f-cq2ql.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" already present on machine kubelet logger.go:42: 07:17:47 | production | 2024-09-12 07:16:45 +0000 UTC Normal Pod production-ui-query-596789575f-cq2ql.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:17:47 | production | 2024-09-12 07:16:45 +0000 UTC Normal Pod production-ui-query-596789575f-cq2ql.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:17:47 | production | 2024-09-12 07:17:00 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling production-ui-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:17:47 | production | 2024-09-12 07:17:00 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling production-ui-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:17:47 | production | 2024-09-12 07:17:00 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling production-ui-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:17:47 | production | 2024-09-12 07:17:02 +0000 UTC Normal Pod production-ui-query-596789575f-cq2ql.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:17:47 | production | 2024-09-12 07:17:02 +0000 UTC Normal Pod production-ui-query-596789575f-cq2ql.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:17:47 | production | 2024-09-12 07:17:02 +0000 UTC Normal Pod production-ui-query-596789575f-cq2ql.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:17:47 | production | 2024-09-12 07:17:02 +0000 UTC Normal ReplicaSet.apps production-ui-query-596789575f SuccessfulDelete Deleted pod: production-ui-query-596789575f-cq2ql replicaset-controller logger.go:42: 07:17:47 | production | 2024-09-12 07:17:02 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled down replica set production-ui-query-596789575f to 0 from 1 deployment-controller logger.go:42: 07:17:47 | production | 2024-09-12 07:17:03 +0000 UTC Normal Pod production-ui-query-67db7df5c7-js6g2 Binding Scheduled Successfully assigned kuttl-test-dashing-elephant/production-ui-query-67db7df5c7-js6g2 to ip-10-0-52-42.us-west-2.compute.internal default-scheduler logger.go:42: 07:17:47 | production | 2024-09-12 07:17:03 +0000 UTC Normal Pod production-ui-query-67db7df5c7-js6g2 AddedInterface Add eth0 [10.128.0.51/23] from ovn-kubernetes logger.go:42: 07:17:47 | production | 2024-09-12 07:17:03 +0000 UTC Normal Pod production-ui-query-67db7df5c7-js6g2.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a9186dcd910256c0f464b0a3928844a01de166a10c186c97ef4581bf288c23cb" already present on machine kubelet logger.go:42: 07:17:47 | production | 2024-09-12 07:17:03 +0000 UTC Normal ReplicaSet.apps production-ui-query-67db7df5c7 SuccessfulCreate Created pod: production-ui-query-67db7df5c7-js6g2 replicaset-controller logger.go:42: 07:17:47 | production | 2024-09-12 07:17:03 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled up replica set production-ui-query-67db7df5c7 to 1 deployment-controller logger.go:42: 07:17:47 | production | 2024-09-12 07:17:04 +0000 UTC Normal Pod production-ui-query-67db7df5c7-js6g2.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:17:47 | production | 2024-09-12 07:17:04 +0000 UTC Normal Pod production-ui-query-67db7df5c7-js6g2.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:17:47 | production | 2024-09-12 07:17:04 +0000 UTC Normal Pod production-ui-query-67db7df5c7-js6g2.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:17:47 | production | 2024-09-12 07:17:04 +0000 UTC Normal Pod production-ui-query-67db7df5c7-js6g2.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:17:47 | production | 2024-09-12 07:17:04 +0000 UTC Normal Pod production-ui-query-67db7df5c7-js6g2.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:17:47 | production | 2024-09-12 07:17:04 +0000 UTC Normal Pod production-ui-query-67db7df5c7-js6g2.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" already present on machine kubelet logger.go:42: 07:17:47 | production | 2024-09-12 07:17:04 +0000 UTC Normal Pod production-ui-query-67db7df5c7-js6g2.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:17:47 | production | 2024-09-12 07:17:04 +0000 UTC Normal Pod production-ui-query-67db7df5c7-js6g2.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:17:47 | production | 2024-09-12 07:17:09 +0000 UTC Normal Pod production-ui-query-67db7df5c7-js6g2.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:17:47 | production | 2024-09-12 07:17:09 +0000 UTC Normal Pod production-ui-query-67db7df5c7-js6g2.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:17:47 | production | 2024-09-12 07:17:09 +0000 UTC Normal Pod production-ui-query-67db7df5c7-js6g2.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:17:47 | production | 2024-09-12 07:17:09 +0000 UTC Normal ReplicaSet.apps production-ui-query-67db7df5c7 SuccessfulDelete Deleted pod: production-ui-query-67db7df5c7-js6g2 replicaset-controller logger.go:42: 07:17:47 | production | 2024-09-12 07:17:09 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled down replica set production-ui-query-67db7df5c7 to 0 from 1 deployment-controller logger.go:42: 07:17:47 | production | 2024-09-12 07:17:10 +0000 UTC Normal Pod production-ui-query-56f47594dc-4tn6q Binding Scheduled Successfully assigned kuttl-test-dashing-elephant/production-ui-query-56f47594dc-4tn6q to ip-10-0-52-42.us-west-2.compute.internal default-scheduler logger.go:42: 07:17:47 | production | 2024-09-12 07:17:10 +0000 UTC Normal Pod production-ui-query-56f47594dc-4tn6q AddedInterface Add eth0 [10.128.0.52/23] from ovn-kubernetes logger.go:42: 07:17:47 | production | 2024-09-12 07:17:10 +0000 UTC Normal Pod production-ui-query-56f47594dc-4tn6q.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a9186dcd910256c0f464b0a3928844a01de166a10c186c97ef4581bf288c23cb" already present on machine kubelet logger.go:42: 07:17:47 | production | 2024-09-12 07:17:10 +0000 UTC Normal ReplicaSet.apps production-ui-query-56f47594dc SuccessfulCreate Created pod: production-ui-query-56f47594dc-4tn6q replicaset-controller logger.go:42: 07:17:47 | production | 2024-09-12 07:17:10 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled up replica set production-ui-query-56f47594dc to 1 deployment-controller logger.go:42: 07:17:47 | production | 2024-09-12 07:17:11 +0000 UTC Normal Pod production-ui-query-56f47594dc-4tn6q.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:17:47 | production | 2024-09-12 07:17:11 +0000 UTC Normal Pod production-ui-query-56f47594dc-4tn6q.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:17:47 | production | 2024-09-12 07:17:11 +0000 UTC Normal Pod production-ui-query-56f47594dc-4tn6q.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" already present on machine kubelet logger.go:42: 07:17:47 | production | 2024-09-12 07:17:11 +0000 UTC Normal Pod production-ui-query-56f47594dc-4tn6q.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:17:47 | production | 2024-09-12 07:17:11 +0000 UTC Normal Pod production-ui-query-56f47594dc-4tn6q.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:17:47 | production | 2024-09-12 07:17:38 +0000 UTC Normal Pod production-ui-query-56f47594dc-4tn6q.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:17:47 | production | 2024-09-12 07:17:38 +0000 UTC Normal Pod production-ui-query-56f47594dc-4tn6q.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:17:47 | production | 2024-09-12 07:17:38 +0000 UTC Normal ReplicaSet.apps production-ui-query-56f47594dc SuccessfulDelete Deleted pod: production-ui-query-56f47594dc-4tn6q replicaset-controller logger.go:42: 07:17:47 | production | 2024-09-12 07:17:38 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled down replica set production-ui-query-56f47594dc to 0 from 1 deployment-controller logger.go:42: 07:17:47 | production | 2024-09-12 07:17:39 +0000 UTC Normal Pod production-ui-query-6dff945976-bhr82 Binding Scheduled Successfully assigned kuttl-test-dashing-elephant/production-ui-query-6dff945976-bhr82 to ip-10-0-52-42.us-west-2.compute.internal default-scheduler logger.go:42: 07:17:47 | production | 2024-09-12 07:17:39 +0000 UTC Normal ReplicaSet.apps production-ui-query-6dff945976 SuccessfulCreate Created pod: production-ui-query-6dff945976-bhr82 replicaset-controller logger.go:42: 07:17:47 | production | 2024-09-12 07:17:39 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled up replica set production-ui-query-6dff945976 to 1 deployment-controller logger.go:42: 07:17:47 | production | 2024-09-12 07:17:40 +0000 UTC Normal Pod production-ui-query-6dff945976-bhr82 AddedInterface Add eth0 [10.128.0.53/23] from ovn-kubernetes logger.go:42: 07:17:47 | production | 2024-09-12 07:17:40 +0000 UTC Normal Pod production-ui-query-6dff945976-bhr82.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a9186dcd910256c0f464b0a3928844a01de166a10c186c97ef4581bf288c23cb" already present on machine kubelet logger.go:42: 07:17:47 | production | 2024-09-12 07:17:40 +0000 UTC Normal Pod production-ui-query-6dff945976-bhr82.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:17:47 | production | 2024-09-12 07:17:40 +0000 UTC Normal Pod production-ui-query-6dff945976-bhr82.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:17:47 | production | 2024-09-12 07:17:40 +0000 UTC Normal Pod production-ui-query-6dff945976-bhr82.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" already present on machine kubelet logger.go:42: 07:17:47 | production | 2024-09-12 07:17:40 +0000 UTC Normal Pod production-ui-query-6dff945976-bhr82.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:17:47 | production | 2024-09-12 07:17:40 +0000 UTC Normal Pod production-ui-query-6dff945976-bhr82.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:17:47 | production | 2024-09-12 07:17:45 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling production-ui-collector FailedGetResourceMetric failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 07:17:47 | production | 2024-09-12 07:17:45 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling production-ui-collector FailedGetResourceMetric failed to get memory utilization: missing request for memory in container jaeger-collector of Pod production-ui-collector-559745849d-v9x6t horizontal-pod-autoscaler logger.go:42: 07:17:47 | production | 2024-09-12 07:17:45 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling production-ui-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 07:17:47 | production | Deleting namespace: kuttl-test-dashing-elephant === CONT kuttl/harness/artifacts logger.go:42: 07:17:54 | artifacts | Creating namespace: kuttl-test-current-starling logger.go:42: 07:17:54 | artifacts | artifacts events from ns kuttl-test-current-starling: logger.go:42: 07:17:54 | artifacts | Deleting namespace: kuttl-test-current-starling === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (138.86s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/allinone (30.94s) --- PASS: kuttl/harness/production (101.06s) --- PASS: kuttl/harness/artifacts (6.09s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name ui --report --output /logs/artifacts/ui.xml ./artifacts/kuttl-report.xml time="2024-09-12T07:18:00Z" level=debug msg="Setting a new name for the test suites" time="2024-09-12T07:18:00Z" level=debug msg="Removing 'artifacts' TestCase" time="2024-09-12T07:18:00Z" level=debug msg="normalizing test case names" time="2024-09-12T07:18:00Z" level=debug msg="ui/allinone -> ui_allinone" time="2024-09-12T07:18:00Z" level=debug msg="ui/production -> ui_production" time="2024-09-12T07:18:00Z" level=debug msg="ui/artifacts -> ui_artifacts" +---------------+--------+ | NAME | RESULT | +---------------+--------+ | ui_allinone | passed | | ui_production | passed | | ui_artifacts | passed | +---------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh elasticsearch false true + '[' 3 -ne 3 ']' + test_suite_name=elasticsearch + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/elasticsearch.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-elasticsearch make[2]: Entering directory '/tmp/jaeger-tests' >>>> Elasticsearch image not loaded because SKIP_ES_EXTERNAL is true SKIP_ES_EXTERNAL=true \ KAFKA_VERSION=3.6.0 \ SKIP_KAFKA=false \ ./tests/e2e/elasticsearch/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.29 True False 29m Cluster version is 4.15.29' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.29 True False 29m Cluster version is 4.15.29' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/elasticsearch/render.sh ++ export SUITE_DIR=./tests/e2e/elasticsearch ++ SUITE_DIR=./tests/e2e/elasticsearch ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/elasticsearch ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + is_secured=false + '[' true = true ']' + is_secured=true + start_test es-from-aio-to-production + '[' 1 -ne 1 ']' + test_name=es-from-aio-to-production + echo =========================================================================== =========================================================================== + info 'Rendering files for test es-from-aio-to-production' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test es-from-aio-to-production\e[0m' Rendering files for test es-from-aio-to-production + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build + '[' _build '!=' _build ']' + mkdir -p es-from-aio-to-production + cd es-from-aio-to-production + jaeger_name=my-jaeger + render_install_jaeger my-jaeger allInOne 00 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=allInOne + test_step=00 + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test my-jaeger true 01 + '[' 3 -ne 3 ']' + jaeger=my-jaeger + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + jaeger_deploy_mode=production + [[ true = true ]] + [[ true = true ]] + jaeger_deploy_mode=production_autoprovisioned + render_install_jaeger my-jaeger production_autoprovisioned 03 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=production_autoprovisioned + test_step=03 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./03-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./03-assert.yaml + [[ true = true ]] + [[ true = true ]] + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.elasticsearch.redundancyPolicy="ZeroRedundancy"' ./03-install.yaml + render_smoke_test my-jaeger true 04 + '[' 3 -ne 3 ']' + jaeger=my-jaeger + is_secured=true + test_step=04 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./04-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./04-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test es-increasing-replicas + '[' 1 -ne 1 ']' + test_name=es-increasing-replicas + echo =========================================================================== =========================================================================== + info 'Rendering files for test es-increasing-replicas' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test es-increasing-replicas\e[0m' Rendering files for test es-increasing-replicas + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build/es-from-aio-to-production + '[' es-from-aio-to-production '!=' _build ']' + cd .. + mkdir -p es-increasing-replicas + cd es-increasing-replicas + jaeger_name=simple-prod + '[' true = true ']' + jaeger_deployment_mode=production_autoprovisioned + render_install_jaeger simple-prod production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=simple-prod + JAEGER_NAME=simple-prod + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + cp ./01-install.yaml ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.collector.replicas=2 ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.query.replicas=2 ./02-install.yaml + cp ./01-assert.yaml ./02-assert.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.replicas=2 ./02-assert.yaml + /tmp/jaeger-tests/bin/yq e -i .status.readyReplicas=2 ./02-assert.yaml + render_smoke_test simple-prod true 03 + '[' 3 -ne 3 ']' + jaeger=simple-prod + is_secured=true + test_step=03 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + export JAEGER_NAME=simple-prod + JAEGER_NAME=simple-prod + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./03-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./03-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + '[' true = true ']' + cp ./02-install.yaml ./04-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.storage.elasticsearch.nodeCount=2 ./04-install.yaml + /tmp/jaeger-tests/bin/gomplate -f ./openshift-check-es-nodes.yaml.template -o ./05-check-es-nodes.yaml + '[' true = true ']' + skip_test es-index-cleaner-upstream 'SKIP_ES_EXTERNAL is true' + '[' 2 -ne 2 ']' + test_name=es-index-cleaner-upstream + message='SKIP_ES_EXTERNAL is true' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build/es-increasing-replicas + '[' es-increasing-replicas '!=' _build ']' + cd .. + rm -rf es-index-cleaner-upstream + warning 'es-index-cleaner-upstream: SKIP_ES_EXTERNAL is true' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: es-index-cleaner-upstream: SKIP_ES_EXTERNAL is true\e[0m' WAR: es-index-cleaner-upstream: SKIP_ES_EXTERNAL is true + '[' true = true ']' + es_index_cleaner -autoprov production_autoprovisioned + '[' 2 -ne 2 ']' + postfix=-autoprov + jaeger_deployment_strategy=production_autoprovisioned + start_test es-index-cleaner-autoprov + '[' 1 -ne 1 ']' + test_name=es-index-cleaner-autoprov + echo =========================================================================== =========================================================================== + info 'Rendering files for test es-index-cleaner-autoprov' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test es-index-cleaner-autoprov\e[0m' Rendering files for test es-index-cleaner-autoprov + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build + '[' _build '!=' _build ']' + mkdir -p es-index-cleaner-autoprov + cd es-index-cleaner-autoprov + jaeger_name=test-es-index-cleaner-with-prefix + cronjob_name=test-es-index-cleaner-with-prefix-es-index-cleaner + secured_es_connection=false + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_managed_es ']' + ELASTICSEARCH_URL=https://elasticsearch + secured_es_connection=true + cp ../../es-index-cleaner-upstream/04-assert.yaml ../../es-index-cleaner-upstream/README.md . + render_install_jaeger test-es-index-cleaner-with-prefix production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=test-es-index-cleaner-with-prefix + JAEGER_NAME=test-es-index-cleaner-with-prefix + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.options.es.index-prefix=""' ./01-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.storage.esIndexCleaner.enabled=false ./01-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.storage.esIndexCleaner.numberOfDays=0 ./01-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.esIndexCleaner.schedule="*/1 * * * *"' ./01-install.yaml + render_report_spans test-es-index-cleaner-with-prefix true 5 00 true 02 + '[' 6 -ne 6 ']' + jaeger=test-es-index-cleaner-with-prefix + is_secured=true + number_of_spans=5 + job_number=00 + ensure_reported_spans=true + test_step=02 + export JAEGER_NAME=test-es-index-cleaner-with-prefix + JAEGER_NAME=test-es-index-cleaner-with-prefix + export JAEGER_COLLECTOR_ENDPOINT=http://test-es-index-cleaner-with-prefix-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://test-es-index-cleaner-with-prefix-collector-headless:14268 + export JOB_NUMBER=00 + JOB_NUMBER=00 + export DAYS=5 + DAYS=5 + '[' true = true ']' + protocol=https:// + query_port= + template=/tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template + '[' true = true ']' + export ENSURE_REPORTED_SPANS=true + ENSURE_REPORTED_SPANS=true + export JAEGER_QUERY_ENDPOINT=https://test-es-index-cleaner-with-prefix-query + JAEGER_QUERY_ENDPOINT=https://test-es-index-cleaner-with-prefix-query + params= + '[' true = true ']' + '[' true = true ']' + '[' '' '!=' allInOne ']' + params='-t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template -t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template -o ./02-report-spans.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-report-spans.yaml.template -o ./02-assert.yaml + unset JAEGER_COLLECTOR_ENDPOINT + unset JAEGER_QUERY_ENDPOINT + unset JOB_NUMBER + unset DAYS + unset ENSURE_REPORTED_SPANS + sed 's~enabled: false~enabled: true~gi' ./01-install.yaml + CRONJOB_NAME=test-es-index-cleaner-with-prefix-es-index-cleaner + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/wait-for-cronjob-execution.yaml.template -o ./04-wait-es-index-cleaner.yaml + /tmp/jaeger-tests/bin/gomplate -f ./01-install.yaml -o ./05-install.yaml + render_check_indices true ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' 00 06 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + job_number=00 + test_step=06 + escape_command ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + '[' 1 -ne 1 ']' + command=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' ++ echo ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=test-es-index-cleaner-with-prefix-curator + JOB_NUMBER=00 + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + MOUNT_SECRET=test-es-index-cleaner-with-prefix-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./06-check-indices.yaml + JOB_NUMBER=00 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./06-assert.yaml + '[' true = true ']' + get_elasticsearch_openshift_operator_version + export ESO_OPERATOR_VERSION + '[' true = true ']' ++ kubectl get pods -l name=elasticsearch-operator --all-namespaces '-o=jsonpath={.items[0].metadata.annotations.operatorframework\.io/properties}' + properties='{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.16},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.12"}}]}' + '[' -z '{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.16},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.12"}}]}' ']' ++ echo '{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.16},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.12"}}]}' ++ /tmp/jaeger-tests/bin/yq e -P '.properties.[] | select(.value.packageName == "elasticsearch-operator") | .value.version' + ESO_OPERATOR_VERSION=5.8.12 ++ version_ge 5.8.12 5.4 +++ echo 5.8.12 5.4 +++ tr ' ' '\n' +++ sort -rV +++ head -n 1 ++ test 5.8.12 == 5.8.12 + '[' -n '' ']' + skip_test es-index-cleaner-managed 'Test only supported with Elasticsearch OpenShift Operator >= 5.4' + '[' 2 -ne 2 ']' + test_name=es-index-cleaner-managed + message='Test only supported with Elasticsearch OpenShift Operator >= 5.4' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build/es-index-cleaner-autoprov + '[' es-index-cleaner-autoprov '!=' _build ']' + cd .. + rm -rf es-index-cleaner-managed + warning 'es-index-cleaner-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: es-index-cleaner-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4\e[0m' WAR: es-index-cleaner-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4 + '[' true = true ']' + start_test es-multiinstance + '[' 1 -ne 1 ']' + test_name=es-multiinstance + echo =========================================================================== =========================================================================== + info 'Rendering files for test es-multiinstance' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test es-multiinstance\e[0m' Rendering files for test es-multiinstance + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build + '[' _build '!=' _build ']' + mkdir -p es-multiinstance + cd es-multiinstance + jaeger_name=instance-1 + render_install_jaeger instance-1 production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=instance-1 + JAEGER_NAME=instance-1 + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + /tmp/jaeger-tests/bin/gomplate -f ./03-create-second-instance.yaml.template -o 03-create-second-instance.yaml + '[' true = true ']' + skip_test es-rollover-upstream 'SKIP_ES_EXTERNAL is true' + '[' 2 -ne 2 ']' + test_name=es-rollover-upstream + message='SKIP_ES_EXTERNAL is true' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build/es-multiinstance + '[' es-multiinstance '!=' _build ']' + cd .. + rm -rf es-rollover-upstream + warning 'es-rollover-upstream: SKIP_ES_EXTERNAL is true' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: es-rollover-upstream: SKIP_ES_EXTERNAL is true\e[0m' WAR: es-rollover-upstream: SKIP_ES_EXTERNAL is true + '[' true = true ']' + es_rollover -autoprov production_autoprovisioned + '[' 2 -ne 2 ']' + postfix=-autoprov + jaeger_deployment_strategy=production_autoprovisioned + start_test es-rollover-autoprov + '[' 1 -ne 1 ']' + test_name=es-rollover-autoprov + echo =========================================================================== =========================================================================== + info 'Rendering files for test es-rollover-autoprov' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test es-rollover-autoprov\e[0m' Rendering files for test es-rollover-autoprov + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build + '[' _build '!=' _build ']' + mkdir -p es-rollover-autoprov + cd es-rollover-autoprov + cp ../../es-rollover-upstream/05-assert.yaml ../../es-rollover-upstream/05-install.yaml ../../es-rollover-upstream/README.md . + jaeger_name=my-jaeger + secured_es_connection=false + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_managed_es ']' + ELASTICSEARCH_URL=https://elasticsearch + secured_es_connection=true + render_install_jaeger my-jaeger production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + render_report_spans my-jaeger true 2 00 true 02 + '[' 6 -ne 6 ']' + jaeger=my-jaeger + is_secured=true + number_of_spans=2 + job_number=00 + ensure_reported_spans=true + test_step=02 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JOB_NUMBER=00 + JOB_NUMBER=00 + export DAYS=2 + DAYS=2 + '[' true = true ']' + protocol=https:// + query_port= + template=/tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template + '[' true = true ']' + export ENSURE_REPORTED_SPANS=true + ENSURE_REPORTED_SPANS=true + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + params= + '[' true = true ']' + '[' true = true ']' + '[' '' '!=' allInOne ']' + params='-t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template -t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template -o ./02-report-spans.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-report-spans.yaml.template -o ./02-assert.yaml + unset JAEGER_COLLECTOR_ENDPOINT + unset JAEGER_QUERY_ENDPOINT + unset JOB_NUMBER + unset DAYS + unset ENSURE_REPORTED_SPANS + render_check_indices true ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' 00 03 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + job_number=00 + test_step=03 + escape_command ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + '[' 1 -ne 1 ']' + command=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' ++ echo ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-exist'\'',' + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-exist'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=00 + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./03-check-indices.yaml + JOB_NUMBER=00 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./03-assert.yaml + render_check_indices true ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' 01 04 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + job_number=01 + test_step=04 + escape_command ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + '[' 1 -ne 1 ']' + command=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' ++ echo ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--pattern'\'', '\''jaeger-span-\\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=01 + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./04-check-indices.yaml + JOB_NUMBER=01 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./04-assert.yaml + render_report_spans my-jaeger true 2 02 true 06 + '[' 6 -ne 6 ']' + jaeger=my-jaeger + is_secured=true + number_of_spans=2 + job_number=02 + ensure_reported_spans=true + test_step=06 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JOB_NUMBER=02 + JOB_NUMBER=02 + export DAYS=2 + DAYS=2 + '[' true = true ']' + protocol=https:// + query_port= + template=/tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template + '[' true = true ']' + export ENSURE_REPORTED_SPANS=true + ENSURE_REPORTED_SPANS=true + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + params= + '[' true = true ']' + '[' true = true ']' + '[' '' '!=' allInOne ']' + params='-t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template -t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template -o ./06-report-spans.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-report-spans.yaml.template -o ./06-assert.yaml + unset JAEGER_COLLECTOR_ENDPOINT + unset JAEGER_QUERY_ENDPOINT + unset JOB_NUMBER + unset DAYS + unset ENSURE_REPORTED_SPANS + render_check_indices true ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' 02 07 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + job_number=02 + test_step=07 + escape_command ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + '[' 1 -ne 1 ']' + command=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' ++ echo ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-exist'\'',' + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-exist'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=02 + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./07-check-indices.yaml + JOB_NUMBER=02 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./07-assert.yaml + render_check_indices true ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' 03 08 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' + job_number=03 + test_step=08 + escape_command ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' + '[' 1 -ne 1 ']' + command=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' ++ echo ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--pattern'\'', '\''jaeger-span-\\d{6}'\'', '\''--assert-exist'\'',' + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\\d{6}'\'', '\''--assert-exist'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=03 + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./08-check-indices.yaml + JOB_NUMBER=03 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./08-assert.yaml + render_check_indices true ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' 04 09 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' + job_number=04 + test_step=09 + escape_command ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' + '[' 1 -ne 1 ']' + command=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' ++ echo ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=04 + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./09-check-indices.yaml + JOB_NUMBER=04 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./09-assert.yaml + render_report_spans my-jaeger true 2 03 true 10 + '[' 6 -ne 6 ']' + jaeger=my-jaeger + is_secured=true + number_of_spans=2 + job_number=03 + ensure_reported_spans=true + test_step=10 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JOB_NUMBER=03 + JOB_NUMBER=03 + export DAYS=2 + DAYS=2 + '[' true = true ']' + protocol=https:// + query_port= + template=/tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template + '[' true = true ']' + export ENSURE_REPORTED_SPANS=true + ENSURE_REPORTED_SPANS=true + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + params= + '[' true = true ']' + '[' true = true ']' + '[' '' '!=' allInOne ']' + params='-t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template -t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template -o ./10-report-spans.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-report-spans.yaml.template -o ./10-assert.yaml + unset JAEGER_COLLECTOR_ENDPOINT + unset JAEGER_QUERY_ENDPOINT + unset JOB_NUMBER + unset DAYS + unset ENSURE_REPORTED_SPANS + CRONJOB_NAME=my-jaeger-es-rollover + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/wait-for-cronjob-execution.yaml.template -o ./11-wait-rollover.yaml + render_check_indices true ''\''--name'\'', '\''jaeger-span-000002'\'',' 05 11 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--name'\'', '\''jaeger-span-000002'\'',' + job_number=05 + test_step=11 + escape_command ''\''--name'\'', '\''jaeger-span-000002'\'',' + '[' 1 -ne 1 ']' + command=''\''--name'\'', '\''jaeger-span-000002'\'',' ++ echo ''\''--name'\'', '\''jaeger-span-000002'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--name'\'', '\''jaeger-span-000002'\'',' + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-000002'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=05 + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-000002'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./11-check-indices.yaml + JOB_NUMBER=05 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./11-assert.yaml + render_check_indices true ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' 06 12 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' + job_number=06 + test_step=12 + escape_command ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' + '[' 1 -ne 1 ']' + command=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' ++ echo ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=06 + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./12-check-indices.yaml + JOB_NUMBER=06 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./12-assert.yaml + '[' true = true ']' + get_elasticsearch_openshift_operator_version + export ESO_OPERATOR_VERSION + '[' true = true ']' ++ kubectl get pods -l name=elasticsearch-operator --all-namespaces '-o=jsonpath={.items[0].metadata.annotations.operatorframework\.io/properties}' + properties='{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.16},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.12"}}]}' + '[' -z '{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.16},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.12"}}]}' ']' ++ echo '{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.16},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.12"}}]}' ++ /tmp/jaeger-tests/bin/yq e -P '.properties.[] | select(.value.packageName == "elasticsearch-operator") | .value.version' + ESO_OPERATOR_VERSION=5.8.12 ++ version_ge 5.8.12 5.4 +++ echo 5.8.12 5.4 +++ tr ' ' '\n' +++ sort -rV +++ head -n 1 ++ test 5.8.12 == 5.8.12 + '[' -n '' ']' + skip_test es-rollover-managed 'Test only supported with Elasticsearch OpenShift Operator >= 5.4' + '[' 2 -ne 2 ']' + test_name=es-rollover-managed + message='Test only supported with Elasticsearch OpenShift Operator >= 5.4' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build/es-rollover-autoprov + '[' es-rollover-autoprov '!=' _build ']' + cd .. + rm -rf es-rollover-managed + warning 'es-rollover-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: es-rollover-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4\e[0m' WAR: es-rollover-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4 + '[' true = true ']' + skip_test es-spark-dependencies 'This test is not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=es-spark-dependencies + message='This test is not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build + '[' _build '!=' _build ']' + rm -rf es-spark-dependencies + warning 'es-spark-dependencies: This test is not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: es-spark-dependencies: This test is not supported in OpenShift\e[0m' WAR: es-spark-dependencies: This test is not supported in OpenShift make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running elasticsearch E2E tests' Running elasticsearch E2E tests + cd tests/e2e/elasticsearch/_build + set +e + KUBECONFIG=/tmp/kubeconfig-2486720052 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-rosa-h-u0tm.f61o.s3.devshift.org:443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 7 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/es-from-aio-to-production === PAUSE kuttl/harness/es-from-aio-to-production === RUN kuttl/harness/es-increasing-replicas === PAUSE kuttl/harness/es-increasing-replicas === RUN kuttl/harness/es-index-cleaner-autoprov === PAUSE kuttl/harness/es-index-cleaner-autoprov === RUN kuttl/harness/es-multiinstance === PAUSE kuttl/harness/es-multiinstance === RUN kuttl/harness/es-rollover-autoprov === PAUSE kuttl/harness/es-rollover-autoprov === RUN kuttl/harness/es-simple-prod === PAUSE kuttl/harness/es-simple-prod === CONT kuttl/harness/artifacts logger.go:42: 07:18:24 | artifacts | Creating namespace: kuttl-test-alert-kitten logger.go:42: 07:18:25 | artifacts | artifacts events from ns kuttl-test-alert-kitten: logger.go:42: 07:18:25 | artifacts | Deleting namespace: kuttl-test-alert-kitten === CONT kuttl/harness/es-multiinstance logger.go:42: 07:18:31 | es-multiinstance | Ignoring 03-create-second-instance.yaml.template as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:18:31 | es-multiinstance | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:18:31 | es-multiinstance | Creating namespace: kuttl-test-ideal-dane logger.go:42: 07:18:31 | es-multiinstance/0-clear-namespace | starting test step 0-clear-namespace logger.go:42: 07:18:31 | es-multiinstance/0-clear-namespace | running command: [sh -c kubectl delete namespace jaeger-e2e-multiinstance-test --ignore-not-found=true] logger.go:42: 07:18:31 | es-multiinstance/0-clear-namespace | test step completed 0-clear-namespace logger.go:42: 07:18:31 | es-multiinstance/1-install | starting test step 1-install logger.go:42: 07:18:31 | es-multiinstance/1-install | Jaeger:kuttl-test-ideal-dane/instance-1 created logger.go:42: 07:19:04 | es-multiinstance/1-install | test step completed 1-install logger.go:42: 07:19:04 | es-multiinstance/2-create-namespace | starting test step 2-create-namespace logger.go:42: 07:19:04 | es-multiinstance/2-create-namespace | running command: [sh -c kubectl create namespace jaeger-e2e-multiinstance-test] logger.go:42: 07:19:04 | es-multiinstance/2-create-namespace | namespace/jaeger-e2e-multiinstance-test created logger.go:42: 07:19:04 | es-multiinstance/2-create-namespace | test step completed 2-create-namespace logger.go:42: 07:19:04 | es-multiinstance/3-create-second-instance | starting test step 3-create-second-instance logger.go:42: 07:19:04 | es-multiinstance/3-create-second-instance | running command: [sh -c kubectl apply -f ./01-install.yaml -n jaeger-e2e-multiinstance-test] logger.go:42: 07:19:05 | es-multiinstance/3-create-second-instance | jaeger.jaegertracing.io/instance-1 created logger.go:42: 07:19:05 | es-multiinstance/3-create-second-instance | running command: [sh -c /tmp/jaeger-tests/bin/kubectl-kuttl assert ./01-assert.yaml -n jaeger-e2e-multiinstance-test --timeout 1000] logger.go:42: 07:19:39 | es-multiinstance/3-create-second-instance | assert is valid logger.go:42: 07:19:39 | es-multiinstance/3-create-second-instance | test step completed 3-create-second-instance logger.go:42: 07:19:39 | es-multiinstance/4-check-secrets | starting test step 4-check-secrets logger.go:42: 07:19:39 | es-multiinstance/4-check-secrets | running command: [sh -c kubectl get secrets elasticsearch -o jsonpath='{.data.logging-es\.crt}' -n $NAMESPACE > secret1] logger.go:42: 07:19:39 | es-multiinstance/4-check-secrets | running command: [sh -c kubectl get secrets elasticsearch -o jsonpath='{.data.logging-es\.crt}' -n jaeger-e2e-multiinstance-test > secret2] logger.go:42: 07:19:40 | es-multiinstance/4-check-secrets | running command: [sh -c cmp --silent secret1 secret2 || exit 0] logger.go:42: 07:19:40 | es-multiinstance/4-check-secrets | test step completed 4-check-secrets logger.go:42: 07:19:40 | es-multiinstance/5-delete | starting test step 5-delete logger.go:42: 07:19:40 | es-multiinstance/5-delete | running command: [sh -c kubectl delete namespace jaeger-e2e-multiinstance-test --wait=false] logger.go:42: 07:19:40 | es-multiinstance/5-delete | namespace "jaeger-e2e-multiinstance-test" deleted logger.go:42: 07:19:40 | es-multiinstance/5-delete | test step completed 5-delete logger.go:42: 07:19:40 | es-multiinstance | es-multiinstance events from ns kuttl-test-ideal-dane: logger.go:42: 07:19:40 | es-multiinstance | 2024-09-12 07:18:34 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestidealdaneinstance1-1-7dbb464b55 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestidealdaneinstance1-1-7dbb464b55dfs9z replicaset-controller logger.go:42: 07:19:40 | es-multiinstance | 2024-09-12 07:18:34 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestidealdaneinstance1-1-7dbb464b55dfs9z Binding Scheduled Successfully assigned kuttl-test-ideal-dane/elasticsearch-cdm-kuttltestidealdaneinstance1-1-7dbb464b55dfs9z to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:19:40 | es-multiinstance | 2024-09-12 07:18:34 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestidealdaneinstance1-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestidealdaneinstance1-1-7dbb464b55 to 1 deployment-controller logger.go:42: 07:19:40 | es-multiinstance | 2024-09-12 07:18:35 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestidealdaneinstance1-1-7dbb464b55dfs9z AddedInterface Add eth0 [10.130.0.53/23] from ovn-kubernetes logger.go:42: 07:19:40 | es-multiinstance | 2024-09-12 07:18:35 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestidealdaneinstance1-1-7dbb464b55dfs9z.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:bc97b8e13087050dfb1a75b02c5b780cbb8fb12a849a655169d072ba8bbf42b4" already present on machine kubelet logger.go:42: 07:19:40 | es-multiinstance | 2024-09-12 07:18:35 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestidealdaneinstance1-1-7dbb464b55dfs9z.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:19:40 | es-multiinstance | 2024-09-12 07:18:35 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestidealdaneinstance1-1-7dbb464b55dfs9z.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:19:40 | es-multiinstance | 2024-09-12 07:18:35 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestidealdaneinstance1-1-7dbb464b55dfs9z.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:1d1d7d47b616995d18692f3a2d6232d7e1a0f41bc1503d87100f31939e080a78" already present on machine kubelet logger.go:42: 07:19:40 | es-multiinstance | 2024-09-12 07:18:35 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestidealdaneinstance1-1-7dbb464b55dfs9z.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:19:40 | es-multiinstance | 2024-09-12 07:18:35 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestidealdaneinstance1-1-7dbb464b55dfs9z.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:19:40 | es-multiinstance | 2024-09-12 07:18:45 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestidealdaneinstance1-1-7dbb464b55dfs9z.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 07:19:40 | es-multiinstance | 2024-09-12 07:18:50 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestidealdaneinstance1-1-7dbb464b55dfs9z.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:19:40 | es-multiinstance | 2024-09-12 07:19:01 +0000 UTC Normal Pod instance-1-collector-54bf55456d-6jnws Binding Scheduled Successfully assigned kuttl-test-ideal-dane/instance-1-collector-54bf55456d-6jnws to ip-10-0-54-110.us-west-2.compute.internal default-scheduler logger.go:42: 07:19:40 | es-multiinstance | 2024-09-12 07:19:01 +0000 UTC Normal ReplicaSet.apps instance-1-collector-54bf55456d SuccessfulCreate Created pod: instance-1-collector-54bf55456d-6jnws replicaset-controller logger.go:42: 07:19:40 | es-multiinstance | 2024-09-12 07:19:01 +0000 UTC Normal Deployment.apps instance-1-collector ScalingReplicaSet Scaled up replica set instance-1-collector-54bf55456d to 1 deployment-controller logger.go:42: 07:19:40 | es-multiinstance | 2024-09-12 07:19:01 +0000 UTC Normal Pod instance-1-query-6bc4445899-t9qs2 Binding Scheduled Successfully assigned kuttl-test-ideal-dane/instance-1-query-6bc4445899-t9qs2 to ip-10-0-52-42.us-west-2.compute.internal default-scheduler logger.go:42: 07:19:40 | es-multiinstance | 2024-09-12 07:19:01 +0000 UTC Normal ReplicaSet.apps instance-1-query-6bc4445899 SuccessfulCreate Created pod: instance-1-query-6bc4445899-t9qs2 replicaset-controller logger.go:42: 07:19:40 | es-multiinstance | 2024-09-12 07:19:01 +0000 UTC Normal Deployment.apps instance-1-query ScalingReplicaSet Scaled up replica set instance-1-query-6bc4445899 to 1 deployment-controller logger.go:42: 07:19:40 | es-multiinstance | 2024-09-12 07:19:02 +0000 UTC Normal Pod instance-1-collector-54bf55456d-6jnws AddedInterface Add eth0 [10.129.0.43/23] from ovn-kubernetes logger.go:42: 07:19:40 | es-multiinstance | 2024-09-12 07:19:02 +0000 UTC Normal Pod instance-1-collector-54bf55456d-6jnws.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:4bab63290ff52e4f6328009f9c8b2c4496b230e9f8a98eac01736a66a291ff6c" already present on machine kubelet logger.go:42: 07:19:40 | es-multiinstance | 2024-09-12 07:19:02 +0000 UTC Normal Pod instance-1-collector-54bf55456d-6jnws.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:19:40 | es-multiinstance | 2024-09-12 07:19:02 +0000 UTC Normal Pod instance-1-collector-54bf55456d-6jnws.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:19:40 | es-multiinstance | 2024-09-12 07:19:02 +0000 UTC Normal Pod instance-1-query-6bc4445899-t9qs2 AddedInterface Add eth0 [10.128.0.54/23] from ovn-kubernetes logger.go:42: 07:19:40 | es-multiinstance | 2024-09-12 07:19:02 +0000 UTC Normal Pod instance-1-query-6bc4445899-t9qs2.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a9186dcd910256c0f464b0a3928844a01de166a10c186c97ef4581bf288c23cb" already present on machine kubelet logger.go:42: 07:19:40 | es-multiinstance | 2024-09-12 07:19:02 +0000 UTC Normal Pod instance-1-query-6bc4445899-t9qs2.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:19:40 | es-multiinstance | 2024-09-12 07:19:02 +0000 UTC Normal Pod instance-1-query-6bc4445899-t9qs2.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:19:40 | es-multiinstance | 2024-09-12 07:19:02 +0000 UTC Normal Pod instance-1-query-6bc4445899-t9qs2.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:19:40 | es-multiinstance | 2024-09-12 07:19:02 +0000 UTC Normal Pod instance-1-query-6bc4445899-t9qs2.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:19:40 | es-multiinstance | 2024-09-12 07:19:02 +0000 UTC Normal Pod instance-1-query-6bc4445899-t9qs2.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:19:40 | es-multiinstance | 2024-09-12 07:19:02 +0000 UTC Normal Pod instance-1-query-6bc4445899-t9qs2.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" already present on machine kubelet logger.go:42: 07:19:40 | es-multiinstance | 2024-09-12 07:19:02 +0000 UTC Normal Pod instance-1-query-6bc4445899-t9qs2.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:19:40 | es-multiinstance | 2024-09-12 07:19:02 +0000 UTC Normal Pod instance-1-query-6bc4445899-t9qs2.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:19:40 | es-multiinstance | 2024-09-12 07:19:16 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling instance-1-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:19:40 | es-multiinstance | 2024-09-12 07:19:16 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling instance-1-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:19:40 | es-multiinstance | 2024-09-12 07:19:16 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling instance-1-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:19:40 | es-multiinstance | Deleting namespace: kuttl-test-ideal-dane === CONT kuttl/harness/es-simple-prod logger.go:42: 07:19:47 | es-simple-prod | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:19:47 | es-simple-prod | Creating namespace: kuttl-test-notable-mongrel logger.go:42: 07:19:47 | es-simple-prod | es-simple-prod events from ns kuttl-test-notable-mongrel: logger.go:42: 07:19:47 | es-simple-prod | Deleting namespace: kuttl-test-notable-mongrel === CONT kuttl/harness/es-rollover-autoprov logger.go:42: 07:19:53 | es-rollover-autoprov | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:19:53 | es-rollover-autoprov | Creating namespace: kuttl-test-sharp-ocelot logger.go:42: 07:19:53 | es-rollover-autoprov/1-install | starting test step 1-install logger.go:42: 07:19:53 | es-rollover-autoprov/1-install | Jaeger:kuttl-test-sharp-ocelot/my-jaeger created logger.go:42: 07:20:28 | es-rollover-autoprov/1-install | test step completed 1-install logger.go:42: 07:20:28 | es-rollover-autoprov/2-report-spans | starting test step 2-report-spans logger.go:42: 07:20:28 | es-rollover-autoprov/2-report-spans | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:20:30 | es-rollover-autoprov/2-report-spans | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:20:37 | es-rollover-autoprov/2-report-spans | running command: [sh -c DAYS=2 ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JOB_NUMBER=00 JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/report-spans.yaml.template -o report-span-00-job.yaml] logger.go:42: 07:20:37 | es-rollover-autoprov/2-report-spans | running command: [sh -c kubectl apply -f report-span-00-job.yaml -n $NAMESPACE] logger.go:42: 07:20:38 | es-rollover-autoprov/2-report-spans | job.batch/00-report-span created logger.go:42: 07:21:02 | es-rollover-autoprov/2-report-spans | test step completed 2-report-spans logger.go:42: 07:21:02 | es-rollover-autoprov/3-check-indices | starting test step 3-check-indices logger.go:42: 07:21:03 | es-rollover-autoprov/3-check-indices | Job:kuttl-test-sharp-ocelot/00-check-indices created logger.go:42: 07:21:08 | es-rollover-autoprov/3-check-indices | test step completed 3-check-indices logger.go:42: 07:21:08 | es-rollover-autoprov/4-check-indices | starting test step 4-check-indices logger.go:42: 07:21:08 | es-rollover-autoprov/4-check-indices | Job:kuttl-test-sharp-ocelot/01-check-indices created logger.go:42: 07:21:14 | es-rollover-autoprov/4-check-indices | test step completed 4-check-indices logger.go:42: 07:21:14 | es-rollover-autoprov/5-install | starting test step 5-install logger.go:42: 07:21:14 | es-rollover-autoprov/5-install | Jaeger:kuttl-test-sharp-ocelot/my-jaeger updated logger.go:42: 07:21:22 | es-rollover-autoprov/5-install | test step completed 5-install logger.go:42: 07:21:22 | es-rollover-autoprov/6-report-spans | starting test step 6-report-spans logger.go:42: 07:21:22 | es-rollover-autoprov/6-report-spans | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:21:30 | es-rollover-autoprov/6-report-spans | running command: [sh -c DAYS=2 ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JOB_NUMBER=02 JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/report-spans.yaml.template -o report-span-02-job.yaml] logger.go:42: 07:21:31 | es-rollover-autoprov/6-report-spans | running command: [sh -c kubectl apply -f report-span-02-job.yaml -n $NAMESPACE] logger.go:42: 07:21:31 | es-rollover-autoprov/6-report-spans | job.batch/02-report-span created logger.go:42: 07:21:56 | es-rollover-autoprov/6-report-spans | test step completed 6-report-spans logger.go:42: 07:21:56 | es-rollover-autoprov/7-check-indices | starting test step 7-check-indices logger.go:42: 07:21:56 | es-rollover-autoprov/7-check-indices | Job:kuttl-test-sharp-ocelot/02-check-indices created logger.go:42: 07:22:00 | es-rollover-autoprov/7-check-indices | test step completed 7-check-indices logger.go:42: 07:22:00 | es-rollover-autoprov/8-check-indices | starting test step 8-check-indices logger.go:42: 07:22:01 | es-rollover-autoprov/8-check-indices | Job:kuttl-test-sharp-ocelot/03-check-indices created logger.go:42: 07:22:05 | es-rollover-autoprov/8-check-indices | test step completed 8-check-indices logger.go:42: 07:22:05 | es-rollover-autoprov/9-check-indices | starting test step 9-check-indices logger.go:42: 07:22:05 | es-rollover-autoprov/9-check-indices | Job:kuttl-test-sharp-ocelot/04-check-indices created logger.go:42: 07:22:11 | es-rollover-autoprov/9-check-indices | test step completed 9-check-indices logger.go:42: 07:22:11 | es-rollover-autoprov/10-report-spans | starting test step 10-report-spans logger.go:42: 07:22:11 | es-rollover-autoprov/10-report-spans | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:22:20 | es-rollover-autoprov/10-report-spans | running command: [sh -c DAYS=2 ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JOB_NUMBER=03 JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/report-spans.yaml.template -o report-span-03-job.yaml] logger.go:42: 07:22:21 | es-rollover-autoprov/10-report-spans | running command: [sh -c kubectl apply -f report-span-03-job.yaml -n $NAMESPACE] logger.go:42: 07:22:21 | es-rollover-autoprov/10-report-spans | job.batch/03-report-span created logger.go:42: 07:22:46 | es-rollover-autoprov/10-report-spans | test step completed 10-report-spans logger.go:42: 07:22:46 | es-rollover-autoprov/11-check-indices | starting test step 11-check-indices logger.go:42: 07:22:46 | es-rollover-autoprov/11-check-indices | running command: [sh -c go run ../../../../cmd-utils/wait-cronjob/main.go --cronjob my-jaeger-es-rollover --namespace $NAMESPACE] logger.go:42: 07:22:55 | es-rollover-autoprov/11-check-indices | time="2024-09-12T07:22:55Z" level=debug msg="Checking if the my-jaeger-es-rollover CronJob exists" logger.go:42: 07:22:55 | es-rollover-autoprov/11-check-indices | time="2024-09-12T07:22:55Z" level=debug msg="No BatchV1beta1/Cronjobs were found" logger.go:42: 07:22:55 | es-rollover-autoprov/11-check-indices | time="2024-09-12T07:22:55Z" level=info msg="Cronjob my-jaeger-es-rollover found successfully" logger.go:42: 07:22:55 | es-rollover-autoprov/11-check-indices | time="2024-09-12T07:22:55Z" level=debug msg="Waiting for the next scheduled job from my-jaeger-es-rollover cronjob" logger.go:42: 07:22:56 | es-rollover-autoprov/11-check-indices | time="2024-09-12T07:22:56Z" level=debug msg="Waiting for next job from my-jaeger-es-rollover to succeed" logger.go:42: 07:23:06 | es-rollover-autoprov/11-check-indices | time="2024-09-12T07:23:06Z" level=info msg="Job of owner my-jaeger-es-rollover succeeded after my-jaeger-es-rollover 10.164442184s" logger.go:42: 07:23:06 | es-rollover-autoprov/11-check-indices | Job:kuttl-test-sharp-ocelot/05-check-indices created logger.go:42: 07:23:11 | es-rollover-autoprov/11-check-indices | test step completed 11-check-indices logger.go:42: 07:23:11 | es-rollover-autoprov/12-check-indices | starting test step 12-check-indices logger.go:42: 07:23:12 | es-rollover-autoprov/12-check-indices | Job:kuttl-test-sharp-ocelot/06-check-indices created logger.go:42: 07:23:16 | es-rollover-autoprov/12-check-indices | test step completed 12-check-indices logger.go:42: 07:23:16 | es-rollover-autoprov | es-rollover-autoprov events from ns kuttl-test-sharp-ocelot: logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:19:58 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestsharpocelotmyjaeger-1-6dfdfbbd8f SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestsharpocelotmyjaeger-1-6dfdfbbd8vn6zs replicaset-controller logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:19:58 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsharpocelotmyjaeger-1-6dfdfbbd8vn6zs Binding Scheduled Successfully assigned kuttl-test-sharp-ocelot/elasticsearch-cdm-kuttltestsharpocelotmyjaeger-1-6dfdfbbd8vn6zs to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:19:58 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsharpocelotmyjaeger-1-6dfdfbbd8vn6zs AddedInterface Add eth0 [10.130.0.55/23] from ovn-kubernetes logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:19:58 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsharpocelotmyjaeger-1-6dfdfbbd8vn6zs.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:bc97b8e13087050dfb1a75b02c5b780cbb8fb12a849a655169d072ba8bbf42b4" already present on machine kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:19:58 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsharpocelotmyjaeger-1-6dfdfbbd8vn6zs.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:19:58 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsharpocelotmyjaeger-1-6dfdfbbd8vn6zs.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:19:58 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsharpocelotmyjaeger-1-6dfdfbbd8vn6zs.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:1d1d7d47b616995d18692f3a2d6232d7e1a0f41bc1503d87100f31939e080a78" already present on machine kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:19:58 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsharpocelotmyjaeger-1-6dfdfbbd8vn6zs.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:19:58 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestsharpocelotmyjaeger-1-6dfdfbbd8vn6zs.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:19:58 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestsharpocelotmyjaeger-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestsharpocelotmyjaeger-1-6dfdfbbd8f to 1 deployment-controller logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:20:08 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestsharpocelotmyjaeger-1-6dfdfbbd8vn6zs.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:20:13 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestsharpocelotmyjaeger-1-6dfdfbbd8vn6zs.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:20:25 +0000 UTC Normal Pod my-jaeger-collector-7f7bb876cf-98rv4 Binding Scheduled Successfully assigned kuttl-test-sharp-ocelot/my-jaeger-collector-7f7bb876cf-98rv4 to ip-10-0-54-110.us-west-2.compute.internal default-scheduler logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:20:25 +0000 UTC Normal Pod my-jaeger-collector-7f7bb876cf-98rv4 AddedInterface Add eth0 [10.129.0.46/23] from ovn-kubernetes logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:20:25 +0000 UTC Normal Pod my-jaeger-collector-7f7bb876cf-98rv4.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:4bab63290ff52e4f6328009f9c8b2c4496b230e9f8a98eac01736a66a291ff6c" already present on machine kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:20:25 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-7f7bb876cf SuccessfulCreate Created pod: my-jaeger-collector-7f7bb876cf-98rv4 replicaset-controller logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:20:25 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-7f7bb876cf to 1 deployment-controller logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:20:25 +0000 UTC Normal Pod my-jaeger-query-669fb87d69-8tqrj Binding Scheduled Successfully assigned kuttl-test-sharp-ocelot/my-jaeger-query-669fb87d69-8tqrj to ip-10-0-52-42.us-west-2.compute.internal default-scheduler logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:20:25 +0000 UTC Warning Pod my-jaeger-query-669fb87d69-8tqrj FailedMount MountVolume.SetUp failed for volume "my-jaeger-ui-oauth-proxy-tls" : secret "my-jaeger-ui-oauth-proxy-tls" not found kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:20:25 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-669fb87d69 SuccessfulCreate Created pod: my-jaeger-query-669fb87d69-8tqrj replicaset-controller logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:20:25 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-669fb87d69 to 1 deployment-controller logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:20:26 +0000 UTC Normal Pod my-jaeger-collector-7f7bb876cf-98rv4.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:20:26 +0000 UTC Normal Pod my-jaeger-collector-7f7bb876cf-98rv4.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:20:26 +0000 UTC Normal Pod my-jaeger-query-669fb87d69-8tqrj AddedInterface Add eth0 [10.128.0.56/23] from ovn-kubernetes logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:20:26 +0000 UTC Normal Pod my-jaeger-query-669fb87d69-8tqrj.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a9186dcd910256c0f464b0a3928844a01de166a10c186c97ef4581bf288c23cb" already present on machine kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:20:26 +0000 UTC Normal Pod my-jaeger-query-669fb87d69-8tqrj.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:20:26 +0000 UTC Normal Pod my-jaeger-query-669fb87d69-8tqrj.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:20:26 +0000 UTC Normal Pod my-jaeger-query-669fb87d69-8tqrj.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:20:26 +0000 UTC Normal Pod my-jaeger-query-669fb87d69-8tqrj.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:20:26 +0000 UTC Normal Pod my-jaeger-query-669fb87d69-8tqrj.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:20:26 +0000 UTC Normal Pod my-jaeger-query-669fb87d69-8tqrj.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" already present on machine kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:20:26 +0000 UTC Normal Pod my-jaeger-query-669fb87d69-8tqrj.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:20:26 +0000 UTC Normal Pod my-jaeger-query-669fb87d69-8tqrj.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:20:30 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-669fb87d69 SuccessfulDelete Deleted pod: my-jaeger-query-669fb87d69-8tqrj replicaset-controller logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:20:30 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled down replica set my-jaeger-query-669fb87d69 to 0 from 1 deployment-controller logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:20:31 +0000 UTC Normal Pod my-jaeger-query-669fb87d69-8tqrj.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:20:31 +0000 UTC Normal Pod my-jaeger-query-669fb87d69-8tqrj.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:20:31 +0000 UTC Normal Pod my-jaeger-query-669fb87d69-8tqrj.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:20:32 +0000 UTC Normal Pod my-jaeger-query-5c5864c656-2xmx5 Binding Scheduled Successfully assigned kuttl-test-sharp-ocelot/my-jaeger-query-5c5864c656-2xmx5 to ip-10-0-52-42.us-west-2.compute.internal default-scheduler logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:20:32 +0000 UTC Normal Pod my-jaeger-query-5c5864c656-2xmx5 AddedInterface Add eth0 [10.128.0.57/23] from ovn-kubernetes logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:20:32 +0000 UTC Normal Pod my-jaeger-query-5c5864c656-2xmx5.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a9186dcd910256c0f464b0a3928844a01de166a10c186c97ef4581bf288c23cb" already present on machine kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:20:32 +0000 UTC Normal Pod my-jaeger-query-5c5864c656-2xmx5.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:20:32 +0000 UTC Normal Pod my-jaeger-query-5c5864c656-2xmx5.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:20:32 +0000 UTC Normal Pod my-jaeger-query-5c5864c656-2xmx5.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:20:32 +0000 UTC Normal Pod my-jaeger-query-5c5864c656-2xmx5.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:20:32 +0000 UTC Normal Pod my-jaeger-query-5c5864c656-2xmx5.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:20:32 +0000 UTC Normal Pod my-jaeger-query-5c5864c656-2xmx5.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" already present on machine kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:20:32 +0000 UTC Normal Pod my-jaeger-query-5c5864c656-2xmx5.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:20:32 +0000 UTC Normal Pod my-jaeger-query-5c5864c656-2xmx5.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:20:32 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-5c5864c656 SuccessfulCreate Created pod: my-jaeger-query-5c5864c656-2xmx5 replicaset-controller logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:20:32 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-5c5864c656 to 1 deployment-controller logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:20:38 +0000 UTC Normal Pod 00-report-span-f98vp Binding Scheduled Successfully assigned kuttl-test-sharp-ocelot/00-report-span-f98vp to ip-10-0-54-110.us-west-2.compute.internal default-scheduler logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:20:38 +0000 UTC Normal Pod 00-report-span-f98vp AddedInterface Add eth0 [10.129.0.47/23] from ovn-kubernetes logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:20:38 +0000 UTC Normal Pod 00-report-span-f98vp.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:20:38 +0000 UTC Normal Job.batch 00-report-span SuccessfulCreate Created pod: 00-report-span-f98vp job-controller logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:20:39 +0000 UTC Normal Pod 00-report-span-f98vp.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 882ms (882ms including waiting) kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:20:39 +0000 UTC Normal Pod 00-report-span-f98vp.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:20:39 +0000 UTC Normal Pod 00-report-span-f98vp.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:20:40 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:20:40 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:20:40 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:21:02 +0000 UTC Normal Job.batch 00-report-span Completed Job completed job-controller logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:21:03 +0000 UTC Normal Pod 00-check-indices-9zqd2 Binding Scheduled Successfully assigned kuttl-test-sharp-ocelot/00-check-indices-9zqd2 to ip-10-0-54-110.us-west-2.compute.internal default-scheduler logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:21:03 +0000 UTC Normal Pod 00-check-indices-9zqd2 AddedInterface Add eth0 [10.129.0.48/23] from ovn-kubernetes logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:21:03 +0000 UTC Normal Pod 00-check-indices-9zqd2.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:21:03 +0000 UTC Normal Job.batch 00-check-indices SuccessfulCreate Created pod: 00-check-indices-9zqd2 job-controller logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:21:04 +0000 UTC Normal Pod 00-check-indices-9zqd2.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 826ms (826ms including waiting) kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:21:04 +0000 UTC Normal Pod 00-check-indices-9zqd2.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:21:04 +0000 UTC Normal Pod 00-check-indices-9zqd2.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:21:07 +0000 UTC Normal Job.batch 00-check-indices Completed Job completed job-controller logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:21:08 +0000 UTC Normal Pod 01-check-indices-tf6gb Binding Scheduled Successfully assigned kuttl-test-sharp-ocelot/01-check-indices-tf6gb to ip-10-0-54-110.us-west-2.compute.internal default-scheduler logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:21:08 +0000 UTC Normal Job.batch 01-check-indices SuccessfulCreate Created pod: 01-check-indices-tf6gb job-controller logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:21:09 +0000 UTC Normal Pod 01-check-indices-tf6gb AddedInterface Add eth0 [10.129.0.49/23] from ovn-kubernetes logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:21:09 +0000 UTC Normal Pod 01-check-indices-tf6gb.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:21:10 +0000 UTC Normal Pod 01-check-indices-tf6gb.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 839ms (839ms including waiting) kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:21:10 +0000 UTC Normal Pod 01-check-indices-tf6gb.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:21:10 +0000 UTC Normal Pod 01-check-indices-tf6gb.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:21:10 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:21:10 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get memory utilization: missing request for memory in container jaeger-collector of Pod my-jaeger-collector-7f7bb876cf-98rv4 horizontal-pod-autoscaler logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:21:10 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:21:13 +0000 UTC Normal Job.batch 01-check-indices Completed Job completed job-controller logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:21:15 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-vr6s7 Binding Scheduled Successfully assigned kuttl-test-sharp-ocelot/my-jaeger-es-rollover-create-mapping-vr6s7 to ip-10-0-54-110.us-west-2.compute.internal default-scheduler logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:21:15 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-vr6s7 AddedInterface Add eth0 [10.129.0.50/23] from ovn-kubernetes logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:21:15 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-vr6s7.spec.containers{my-jaeger-es-rollover-create-mapping} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:b3897239050ab12ae12c28bb51d955438334aad957dceb83a53cf2e8a466b69b" kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:21:15 +0000 UTC Normal Job.batch my-jaeger-es-rollover-create-mapping SuccessfulCreate Created pod: my-jaeger-es-rollover-create-mapping-vr6s7 job-controller logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:21:18 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-vr6s7.spec.containers{my-jaeger-es-rollover-create-mapping} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:b3897239050ab12ae12c28bb51d955438334aad957dceb83a53cf2e8a466b69b" in 2.832s (2.832s including waiting) kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:21:18 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-vr6s7.spec.containers{my-jaeger-es-rollover-create-mapping} Created Created container my-jaeger-es-rollover-create-mapping kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:21:18 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-vr6s7.spec.containers{my-jaeger-es-rollover-create-mapping} Started Started container my-jaeger-es-rollover-create-mapping kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:21:21 +0000 UTC Normal Job.batch my-jaeger-es-rollover-create-mapping Completed Job completed job-controller logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:21:22 +0000 UTC Normal Pod my-jaeger-collector-7f7bb876cf-98rv4.spec.containers{jaeger-collector} Killing Stopping container jaeger-collector kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:21:22 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-7f7bb876cf SuccessfulDelete Deleted pod: my-jaeger-collector-7f7bb876cf-98rv4 replicaset-controller logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:21:22 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled down replica set my-jaeger-collector-7f7bb876cf to 0 from 1 deployment-controller logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:21:22 +0000 UTC Normal Pod my-jaeger-query-5c5864c656-2xmx5.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:21:22 +0000 UTC Normal Pod my-jaeger-query-5c5864c656-2xmx5.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:21:22 +0000 UTC Normal Pod my-jaeger-query-5c5864c656-2xmx5.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:21:22 +0000 UTC Warning Pod my-jaeger-query-5c5864c656-2xmx5.spec.containers{jaeger-agent} Unhealthy Readiness probe failed: Get "http://10.128.0.57:14271/": dial tcp 10.128.0.57:14271: connect: connection refused kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:21:22 +0000 UTC Warning Pod my-jaeger-query-5c5864c656-2xmx5.spec.containers{jaeger-query} Unhealthy Readiness probe failed: Get "http://10.128.0.57:16687/": dial tcp 10.128.0.57:16687: connect: connection refused kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:21:22 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-5c5864c656 SuccessfulDelete Deleted pod: my-jaeger-query-5c5864c656-2xmx5 replicaset-controller logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:21:22 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled down replica set my-jaeger-query-5c5864c656 to 0 from 1 deployment-controller logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:21:23 +0000 UTC Normal Pod my-jaeger-collector-7cd5cf6494-wk9q6 Binding Scheduled Successfully assigned kuttl-test-sharp-ocelot/my-jaeger-collector-7cd5cf6494-wk9q6 to ip-10-0-54-110.us-west-2.compute.internal default-scheduler logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:21:23 +0000 UTC Normal Pod my-jaeger-collector-7cd5cf6494-wk9q6 AddedInterface Add eth0 [10.129.0.51/23] from ovn-kubernetes logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:21:23 +0000 UTC Normal Pod my-jaeger-collector-7cd5cf6494-wk9q6.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:4bab63290ff52e4f6328009f9c8b2c4496b230e9f8a98eac01736a66a291ff6c" already present on machine kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:21:23 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-7cd5cf6494 SuccessfulCreate Created pod: my-jaeger-collector-7cd5cf6494-wk9q6 replicaset-controller logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:21:23 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-7cd5cf6494 to 1 deployment-controller logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:21:23 +0000 UTC Normal Pod my-jaeger-query-5fc6c949c9-bh5cf Binding Scheduled Successfully assigned kuttl-test-sharp-ocelot/my-jaeger-query-5fc6c949c9-bh5cf to ip-10-0-52-42.us-west-2.compute.internal default-scheduler logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:21:23 +0000 UTC Normal Pod my-jaeger-query-5fc6c949c9-bh5cf AddedInterface Add eth0 [10.128.0.58/23] from ovn-kubernetes logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:21:23 +0000 UTC Normal Pod my-jaeger-query-5fc6c949c9-bh5cf.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a9186dcd910256c0f464b0a3928844a01de166a10c186c97ef4581bf288c23cb" already present on machine kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:21:23 +0000 UTC Normal Pod my-jaeger-query-5fc6c949c9-bh5cf.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:21:23 +0000 UTC Normal Pod my-jaeger-query-5fc6c949c9-bh5cf.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:21:23 +0000 UTC Normal Pod my-jaeger-query-5fc6c949c9-bh5cf.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:21:23 +0000 UTC Normal Pod my-jaeger-query-5fc6c949c9-bh5cf.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:21:23 +0000 UTC Normal Pod my-jaeger-query-5fc6c949c9-bh5cf.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:21:23 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-5fc6c949c9 SuccessfulCreate Created pod: my-jaeger-query-5fc6c949c9-bh5cf replicaset-controller logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:21:23 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-5fc6c949c9 to 1 deployment-controller logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:21:24 +0000 UTC Normal Pod my-jaeger-collector-7cd5cf6494-wk9q6.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:21:24 +0000 UTC Normal Pod my-jaeger-collector-7cd5cf6494-wk9q6.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:21:24 +0000 UTC Normal Pod my-jaeger-query-5fc6c949c9-bh5cf.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" already present on machine kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:21:24 +0000 UTC Normal Pod my-jaeger-query-5fc6c949c9-bh5cf.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:21:24 +0000 UTC Normal Pod my-jaeger-query-5fc6c949c9-bh5cf.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:21:31 +0000 UTC Normal Pod 02-report-span-qlnjh Binding Scheduled Successfully assigned kuttl-test-sharp-ocelot/02-report-span-qlnjh to ip-10-0-54-110.us-west-2.compute.internal default-scheduler logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:21:31 +0000 UTC Normal Pod 02-report-span-qlnjh AddedInterface Add eth0 [10.129.0.52/23] from ovn-kubernetes logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:21:31 +0000 UTC Normal Job.batch 02-report-span SuccessfulCreate Created pod: 02-report-span-qlnjh job-controller logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:21:32 +0000 UTC Normal Pod 02-report-span-qlnjh.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:21:32 +0000 UTC Normal Pod 02-report-span-qlnjh.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 896ms (896ms including waiting) kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:21:33 +0000 UTC Normal Pod 02-report-span-qlnjh.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:21:33 +0000 UTC Normal Pod 02-report-span-qlnjh.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:21:55 +0000 UTC Normal Job.batch 02-report-span Completed Job completed job-controller logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:21:56 +0000 UTC Normal Pod 02-check-indices-zgnt8 Binding Scheduled Successfully assigned kuttl-test-sharp-ocelot/02-check-indices-zgnt8 to ip-10-0-54-110.us-west-2.compute.internal default-scheduler logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:21:56 +0000 UTC Normal Job.batch 02-check-indices SuccessfulCreate Created pod: 02-check-indices-zgnt8 job-controller logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:21:57 +0000 UTC Normal Pod 02-check-indices-zgnt8 AddedInterface Add eth0 [10.129.0.53/23] from ovn-kubernetes logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:21:57 +0000 UTC Normal Pod 02-check-indices-zgnt8.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:21:58 +0000 UTC Normal Pod 02-check-indices-zgnt8.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 827ms (827ms including waiting) kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:21:58 +0000 UTC Normal Pod 02-check-indices-zgnt8.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:21:58 +0000 UTC Normal Pod 02-check-indices-zgnt8.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:22:00 +0000 UTC Normal Job.batch 02-check-indices Completed Job completed job-controller logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:22:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28768762-v4mvq Binding Scheduled Successfully assigned kuttl-test-sharp-ocelot/my-jaeger-es-lookback-28768762-v4mvq to ip-10-0-54-110.us-west-2.compute.internal default-scheduler logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:22:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28768762-v4mvq AddedInterface Add eth0 [10.129.0.54/23] from ovn-kubernetes logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:22:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28768762-v4mvq.spec.containers{my-jaeger-es-lookback} Pulled Container image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:b3897239050ab12ae12c28bb51d955438334aad957dceb83a53cf2e8a466b69b" already present on machine kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:22:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28768762-v4mvq.spec.containers{my-jaeger-es-lookback} Created Created container my-jaeger-es-lookback kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:22:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28768762-v4mvq.spec.containers{my-jaeger-es-lookback} Started Started container my-jaeger-es-lookback kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:22:00 +0000 UTC Normal Job.batch my-jaeger-es-lookback-28768762 SuccessfulCreate Created pod: my-jaeger-es-lookback-28768762-v4mvq job-controller logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:22:00 +0000 UTC Normal CronJob.batch my-jaeger-es-lookback SuccessfulCreate Created job my-jaeger-es-lookback-28768762 cronjob-controller logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:22:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28768762-x54bc Binding Scheduled Successfully assigned kuttl-test-sharp-ocelot/my-jaeger-es-rollover-28768762-x54bc to ip-10-0-54-110.us-west-2.compute.internal default-scheduler logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:22:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28768762-x54bc AddedInterface Add eth0 [10.129.0.55/23] from ovn-kubernetes logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:22:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28768762-x54bc.spec.containers{my-jaeger-es-rollover} Pulled Container image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:b3897239050ab12ae12c28bb51d955438334aad957dceb83a53cf2e8a466b69b" already present on machine kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:22:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28768762-x54bc.spec.containers{my-jaeger-es-rollover} Created Created container my-jaeger-es-rollover kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:22:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28768762-x54bc.spec.containers{my-jaeger-es-rollover} Started Started container my-jaeger-es-rollover kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:22:00 +0000 UTC Normal Job.batch my-jaeger-es-rollover-28768762 SuccessfulCreate Created pod: my-jaeger-es-rollover-28768762-x54bc job-controller logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:22:00 +0000 UTC Normal CronJob.batch my-jaeger-es-rollover SuccessfulCreate Created job my-jaeger-es-rollover-28768762 cronjob-controller logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:22:01 +0000 UTC Normal Pod 03-check-indices-qd7q4 Binding Scheduled Successfully assigned kuttl-test-sharp-ocelot/03-check-indices-qd7q4 to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:22:01 +0000 UTC Normal Pod 03-check-indices-qd7q4 AddedInterface Add eth0 [10.130.0.56/23] from ovn-kubernetes logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:22:01 +0000 UTC Normal Pod 03-check-indices-qd7q4.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:22:01 +0000 UTC Normal Job.batch 03-check-indices SuccessfulCreate Created pod: 03-check-indices-qd7q4 job-controller logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:22:02 +0000 UTC Normal Pod 03-check-indices-qd7q4.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 867ms (867ms including waiting) kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:22:02 +0000 UTC Normal Pod 03-check-indices-qd7q4.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:22:02 +0000 UTC Normal Pod 03-check-indices-qd7q4.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:22:03 +0000 UTC Normal Job.batch my-jaeger-es-lookback-28768762 Completed Job completed job-controller logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:22:03 +0000 UTC Normal CronJob.batch my-jaeger-es-lookback SawCompletedJob Saw completed job: my-jaeger-es-lookback-28768762, status: Complete cronjob-controller logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:22:03 +0000 UTC Normal Job.batch my-jaeger-es-rollover-28768762 Completed Job completed job-controller logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:22:03 +0000 UTC Normal CronJob.batch my-jaeger-es-rollover SawCompletedJob Saw completed job: my-jaeger-es-rollover-28768762, status: Complete cronjob-controller logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:22:05 +0000 UTC Normal Job.batch 03-check-indices Completed Job completed job-controller logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:22:05 +0000 UTC Normal Pod 04-check-indices-qks2n Binding Scheduled Successfully assigned kuttl-test-sharp-ocelot/04-check-indices-qks2n to ip-10-0-54-110.us-west-2.compute.internal default-scheduler logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:22:05 +0000 UTC Normal Job.batch 04-check-indices SuccessfulCreate Created pod: 04-check-indices-qks2n job-controller logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:22:06 +0000 UTC Normal Pod 04-check-indices-qks2n AddedInterface Add eth0 [10.129.0.56/23] from ovn-kubernetes logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:22:06 +0000 UTC Normal Pod 04-check-indices-qks2n.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:22:07 +0000 UTC Normal Pod 04-check-indices-qks2n.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 812ms (812ms including waiting) kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:22:07 +0000 UTC Normal Pod 04-check-indices-qks2n.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:22:07 +0000 UTC Normal Pod 04-check-indices-qks2n.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:22:10 +0000 UTC Normal Job.batch 04-check-indices Completed Job completed job-controller logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:22:10 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get memory utilization: missing request for memory in container jaeger-collector of Pod my-jaeger-collector-7cd5cf6494-wk9q6 horizontal-pod-autoscaler logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:22:21 +0000 UTC Normal Pod 03-report-span-5dkkt Binding Scheduled Successfully assigned kuttl-test-sharp-ocelot/03-report-span-5dkkt to ip-10-0-54-110.us-west-2.compute.internal default-scheduler logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:22:21 +0000 UTC Normal Job.batch 03-report-span SuccessfulCreate Created pod: 03-report-span-5dkkt job-controller logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:22:22 +0000 UTC Normal Pod 03-report-span-5dkkt AddedInterface Add eth0 [10.129.0.57/23] from ovn-kubernetes logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:22:22 +0000 UTC Normal Pod 03-report-span-5dkkt.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:22:23 +0000 UTC Normal Pod 03-report-span-5dkkt.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 859ms (859ms including waiting) kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:22:23 +0000 UTC Normal Pod 03-report-span-5dkkt.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:22:23 +0000 UTC Normal Pod 03-report-span-5dkkt.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:22:45 +0000 UTC Normal Job.batch 03-report-span Completed Job completed job-controller logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:23:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28768763-qbwms Binding Scheduled Successfully assigned kuttl-test-sharp-ocelot/my-jaeger-es-lookback-28768763-qbwms to ip-10-0-54-110.us-west-2.compute.internal default-scheduler logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:23:00 +0000 UTC Normal Job.batch my-jaeger-es-lookback-28768763 SuccessfulCreate Created pod: my-jaeger-es-lookback-28768763-qbwms job-controller logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:23:00 +0000 UTC Normal CronJob.batch my-jaeger-es-lookback SuccessfulCreate Created job my-jaeger-es-lookback-28768763 cronjob-controller logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:23:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28768763-r9x7k Binding Scheduled Successfully assigned kuttl-test-sharp-ocelot/my-jaeger-es-rollover-28768763-r9x7k to ip-10-0-54-110.us-west-2.compute.internal default-scheduler logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:23:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28768763-r9x7k AddedInterface Add eth0 [10.129.0.59/23] from ovn-kubernetes logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:23:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28768763-r9x7k.spec.containers{my-jaeger-es-rollover} Pulled Container image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:b3897239050ab12ae12c28bb51d955438334aad957dceb83a53cf2e8a466b69b" already present on machine kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:23:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28768763-r9x7k.spec.containers{my-jaeger-es-rollover} Created Created container my-jaeger-es-rollover kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:23:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28768763-r9x7k.spec.containers{my-jaeger-es-rollover} Started Started container my-jaeger-es-rollover kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:23:00 +0000 UTC Normal Job.batch my-jaeger-es-rollover-28768763 SuccessfulCreate Created pod: my-jaeger-es-rollover-28768763-r9x7k job-controller logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:23:00 +0000 UTC Normal CronJob.batch my-jaeger-es-rollover SuccessfulCreate Created job my-jaeger-es-rollover-28768763 cronjob-controller logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:23:01 +0000 UTC Normal Pod my-jaeger-es-lookback-28768763-qbwms AddedInterface Add eth0 [10.129.0.58/23] from ovn-kubernetes logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:23:01 +0000 UTC Normal Pod my-jaeger-es-lookback-28768763-qbwms.spec.containers{my-jaeger-es-lookback} Pulled Container image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:b3897239050ab12ae12c28bb51d955438334aad957dceb83a53cf2e8a466b69b" already present on machine kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:23:01 +0000 UTC Normal Pod my-jaeger-es-lookback-28768763-qbwms.spec.containers{my-jaeger-es-lookback} Created Created container my-jaeger-es-lookback kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:23:01 +0000 UTC Normal Pod my-jaeger-es-lookback-28768763-qbwms.spec.containers{my-jaeger-es-lookback} Started Started container my-jaeger-es-lookback kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:23:03 +0000 UTC Normal Job.batch my-jaeger-es-lookback-28768763 Completed Job completed job-controller logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:23:03 +0000 UTC Normal CronJob.batch my-jaeger-es-lookback SawCompletedJob Saw completed job: my-jaeger-es-lookback-28768763, status: Complete cronjob-controller logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:23:03 +0000 UTC Normal Job.batch my-jaeger-es-rollover-28768763 Completed Job completed job-controller logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:23:03 +0000 UTC Normal CronJob.batch my-jaeger-es-rollover SawCompletedJob Saw completed job: my-jaeger-es-rollover-28768763, status: Complete cronjob-controller logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:23:06 +0000 UTC Normal Pod 05-check-indices-zgw9d Binding Scheduled Successfully assigned kuttl-test-sharp-ocelot/05-check-indices-zgw9d to ip-10-0-54-110.us-west-2.compute.internal default-scheduler logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:23:06 +0000 UTC Normal Pod 05-check-indices-zgw9d AddedInterface Add eth0 [10.129.0.60/23] from ovn-kubernetes logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:23:06 +0000 UTC Normal Job.batch 05-check-indices SuccessfulCreate Created pod: 05-check-indices-zgw9d job-controller logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:23:07 +0000 UTC Normal Pod 05-check-indices-zgw9d.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:23:07 +0000 UTC Normal Pod 05-check-indices-zgw9d.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 873ms (874ms including waiting) kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:23:07 +0000 UTC Normal Pod 05-check-indices-zgw9d.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:23:07 +0000 UTC Normal Pod 05-check-indices-zgw9d.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:23:10 +0000 UTC Normal Job.batch 05-check-indices Completed Job completed job-controller logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:23:12 +0000 UTC Normal Pod 06-check-indices-dddb2 Binding Scheduled Successfully assigned kuttl-test-sharp-ocelot/06-check-indices-dddb2 to ip-10-0-54-110.us-west-2.compute.internal default-scheduler logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:23:12 +0000 UTC Normal Pod 06-check-indices-dddb2 AddedInterface Add eth0 [10.129.0.61/23] from ovn-kubernetes logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:23:12 +0000 UTC Normal Pod 06-check-indices-dddb2.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:23:12 +0000 UTC Normal Job.batch 06-check-indices SuccessfulCreate Created pod: 06-check-indices-dddb2 job-controller logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:23:13 +0000 UTC Normal Pod 06-check-indices-dddb2.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 794ms (794ms including waiting) kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:23:13 +0000 UTC Normal Pod 06-check-indices-dddb2.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:23:13 +0000 UTC Normal Pod 06-check-indices-dddb2.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:23:16 | es-rollover-autoprov | 2024-09-12 07:23:15 +0000 UTC Normal Job.batch 06-check-indices Completed Job completed job-controller logger.go:42: 07:23:17 | es-rollover-autoprov | Deleting namespace: kuttl-test-sharp-ocelot === CONT kuttl/harness/es-increasing-replicas logger.go:42: 07:23:25 | es-increasing-replicas | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:23:25 | es-increasing-replicas | Ignoring check-es-nodes.sh as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:23:25 | es-increasing-replicas | Ignoring openshift-check-es-nodes.yaml.template as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:23:25 | es-increasing-replicas | Creating namespace: kuttl-test-awaited-swan logger.go:42: 07:23:25 | es-increasing-replicas/1-install | starting test step 1-install logger.go:42: 07:23:25 | es-increasing-replicas/1-install | Jaeger:kuttl-test-awaited-swan/simple-prod created logger.go:42: 07:24:04 | es-increasing-replicas/1-install | test step completed 1-install logger.go:42: 07:24:04 | es-increasing-replicas/2-install | starting test step 2-install logger.go:42: 07:24:05 | es-increasing-replicas/2-install | Jaeger:kuttl-test-awaited-swan/simple-prod updated logger.go:42: 07:24:11 | es-increasing-replicas/2-install | test step completed 2-install logger.go:42: 07:24:11 | es-increasing-replicas/3-smoke-test | starting test step 3-smoke-test logger.go:42: 07:24:11 | es-increasing-replicas/3-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE simple-prod /dev/null] logger.go:42: 07:24:13 | es-increasing-replicas/3-smoke-test | Warning: resource jaegers/simple-prod is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:24:20 | es-increasing-replicas/3-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:24:21 | es-increasing-replicas/3-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:24:21 | es-increasing-replicas/3-smoke-test | job.batch/report-span created logger.go:42: 07:24:21 | es-increasing-replicas/3-smoke-test | job.batch/check-span created logger.go:42: 07:24:34 | es-increasing-replicas/3-smoke-test | test step completed 3-smoke-test logger.go:42: 07:24:34 | es-increasing-replicas/4-install | starting test step 4-install logger.go:42: 07:24:35 | es-increasing-replicas/4-install | Jaeger:kuttl-test-awaited-swan/simple-prod updated logger.go:42: 07:24:35 | es-increasing-replicas/4-install | test step completed 4-install logger.go:42: 07:24:35 | es-increasing-replicas/5-check-es-nodes | starting test step 5-check-es-nodes logger.go:42: 07:24:35 | es-increasing-replicas/5-check-es-nodes | running command: [sh -c ./check-es-nodes.sh $NAMESPACE] logger.go:42: 07:24:35 | es-increasing-replicas/5-check-es-nodes | Checking if the number of ES instances is the expected logger.go:42: 07:24:35 | es-increasing-replicas/5-check-es-nodes | false logger.go:42: 07:24:35 | es-increasing-replicas/5-check-es-nodes | Error: no matches found logger.go:42: 07:24:40 | es-increasing-replicas/5-check-es-nodes | Checking if the number of ES instances is the expected logger.go:42: 07:24:40 | es-increasing-replicas/5-check-es-nodes | true logger.go:42: 07:24:41 | es-increasing-replicas/5-check-es-nodes | test step completed 5-check-es-nodes logger.go:42: 07:24:41 | es-increasing-replicas | es-increasing-replicas events from ns kuttl-test-awaited-swan: logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:23:33 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestawaitedswansimpleprod-1-6f8f4547c SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestawaitedswansimpleprod-1-6f8f454qvfrb replicaset-controller logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:23:33 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestawaitedswansimpleprod-1-6f8f454qvfrb Binding Scheduled Successfully assigned kuttl-test-awaited-swan/elasticsearch-cdm-kuttltestawaitedswansimpleprod-1-6f8f454qvfrb to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:23:33 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestawaitedswansimpleprod-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestawaitedswansimpleprod-1-6f8f4547c to 1 deployment-controller logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:23:34 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestawaitedswansimpleprod-1-6f8f454qvfrb AddedInterface Add eth0 [10.130.0.57/23] from ovn-kubernetes logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:23:34 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestawaitedswansimpleprod-1-6f8f454qvfrb.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:bc97b8e13087050dfb1a75b02c5b780cbb8fb12a849a655169d072ba8bbf42b4" already present on machine kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:23:34 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestawaitedswansimpleprod-1-6f8f454qvfrb.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:23:34 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestawaitedswansimpleprod-1-6f8f454qvfrb.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:23:34 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestawaitedswansimpleprod-1-6f8f454qvfrb.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:1d1d7d47b616995d18692f3a2d6232d7e1a0f41bc1503d87100f31939e080a78" already present on machine kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:23:34 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestawaitedswansimpleprod-1-6f8f454qvfrb.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:23:34 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestawaitedswansimpleprod-1-6f8f454qvfrb.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:23:44 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestawaitedswansimpleprod-1-6f8f454qvfrb.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:23:49 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestawaitedswansimpleprod-1-6f8f454qvfrb.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:00 +0000 UTC Normal Pod simple-prod-collector-f5f5c99f7-xnh4f Binding Scheduled Successfully assigned kuttl-test-awaited-swan/simple-prod-collector-f5f5c99f7-xnh4f to ip-10-0-54-110.us-west-2.compute.internal default-scheduler logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:00 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-f5f5c99f7 SuccessfulCreate Created pod: simple-prod-collector-f5f5c99f7-xnh4f replicaset-controller logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:00 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-f5f5c99f7 to 1 deployment-controller logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:00 +0000 UTC Normal Pod simple-prod-query-79fb569b44-5kch7 Binding Scheduled Successfully assigned kuttl-test-awaited-swan/simple-prod-query-79fb569b44-5kch7 to ip-10-0-52-42.us-west-2.compute.internal default-scheduler logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:00 +0000 UTC Normal ReplicaSet.apps simple-prod-query-79fb569b44 SuccessfulCreate Created pod: simple-prod-query-79fb569b44-5kch7 replicaset-controller logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:00 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-79fb569b44 to 1 deployment-controller logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:01 +0000 UTC Normal Pod simple-prod-collector-f5f5c99f7-xnh4f AddedInterface Add eth0 [10.129.0.62/23] from ovn-kubernetes logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:01 +0000 UTC Normal Pod simple-prod-collector-f5f5c99f7-xnh4f.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:4bab63290ff52e4f6328009f9c8b2c4496b230e9f8a98eac01736a66a291ff6c" already present on machine kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:01 +0000 UTC Normal Pod simple-prod-collector-f5f5c99f7-xnh4f.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:01 +0000 UTC Normal Pod simple-prod-collector-f5f5c99f7-xnh4f.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:01 +0000 UTC Normal Pod simple-prod-query-79fb569b44-5kch7 AddedInterface Add eth0 [10.128.0.59/23] from ovn-kubernetes logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:01 +0000 UTC Normal Pod simple-prod-query-79fb569b44-5kch7.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a9186dcd910256c0f464b0a3928844a01de166a10c186c97ef4581bf288c23cb" already present on machine kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:01 +0000 UTC Normal Pod simple-prod-query-79fb569b44-5kch7.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:01 +0000 UTC Normal Pod simple-prod-query-79fb569b44-5kch7.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:01 +0000 UTC Normal Pod simple-prod-query-79fb569b44-5kch7.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:01 +0000 UTC Normal Pod simple-prod-query-79fb569b44-5kch7.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:01 +0000 UTC Normal Pod simple-prod-query-79fb569b44-5kch7.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:01 +0000 UTC Normal Pod simple-prod-query-79fb569b44-5kch7.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" already present on machine kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:01 +0000 UTC Normal Pod simple-prod-query-79fb569b44-5kch7.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:01 +0000 UTC Normal Pod simple-prod-query-79fb569b44-5kch7.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:05 +0000 UTC Normal Pod simple-prod-collector-f5f5c99f7-tktpn Binding Scheduled Successfully assigned kuttl-test-awaited-swan/simple-prod-collector-f5f5c99f7-tktpn to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:05 +0000 UTC Normal Pod simple-prod-collector-f5f5c99f7-tktpn AddedInterface Add eth0 [10.130.0.58/23] from ovn-kubernetes logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:05 +0000 UTC Normal Pod simple-prod-collector-f5f5c99f7-tktpn.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:4bab63290ff52e4f6328009f9c8b2c4496b230e9f8a98eac01736a66a291ff6c" already present on machine kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:05 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-f5f5c99f7 SuccessfulCreate Created pod: simple-prod-collector-f5f5c99f7-tktpn replicaset-controller logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:05 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-f5f5c99f7 to 2 from 1 deployment-controller logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:05 +0000 UTC Normal Pod simple-prod-query-79fb569b44-gfs6f Binding Scheduled Successfully assigned kuttl-test-awaited-swan/simple-prod-query-79fb569b44-gfs6f to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:05 +0000 UTC Normal Pod simple-prod-query-79fb569b44-gfs6f AddedInterface Add eth0 [10.130.0.59/23] from ovn-kubernetes logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:05 +0000 UTC Normal Pod simple-prod-query-79fb569b44-gfs6f.spec.containers{jaeger-query} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a9186dcd910256c0f464b0a3928844a01de166a10c186c97ef4581bf288c23cb" kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:05 +0000 UTC Normal ReplicaSet.apps simple-prod-query-79fb569b44 SuccessfulCreate Created pod: simple-prod-query-79fb569b44-gfs6f replicaset-controller logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:05 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-79fb569b44 to 2 from 1 deployment-controller logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:06 +0000 UTC Normal Pod simple-prod-collector-f5f5c99f7-tktpn.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:06 +0000 UTC Normal Pod simple-prod-collector-f5f5c99f7-tktpn.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:09 +0000 UTC Normal Pod simple-prod-query-79fb569b44-gfs6f.spec.containers{jaeger-query} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a9186dcd910256c0f464b0a3928844a01de166a10c186c97ef4581bf288c23cb" in 3.217s (3.217s including waiting) kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:09 +0000 UTC Normal Pod simple-prod-query-79fb569b44-gfs6f.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:09 +0000 UTC Normal Pod simple-prod-query-79fb569b44-gfs6f.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:09 +0000 UTC Normal Pod simple-prod-query-79fb569b44-gfs6f.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:09 +0000 UTC Normal Pod simple-prod-query-79fb569b44-gfs6f.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:09 +0000 UTC Normal Pod simple-prod-query-79fb569b44-gfs6f.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:09 +0000 UTC Normal Pod simple-prod-query-79fb569b44-gfs6f.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" already present on machine kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:09 +0000 UTC Normal Pod simple-prod-query-79fb569b44-gfs6f.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:09 +0000 UTC Normal Pod simple-prod-query-79fb569b44-gfs6f.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:14 +0000 UTC Normal Pod simple-prod-query-79fb569b44-5kch7.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:14 +0000 UTC Normal Pod simple-prod-query-79fb569b44-5kch7.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:14 +0000 UTC Normal Pod simple-prod-query-79fb569b44-5kch7.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:14 +0000 UTC Normal Pod simple-prod-query-79fb569b44-gfs6f.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:14 +0000 UTC Normal Pod simple-prod-query-79fb569b44-gfs6f.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:14 +0000 UTC Normal Pod simple-prod-query-79fb569b44-gfs6f.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:14 +0000 UTC Normal ReplicaSet.apps simple-prod-query-79fb569b44 SuccessfulDelete Deleted pod: simple-prod-query-79fb569b44-gfs6f replicaset-controller logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:14 +0000 UTC Normal ReplicaSet.apps simple-prod-query-79fb569b44 SuccessfulDelete Deleted pod: simple-prod-query-79fb569b44-5kch7 replicaset-controller logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:14 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled down replica set simple-prod-query-79fb569b44 to 0 from 2 deployment-controller logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:15 +0000 UTC Normal Pod simple-prod-query-dc896d7bf-bxvww Binding Scheduled Successfully assigned kuttl-test-awaited-swan/simple-prod-query-dc896d7bf-bxvww to ip-10-0-52-42.us-west-2.compute.internal default-scheduler logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:15 +0000 UTC Normal Pod simple-prod-query-dc896d7bf-g5c7n Binding Scheduled Successfully assigned kuttl-test-awaited-swan/simple-prod-query-dc896d7bf-g5c7n to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:15 +0000 UTC Normal ReplicaSet.apps simple-prod-query-dc896d7bf SuccessfulCreate Created pod: simple-prod-query-dc896d7bf-bxvww replicaset-controller logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:15 +0000 UTC Normal ReplicaSet.apps simple-prod-query-dc896d7bf SuccessfulCreate Created pod: simple-prod-query-dc896d7bf-g5c7n replicaset-controller logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:15 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-dc896d7bf to 2 deployment-controller logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:16 +0000 UTC Normal Pod simple-prod-query-dc896d7bf-bxvww AddedInterface Add eth0 [10.128.0.60/23] from ovn-kubernetes logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:16 +0000 UTC Normal Pod simple-prod-query-dc896d7bf-bxvww.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a9186dcd910256c0f464b0a3928844a01de166a10c186c97ef4581bf288c23cb" already present on machine kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:16 +0000 UTC Normal Pod simple-prod-query-dc896d7bf-bxvww.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:16 +0000 UTC Normal Pod simple-prod-query-dc896d7bf-bxvww.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:16 +0000 UTC Normal Pod simple-prod-query-dc896d7bf-bxvww.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:16 +0000 UTC Normal Pod simple-prod-query-dc896d7bf-bxvww.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:16 +0000 UTC Normal Pod simple-prod-query-dc896d7bf-bxvww.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:16 +0000 UTC Normal Pod simple-prod-query-dc896d7bf-bxvww.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" already present on machine kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:16 +0000 UTC Normal Pod simple-prod-query-dc896d7bf-bxvww.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:16 +0000 UTC Normal Pod simple-prod-query-dc896d7bf-bxvww.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:16 +0000 UTC Normal Pod simple-prod-query-dc896d7bf-g5c7n AddedInterface Add eth0 [10.130.0.60/23] from ovn-kubernetes logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:16 +0000 UTC Normal Pod simple-prod-query-dc896d7bf-g5c7n.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a9186dcd910256c0f464b0a3928844a01de166a10c186c97ef4581bf288c23cb" already present on machine kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:16 +0000 UTC Normal Pod simple-prod-query-dc896d7bf-g5c7n.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:16 +0000 UTC Normal Pod simple-prod-query-dc896d7bf-g5c7n.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:16 +0000 UTC Normal Pod simple-prod-query-dc896d7bf-g5c7n.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:16 +0000 UTC Normal Pod simple-prod-query-dc896d7bf-g5c7n.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:16 +0000 UTC Normal Pod simple-prod-query-dc896d7bf-g5c7n.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:16 +0000 UTC Normal Pod simple-prod-query-dc896d7bf-g5c7n.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" already present on machine kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:16 +0000 UTC Normal Pod simple-prod-query-dc896d7bf-g5c7n.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:16 +0000 UTC Normal Pod simple-prod-query-dc896d7bf-g5c7n.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:21 +0000 UTC Normal Pod check-span-92lch Binding Scheduled Successfully assigned kuttl-test-awaited-swan/check-span-92lch to ip-10-0-54-110.us-west-2.compute.internal default-scheduler logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:21 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-92lch job-controller logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:21 +0000 UTC Normal Pod report-span-d9fx2 Binding Scheduled Successfully assigned kuttl-test-awaited-swan/report-span-d9fx2 to ip-10-0-54-110.us-west-2.compute.internal default-scheduler logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:21 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-d9fx2 job-controller logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:22 +0000 UTC Normal Pod check-span-92lch AddedInterface Add eth0 [10.129.0.64/23] from ovn-kubernetes logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:22 +0000 UTC Normal Pod check-span-92lch.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:22 +0000 UTC Normal Pod report-span-d9fx2 AddedInterface Add eth0 [10.129.0.63/23] from ovn-kubernetes logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:22 +0000 UTC Normal Pod report-span-d9fx2.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:23 +0000 UTC Normal Pod check-span-92lch.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 821ms (822ms including waiting) kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:23 +0000 UTC Normal Pod check-span-92lch.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:23 +0000 UTC Normal Pod check-span-92lch.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:23 +0000 UTC Normal Pod report-span-d9fx2.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 875ms (875ms including waiting) kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:23 +0000 UTC Normal Pod report-span-d9fx2.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:23 +0000 UTC Normal Pod report-span-d9fx2.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:34 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:35 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestawaitedswansimpleprod-2-74fdd4c969 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestawaitedswansimpleprod-2-74fdd4cznxc2 replicaset-controller logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:35 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestawaitedswansimpleprod-2-74fdd4cznxc2 Binding Scheduled Successfully assigned kuttl-test-awaited-swan/elasticsearch-cdm-kuttltestawaitedswansimpleprod-2-74fdd4cznxc2 to ip-10-0-52-42.us-west-2.compute.internal default-scheduler logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:35 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestawaitedswansimpleprod-2 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestawaitedswansimpleprod-2-74fdd4c969 to 1 deployment-controller logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:35 +0000 UTC Normal Pod simple-prod-collector-f5f5c99f7-tktpn.spec.containers{jaeger-collector} Killing Stopping container jaeger-collector kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:35 +0000 UTC Warning Pod simple-prod-collector-f5f5c99f7-tktpn.spec.containers{jaeger-collector} Unhealthy Readiness probe failed: Get "http://10.130.0.58:14269/": dial tcp 10.130.0.58:14269: connect: connection refused kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:35 +0000 UTC Normal Pod simple-prod-collector-f5f5c99f7-xnh4f.spec.containers{jaeger-collector} Killing Stopping container jaeger-collector kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:35 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-f5f5c99f7 SuccessfulDelete Deleted pod: simple-prod-collector-f5f5c99f7-tktpn replicaset-controller logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:35 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-f5f5c99f7 SuccessfulDelete Deleted pod: simple-prod-collector-f5f5c99f7-xnh4f replicaset-controller logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:35 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled down replica set simple-prod-collector-f5f5c99f7 to 0 from 2 deployment-controller logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:35 +0000 UTC Normal Pod simple-prod-query-dc896d7bf-bxvww.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:35 +0000 UTC Normal Pod simple-prod-query-dc896d7bf-bxvww.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:35 +0000 UTC Normal Pod simple-prod-query-dc896d7bf-bxvww.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:35 +0000 UTC Normal Pod simple-prod-query-dc896d7bf-g5c7n.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:35 +0000 UTC Normal Pod simple-prod-query-dc896d7bf-g5c7n.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:35 +0000 UTC Normal Pod simple-prod-query-dc896d7bf-g5c7n.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:35 +0000 UTC Normal ReplicaSet.apps simple-prod-query-dc896d7bf SuccessfulDelete Deleted pod: simple-prod-query-dc896d7bf-g5c7n replicaset-controller logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:35 +0000 UTC Normal ReplicaSet.apps simple-prod-query-dc896d7bf SuccessfulDelete Deleted pod: simple-prod-query-dc896d7bf-bxvww replicaset-controller logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:35 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled down replica set simple-prod-query-dc896d7bf to 0 from 2 deployment-controller logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:36 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestawaitedswansimpleprod-2-74fdd4cznxc2 AddedInterface Add eth0 [10.128.0.61/23] from ovn-kubernetes logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:36 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestawaitedswansimpleprod-2-74fdd4cznxc2.spec.containers{elasticsearch} Pulling Pulling image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:bc97b8e13087050dfb1a75b02c5b780cbb8fb12a849a655169d072ba8bbf42b4" kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:36 +0000 UTC Normal Pod simple-prod-collector-689689c978-77gg8 Binding Scheduled Successfully assigned kuttl-test-awaited-swan/simple-prod-collector-689689c978-77gg8 to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:36 +0000 UTC Normal Pod simple-prod-collector-689689c978-77gg8 AddedInterface Add eth0 [10.130.0.61/23] from ovn-kubernetes logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:36 +0000 UTC Normal Pod simple-prod-collector-689689c978-77gg8.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:4bab63290ff52e4f6328009f9c8b2c4496b230e9f8a98eac01736a66a291ff6c" already present on machine kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:36 +0000 UTC Normal Pod simple-prod-collector-689689c978-f4p6h Binding Scheduled Successfully assigned kuttl-test-awaited-swan/simple-prod-collector-689689c978-f4p6h to ip-10-0-54-110.us-west-2.compute.internal default-scheduler logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:36 +0000 UTC Normal Pod simple-prod-collector-689689c978-f4p6h AddedInterface Add eth0 [10.129.0.65/23] from ovn-kubernetes logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:36 +0000 UTC Normal Pod simple-prod-collector-689689c978-f4p6h.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:4bab63290ff52e4f6328009f9c8b2c4496b230e9f8a98eac01736a66a291ff6c" already present on machine kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:36 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-689689c978 SuccessfulCreate Created pod: simple-prod-collector-689689c978-f4p6h replicaset-controller logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:36 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-689689c978 SuccessfulCreate Created pod: simple-prod-collector-689689c978-77gg8 replicaset-controller logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:36 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-689689c978 to 2 deployment-controller logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:36 +0000 UTC Normal Pod simple-prod-query-5dd76898fb-hjp5j Binding Scheduled Successfully assigned kuttl-test-awaited-swan/simple-prod-query-5dd76898fb-hjp5j to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:36 +0000 UTC Normal ReplicaSet.apps simple-prod-query-5dd76898fb SuccessfulCreate Created pod: simple-prod-query-5dd76898fb-hjp5j replicaset-controller logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:36 +0000 UTC Warning Pod simple-prod-query-dc896d7bf-bxvww.spec.containers{jaeger-query} Unhealthy Readiness probe failed: Get "http://10.128.0.60:16687/": dial tcp 10.128.0.60:16687: connect: connection refused kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:36 +0000 UTC Warning Pod simple-prod-query-dc896d7bf-bxvww.spec.containers{jaeger-agent} Unhealthy Readiness probe failed: Get "http://10.128.0.60:14271/": dial tcp 10.128.0.60:14271: connect: connection refused kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:36 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-5dd76898fb to 2 deployment-controller logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:37 +0000 UTC Normal Pod simple-prod-collector-689689c978-77gg8.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:37 +0000 UTC Normal Pod simple-prod-collector-689689c978-77gg8.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:37 +0000 UTC Normal Pod simple-prod-collector-689689c978-f4p6h.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:37 +0000 UTC Normal Pod simple-prod-collector-689689c978-f4p6h.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:37 +0000 UTC Normal Pod simple-prod-query-5dd76898fb-hjp5j AddedInterface Add eth0 [10.130.0.62/23] from ovn-kubernetes logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:37 +0000 UTC Normal Pod simple-prod-query-5dd76898fb-hjp5j.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a9186dcd910256c0f464b0a3928844a01de166a10c186c97ef4581bf288c23cb" already present on machine kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:37 +0000 UTC Normal Pod simple-prod-query-5dd76898fb-hjp5j.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:37 +0000 UTC Normal Pod simple-prod-query-5dd76898fb-hjp5j.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:37 +0000 UTC Normal Pod simple-prod-query-5dd76898fb-hjp5j.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:37 +0000 UTC Normal Pod simple-prod-query-5dd76898fb-hjp5j.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:37 +0000 UTC Normal Pod simple-prod-query-5dd76898fb-hjp5j.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:37 +0000 UTC Normal Pod simple-prod-query-5dd76898fb-hjp5j.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" already present on machine kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:37 +0000 UTC Normal Pod simple-prod-query-5dd76898fb-hjp5j.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:37 +0000 UTC Normal Pod simple-prod-query-5dd76898fb-hjp5j.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:37 +0000 UTC Normal Pod simple-prod-query-5dd76898fb-mhpcp Binding Scheduled Successfully assigned kuttl-test-awaited-swan/simple-prod-query-5dd76898fb-mhpcp to ip-10-0-52-42.us-west-2.compute.internal default-scheduler logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:37 +0000 UTC Normal Pod simple-prod-query-5dd76898fb-mhpcp AddedInterface Add eth0 [10.128.0.62/23] from ovn-kubernetes logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:37 +0000 UTC Normal Pod simple-prod-query-5dd76898fb-mhpcp.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a9186dcd910256c0f464b0a3928844a01de166a10c186c97ef4581bf288c23cb" already present on machine kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:37 +0000 UTC Normal Pod simple-prod-query-5dd76898fb-mhpcp.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:37 +0000 UTC Normal Pod simple-prod-query-5dd76898fb-mhpcp.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:37 +0000 UTC Normal Pod simple-prod-query-5dd76898fb-mhpcp.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:37 +0000 UTC Normal Pod simple-prod-query-5dd76898fb-mhpcp.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:37 +0000 UTC Normal ReplicaSet.apps simple-prod-query-5dd76898fb SuccessfulCreate Created pod: simple-prod-query-5dd76898fb-mhpcp replicaset-controller logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:38 +0000 UTC Normal Pod simple-prod-query-5dd76898fb-mhpcp.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:38 +0000 UTC Normal Pod simple-prod-query-5dd76898fb-mhpcp.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" already present on machine kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:38 +0000 UTC Normal Pod simple-prod-query-5dd76898fb-mhpcp.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:24:41 | es-increasing-replicas | 2024-09-12 07:24:38 +0000 UTC Normal Pod simple-prod-query-5dd76898fb-mhpcp.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:24:41 | es-increasing-replicas | Deleting namespace: kuttl-test-awaited-swan === CONT kuttl/harness/es-index-cleaner-autoprov logger.go:42: 07:25:21 | es-index-cleaner-autoprov | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:25:21 | es-index-cleaner-autoprov | Creating namespace: kuttl-test-stirred-chipmunk logger.go:42: 07:25:21 | es-index-cleaner-autoprov/1-install | starting test step 1-install logger.go:42: 07:25:21 | es-index-cleaner-autoprov/1-install | Jaeger:kuttl-test-stirred-chipmunk/test-es-index-cleaner-with-prefix created logger.go:42: 07:25:56 | es-index-cleaner-autoprov/1-install | test step completed 1-install logger.go:42: 07:25:56 | es-index-cleaner-autoprov/2-report-spans | starting test step 2-report-spans logger.go:42: 07:25:56 | es-index-cleaner-autoprov/2-report-spans | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE test-es-index-cleaner-with-prefix /dev/null] logger.go:42: 07:25:58 | es-index-cleaner-autoprov/2-report-spans | Warning: resource jaegers/test-es-index-cleaner-with-prefix is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:26:05 | es-index-cleaner-autoprov/2-report-spans | running command: [sh -c DAYS=5 ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JOB_NUMBER=00 JAEGER_COLLECTOR_ENDPOINT=http://test-es-index-cleaner-with-prefix-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://test-es-index-cleaner-with-prefix-query MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/report-spans.yaml.template -o report-span-00-job.yaml] logger.go:42: 07:26:06 | es-index-cleaner-autoprov/2-report-spans | running command: [sh -c kubectl apply -f report-span-00-job.yaml -n $NAMESPACE] logger.go:42: 07:26:06 | es-index-cleaner-autoprov/2-report-spans | job.batch/00-report-span created logger.go:42: 07:26:46 | es-index-cleaner-autoprov/2-report-spans | test step completed 2-report-spans logger.go:42: 07:26:46 | es-index-cleaner-autoprov/3-install | starting test step 3-install logger.go:42: 07:26:46 | es-index-cleaner-autoprov/3-install | Jaeger:kuttl-test-stirred-chipmunk/test-es-index-cleaner-with-prefix updated logger.go:42: 07:26:46 | es-index-cleaner-autoprov/3-install | test step completed 3-install logger.go:42: 07:26:46 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | starting test step 4-wait-es-index-cleaner logger.go:42: 07:26:46 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | running command: [sh -c go run ../../../../cmd-utils/wait-cronjob/main.go --cronjob test-es-index-cleaner-with-prefix-es-index-cleaner --namespace $NAMESPACE] logger.go:42: 07:26:47 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | time="2024-09-12T07:26:47Z" level=debug msg="Checking if the test-es-index-cleaner-with-prefix-es-index-cleaner CronJob exists" logger.go:42: 07:26:47 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | time="2024-09-12T07:26:47Z" level=debug msg="No BatchV1beta1/Cronjobs were found" logger.go:42: 07:26:47 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | time="2024-09-12T07:26:47Z" level=info msg="Cronjob test-es-index-cleaner-with-prefix-es-index-cleaner found successfully" logger.go:42: 07:26:47 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | time="2024-09-12T07:26:47Z" level=debug msg="Waiting for the next scheduled job from test-es-index-cleaner-with-prefix-es-index-cleaner cronjob" logger.go:42: 07:26:47 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | time="2024-09-12T07:26:47Z" level=debug msg="Waiting for next job from test-es-index-cleaner-with-prefix-es-index-cleaner to succeed" logger.go:42: 07:26:57 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | time="2024-09-12T07:26:57Z" level=debug msg="Waiting for next job from test-es-index-cleaner-with-prefix-es-index-cleaner to succeed" logger.go:42: 07:27:07 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | time="2024-09-12T07:27:07Z" level=debug msg="Waiting for next job from test-es-index-cleaner-with-prefix-es-index-cleaner to succeed" logger.go:42: 07:27:17 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | time="2024-09-12T07:27:17Z" level=info msg="Job of owner test-es-index-cleaner-with-prefix-es-index-cleaner succeeded after test-es-index-cleaner-with-prefix-es-index-cleaner 30.056128989s" logger.go:42: 07:27:17 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | test step completed 4-wait-es-index-cleaner logger.go:42: 07:27:17 | es-index-cleaner-autoprov/5-install | starting test step 5-install logger.go:42: 07:27:18 | es-index-cleaner-autoprov/5-install | Jaeger:kuttl-test-stirred-chipmunk/test-es-index-cleaner-with-prefix updated logger.go:42: 07:27:18 | es-index-cleaner-autoprov/5-install | test step completed 5-install logger.go:42: 07:27:18 | es-index-cleaner-autoprov/6-check-indices | starting test step 6-check-indices logger.go:42: 07:27:18 | es-index-cleaner-autoprov/6-check-indices | Job:kuttl-test-stirred-chipmunk/00-check-indices created logger.go:42: 07:27:23 | es-index-cleaner-autoprov/6-check-indices | test step completed 6-check-indices logger.go:42: 07:27:24 | es-index-cleaner-autoprov | es-index-cleaner-autoprov events from ns kuttl-test-stirred-chipmunk: logger.go:42: 07:27:24 | es-index-cleaner-autoprov | 2024-09-12 07:25:26 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttlteststirredchipmunktestesindexc-1-f875fbd5c SuccessfulCreate Created pod: elasticsearch-cdm-kuttlteststirredchipmunktestesindexc-1-fzsntr replicaset-controller logger.go:42: 07:27:24 | es-index-cleaner-autoprov | 2024-09-12 07:25:26 +0000 UTC Normal Pod elasticsearch-cdm-kuttlteststirredchipmunktestesindexc-1-fzsntr Binding Scheduled Successfully assigned kuttl-test-stirred-chipmunk/elasticsearch-cdm-kuttlteststirredchipmunktestesindexc-1-fzsntr to ip-10-0-52-42.us-west-2.compute.internal default-scheduler logger.go:42: 07:27:24 | es-index-cleaner-autoprov | 2024-09-12 07:25:26 +0000 UTC Normal Pod elasticsearch-cdm-kuttlteststirredchipmunktestesindexc-1-fzsntr AddedInterface Add eth0 [10.128.0.63/23] from ovn-kubernetes logger.go:42: 07:27:24 | es-index-cleaner-autoprov | 2024-09-12 07:25:26 +0000 UTC Normal Pod elasticsearch-cdm-kuttlteststirredchipmunktestesindexc-1-fzsntr.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:bc97b8e13087050dfb1a75b02c5b780cbb8fb12a849a655169d072ba8bbf42b4" already present on machine kubelet logger.go:42: 07:27:24 | es-index-cleaner-autoprov | 2024-09-12 07:25:26 +0000 UTC Normal Pod elasticsearch-cdm-kuttlteststirredchipmunktestesindexc-1-fzsntr.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:27:24 | es-index-cleaner-autoprov | 2024-09-12 07:25:26 +0000 UTC Normal Pod elasticsearch-cdm-kuttlteststirredchipmunktestesindexc-1-fzsntr.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:27:24 | es-index-cleaner-autoprov | 2024-09-12 07:25:26 +0000 UTC Normal Pod elasticsearch-cdm-kuttlteststirredchipmunktestesindexc-1-fzsntr.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:1d1d7d47b616995d18692f3a2d6232d7e1a0f41bc1503d87100f31939e080a78" already present on machine kubelet logger.go:42: 07:27:24 | es-index-cleaner-autoprov | 2024-09-12 07:25:26 +0000 UTC Normal Pod elasticsearch-cdm-kuttlteststirredchipmunktestesindexc-1-fzsntr.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:27:24 | es-index-cleaner-autoprov | 2024-09-12 07:25:26 +0000 UTC Normal Pod elasticsearch-cdm-kuttlteststirredchipmunktestesindexc-1-fzsntr.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:27:24 | es-index-cleaner-autoprov | 2024-09-12 07:25:26 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttlteststirredchipmunktestesindexc-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttlteststirredchipmunktestesindexc-1-f875fbd5c to 1 deployment-controller logger.go:42: 07:27:24 | es-index-cleaner-autoprov | 2024-09-12 07:25:36 +0000 UTC Warning Pod elasticsearch-cdm-kuttlteststirredchipmunktestesindexc-1-fzsntr.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 07:27:24 | es-index-cleaner-autoprov | 2024-09-12 07:25:41 +0000 UTC Warning Pod elasticsearch-cdm-kuttlteststirredchipmunktestesindexc-1-fzsntr.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:27:24 | es-index-cleaner-autoprov | 2024-09-12 07:25:53 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-collector-5484748d55-227tc Binding Scheduled Successfully assigned kuttl-test-stirred-chipmunk/test-es-index-cleaner-with-prefix-collector-5484748d55-227tc to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:27:24 | es-index-cleaner-autoprov | 2024-09-12 07:25:53 +0000 UTC Warning Pod test-es-index-cleaner-with-prefix-collector-5484748d55-227tc FailedMount MountVolume.SetUp failed for volume "test-es-index-cleaner-with-prefix-collector-tls-config-volume" : secret "test-es-index-cleaner-with-prefix-collector-headless-tls" not found kubelet logger.go:42: 07:27:24 | es-index-cleaner-autoprov | 2024-09-12 07:25:53 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-collector-5484748d55-227tc AddedInterface Add eth0 [10.130.0.63/23] from ovn-kubernetes logger.go:42: 07:27:24 | es-index-cleaner-autoprov | 2024-09-12 07:25:53 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-collector-5484748d55-227tc.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:4bab63290ff52e4f6328009f9c8b2c4496b230e9f8a98eac01736a66a291ff6c" already present on machine kubelet logger.go:42: 07:27:24 | es-index-cleaner-autoprov | 2024-09-12 07:25:53 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-collector-5484748d55-227tc.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:27:24 | es-index-cleaner-autoprov | 2024-09-12 07:25:53 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-collector-5484748d55-227tc.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:27:24 | es-index-cleaner-autoprov | 2024-09-12 07:25:53 +0000 UTC Normal ReplicaSet.apps test-es-index-cleaner-with-prefix-collector-5484748d55 SuccessfulCreate Created pod: test-es-index-cleaner-with-prefix-collector-5484748d55-227tc replicaset-controller logger.go:42: 07:27:24 | es-index-cleaner-autoprov | 2024-09-12 07:25:53 +0000 UTC Normal Deployment.apps test-es-index-cleaner-with-prefix-collector ScalingReplicaSet Scaled up replica set test-es-index-cleaner-with-prefix-collector-5484748d55 to 1 deployment-controller logger.go:42: 07:27:24 | es-index-cleaner-autoprov | 2024-09-12 07:25:53 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-55458d45ff-g8hxr Binding Scheduled Successfully assigned kuttl-test-stirred-chipmunk/test-es-index-cleaner-with-prefix-query-55458d45ff-g8hxr to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:27:24 | es-index-cleaner-autoprov | 2024-09-12 07:25:53 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-55458d45ff-g8hxr AddedInterface Add eth0 [10.130.0.64/23] from ovn-kubernetes logger.go:42: 07:27:24 | es-index-cleaner-autoprov | 2024-09-12 07:25:53 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-55458d45ff-g8hxr.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a9186dcd910256c0f464b0a3928844a01de166a10c186c97ef4581bf288c23cb" already present on machine kubelet logger.go:42: 07:27:24 | es-index-cleaner-autoprov | 2024-09-12 07:25:53 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-55458d45ff-g8hxr.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:27:24 | es-index-cleaner-autoprov | 2024-09-12 07:25:53 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-55458d45ff-g8hxr.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:27:24 | es-index-cleaner-autoprov | 2024-09-12 07:25:53 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-55458d45ff-g8hxr.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:27:24 | es-index-cleaner-autoprov | 2024-09-12 07:25:53 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-55458d45ff-g8hxr.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:27:24 | es-index-cleaner-autoprov | 2024-09-12 07:25:53 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-55458d45ff-g8hxr.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:27:24 | es-index-cleaner-autoprov | 2024-09-12 07:25:53 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-55458d45ff-g8hxr.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" already present on machine kubelet logger.go:42: 07:27:24 | es-index-cleaner-autoprov | 2024-09-12 07:25:53 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-55458d45ff-g8hxr.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:27:24 | es-index-cleaner-autoprov | 2024-09-12 07:25:53 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-55458d45ff-g8hxr.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:27:24 | es-index-cleaner-autoprov | 2024-09-12 07:25:53 +0000 UTC Normal ReplicaSet.apps test-es-index-cleaner-with-prefix-query-55458d45ff SuccessfulCreate Created pod: test-es-index-cleaner-with-prefix-query-55458d45ff-g8hxr replicaset-controller logger.go:42: 07:27:24 | es-index-cleaner-autoprov | 2024-09-12 07:25:53 +0000 UTC Normal Deployment.apps test-es-index-cleaner-with-prefix-query ScalingReplicaSet Scaled up replica set test-es-index-cleaner-with-prefix-query-55458d45ff to 1 deployment-controller logger.go:42: 07:27:24 | es-index-cleaner-autoprov | 2024-09-12 07:25:58 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-55458d45ff-g8hxr.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:27:24 | es-index-cleaner-autoprov | 2024-09-12 07:25:58 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-55458d45ff-g8hxr.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:27:24 | es-index-cleaner-autoprov | 2024-09-12 07:25:58 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-55458d45ff-g8hxr.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:27:24 | es-index-cleaner-autoprov | 2024-09-12 07:25:58 +0000 UTC Normal ReplicaSet.apps test-es-index-cleaner-with-prefix-query-55458d45ff SuccessfulDelete Deleted pod: test-es-index-cleaner-with-prefix-query-55458d45ff-g8hxr replicaset-controller logger.go:42: 07:27:24 | es-index-cleaner-autoprov | 2024-09-12 07:25:58 +0000 UTC Normal Deployment.apps test-es-index-cleaner-with-prefix-query ScalingReplicaSet Scaled down replica set test-es-index-cleaner-with-prefix-query-55458d45ff to 0 from 1 deployment-controller logger.go:42: 07:27:24 | es-index-cleaner-autoprov | 2024-09-12 07:25:59 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-59c98fdfbd-vllbs Binding Scheduled Successfully assigned kuttl-test-stirred-chipmunk/test-es-index-cleaner-with-prefix-query-59c98fdfbd-vllbs to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:27:24 | es-index-cleaner-autoprov | 2024-09-12 07:25:59 +0000 UTC Normal ReplicaSet.apps test-es-index-cleaner-with-prefix-query-59c98fdfbd SuccessfulCreate Created pod: test-es-index-cleaner-with-prefix-query-59c98fdfbd-vllbs replicaset-controller logger.go:42: 07:27:24 | es-index-cleaner-autoprov | 2024-09-12 07:25:59 +0000 UTC Normal Deployment.apps test-es-index-cleaner-with-prefix-query ScalingReplicaSet Scaled up replica set test-es-index-cleaner-with-prefix-query-59c98fdfbd to 1 deployment-controller logger.go:42: 07:27:24 | es-index-cleaner-autoprov | 2024-09-12 07:26:00 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-59c98fdfbd-vllbs AddedInterface Add eth0 [10.130.0.65/23] from ovn-kubernetes logger.go:42: 07:27:24 | es-index-cleaner-autoprov | 2024-09-12 07:26:00 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-59c98fdfbd-vllbs.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a9186dcd910256c0f464b0a3928844a01de166a10c186c97ef4581bf288c23cb" already present on machine kubelet logger.go:42: 07:27:24 | es-index-cleaner-autoprov | 2024-09-12 07:26:00 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-59c98fdfbd-vllbs.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:27:24 | es-index-cleaner-autoprov | 2024-09-12 07:26:00 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-59c98fdfbd-vllbs.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:27:24 | es-index-cleaner-autoprov | 2024-09-12 07:26:00 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-59c98fdfbd-vllbs.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:27:24 | es-index-cleaner-autoprov | 2024-09-12 07:26:00 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-59c98fdfbd-vllbs.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:27:24 | es-index-cleaner-autoprov | 2024-09-12 07:26:00 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-59c98fdfbd-vllbs.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:27:24 | es-index-cleaner-autoprov | 2024-09-12 07:26:00 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-59c98fdfbd-vllbs.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" already present on machine kubelet logger.go:42: 07:27:24 | es-index-cleaner-autoprov | 2024-09-12 07:26:00 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-59c98fdfbd-vllbs.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:27:24 | es-index-cleaner-autoprov | 2024-09-12 07:26:00 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-59c98fdfbd-vllbs.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:27:24 | es-index-cleaner-autoprov | 2024-09-12 07:26:06 +0000 UTC Normal Pod 00-report-span-tjxhd Binding Scheduled Successfully assigned kuttl-test-stirred-chipmunk/00-report-span-tjxhd to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:27:24 | es-index-cleaner-autoprov | 2024-09-12 07:26:06 +0000 UTC Normal Job.batch 00-report-span SuccessfulCreate Created pod: 00-report-span-tjxhd job-controller logger.go:42: 07:27:24 | es-index-cleaner-autoprov | 2024-09-12 07:26:07 +0000 UTC Normal Pod 00-report-span-tjxhd AddedInterface Add eth0 [10.130.0.66/23] from ovn-kubernetes logger.go:42: 07:27:24 | es-index-cleaner-autoprov | 2024-09-12 07:26:07 +0000 UTC Normal Pod 00-report-span-tjxhd.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:27:24 | es-index-cleaner-autoprov | 2024-09-12 07:26:07 +0000 UTC Normal Pod 00-report-span-tjxhd.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 861ms (861ms including waiting) kubelet logger.go:42: 07:27:24 | es-index-cleaner-autoprov | 2024-09-12 07:26:08 +0000 UTC Normal Pod 00-report-span-tjxhd.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:27:24 | es-index-cleaner-autoprov | 2024-09-12 07:26:08 +0000 UTC Normal Pod 00-report-span-tjxhd.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:27:24 | es-index-cleaner-autoprov | 2024-09-12 07:26:08 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:27:24 | es-index-cleaner-autoprov | 2024-09-12 07:26:08 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:27:24 | es-index-cleaner-autoprov | 2024-09-12 07:26:08 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:27:24 | es-index-cleaner-autoprov | 2024-09-12 07:26:45 +0000 UTC Normal Job.batch 00-report-span Completed Job completed job-controller logger.go:42: 07:27:24 | es-index-cleaner-autoprov | 2024-09-12 07:26:53 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedGetResourceMetric failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 07:27:24 | es-index-cleaner-autoprov | 2024-09-12 07:26:53 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedGetResourceMetric failed to get memory utilization: missing request for memory in container jaeger-collector of Pod test-es-index-cleaner-with-prefix-collector-5484748d55-227tc horizontal-pod-autoscaler logger.go:42: 07:27:24 | es-index-cleaner-autoprov | 2024-09-12 07:26:53 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 07:27:24 | es-index-cleaner-autoprov | 2024-09-12 07:27:00 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-es-index-cleaner-287687644s6g Binding Scheduled Successfully assigned kuttl-test-stirred-chipmunk/test-es-index-cleaner-with-prefix-es-index-cleaner-287687644s6g to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:27:24 | es-index-cleaner-autoprov | 2024-09-12 07:27:00 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-es-index-cleaner-287687644s6g AddedInterface Add eth0 [10.130.0.67/23] from ovn-kubernetes logger.go:42: 07:27:24 | es-index-cleaner-autoprov | 2024-09-12 07:27:00 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-es-index-cleaner-287687644s6g.spec.containers{test-es-index-cleaner-with-prefix-es-index-cleaner} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-es-index-cleaner-rhel8@sha256:503ae309c9a5dfd756ec3dc336874d6b7a37d1c3c061d771b6e7ee1b5a1e7912" kubelet logger.go:42: 07:27:24 | es-index-cleaner-autoprov | 2024-09-12 07:27:00 +0000 UTC Normal Job.batch test-es-index-cleaner-with-prefix-es-index-cleaner-28768767 SuccessfulCreate Created pod: test-es-index-cleaner-with-prefix-es-index-cleaner-287687644s6g job-controller logger.go:42: 07:27:24 | es-index-cleaner-autoprov | 2024-09-12 07:27:00 +0000 UTC Normal CronJob.batch test-es-index-cleaner-with-prefix-es-index-cleaner SuccessfulCreate Created job test-es-index-cleaner-with-prefix-es-index-cleaner-28768767 cronjob-controller logger.go:42: 07:27:24 | es-index-cleaner-autoprov | 2024-09-12 07:27:05 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-es-index-cleaner-287687644s6g.spec.containers{test-es-index-cleaner-with-prefix-es-index-cleaner} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-es-index-cleaner-rhel8@sha256:503ae309c9a5dfd756ec3dc336874d6b7a37d1c3c061d771b6e7ee1b5a1e7912" in 4.516s (4.516s including waiting) kubelet logger.go:42: 07:27:24 | es-index-cleaner-autoprov | 2024-09-12 07:27:05 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-es-index-cleaner-287687644s6g.spec.containers{test-es-index-cleaner-with-prefix-es-index-cleaner} Created Created container test-es-index-cleaner-with-prefix-es-index-cleaner kubelet logger.go:42: 07:27:24 | es-index-cleaner-autoprov | 2024-09-12 07:27:05 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-es-index-cleaner-287687644s6g.spec.containers{test-es-index-cleaner-with-prefix-es-index-cleaner} Started Started container test-es-index-cleaner-with-prefix-es-index-cleaner kubelet logger.go:42: 07:27:24 | es-index-cleaner-autoprov | 2024-09-12 07:27:07 +0000 UTC Normal Job.batch test-es-index-cleaner-with-prefix-es-index-cleaner-28768767 Completed Job completed job-controller logger.go:42: 07:27:24 | es-index-cleaner-autoprov | 2024-09-12 07:27:07 +0000 UTC Normal CronJob.batch test-es-index-cleaner-with-prefix-es-index-cleaner SawCompletedJob Saw completed job: test-es-index-cleaner-with-prefix-es-index-cleaner-28768767, status: Complete cronjob-controller logger.go:42: 07:27:24 | es-index-cleaner-autoprov | 2024-09-12 07:27:18 +0000 UTC Normal Pod 00-check-indices-sv6gl Binding Scheduled Successfully assigned kuttl-test-stirred-chipmunk/00-check-indices-sv6gl to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:27:24 | es-index-cleaner-autoprov | 2024-09-12 07:27:18 +0000 UTC Normal Job.batch 00-check-indices SuccessfulCreate Created pod: 00-check-indices-sv6gl job-controller logger.go:42: 07:27:24 | es-index-cleaner-autoprov | 2024-09-12 07:27:19 +0000 UTC Normal Pod 00-check-indices-sv6gl AddedInterface Add eth0 [10.130.0.68/23] from ovn-kubernetes logger.go:42: 07:27:24 | es-index-cleaner-autoprov | 2024-09-12 07:27:19 +0000 UTC Normal Pod 00-check-indices-sv6gl.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:27:24 | es-index-cleaner-autoprov | 2024-09-12 07:27:20 +0000 UTC Normal Pod 00-check-indices-sv6gl.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 828ms (828ms including waiting) kubelet logger.go:42: 07:27:24 | es-index-cleaner-autoprov | 2024-09-12 07:27:20 +0000 UTC Normal Pod 00-check-indices-sv6gl.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:27:24 | es-index-cleaner-autoprov | 2024-09-12 07:27:20 +0000 UTC Normal Pod 00-check-indices-sv6gl.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:27:24 | es-index-cleaner-autoprov | 2024-09-12 07:27:22 +0000 UTC Normal Job.batch 00-check-indices Completed Job completed job-controller logger.go:42: 07:27:24 | es-index-cleaner-autoprov | Deleting namespace: kuttl-test-stirred-chipmunk === CONT kuttl/harness/es-from-aio-to-production logger.go:42: 07:27:30 | es-from-aio-to-production | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:27:30 | es-from-aio-to-production | Creating namespace: kuttl-test-climbing-lemming logger.go:42: 07:27:30 | es-from-aio-to-production/0-install | starting test step 0-install logger.go:42: 07:27:31 | es-from-aio-to-production/0-install | Jaeger:kuttl-test-climbing-lemming/my-jaeger created logger.go:42: 07:27:34 | es-from-aio-to-production/0-install | test step completed 0-install logger.go:42: 07:27:34 | es-from-aio-to-production/1-smoke-test | starting test step 1-smoke-test logger.go:42: 07:27:34 | es-from-aio-to-production/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:27:36 | es-from-aio-to-production/1-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:27:43 | es-from-aio-to-production/1-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:27:44 | es-from-aio-to-production/1-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:27:44 | es-from-aio-to-production/1-smoke-test | job.batch/report-span created logger.go:42: 07:27:44 | es-from-aio-to-production/1-smoke-test | job.batch/check-span created logger.go:42: 07:27:57 | es-from-aio-to-production/1-smoke-test | test step completed 1-smoke-test logger.go:42: 07:27:57 | es-from-aio-to-production/3-install | starting test step 3-install logger.go:42: 07:27:58 | es-from-aio-to-production/3-install | Jaeger:kuttl-test-climbing-lemming/my-jaeger updated logger.go:42: 07:28:33 | es-from-aio-to-production/3-install | test step completed 3-install logger.go:42: 07:28:33 | es-from-aio-to-production/4-smoke-test | starting test step 4-smoke-test logger.go:42: 07:28:33 | es-from-aio-to-production/4-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:28:42 | es-from-aio-to-production/4-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:28:42 | es-from-aio-to-production/4-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:28:43 | es-from-aio-to-production/4-smoke-test | job.batch/report-span unchanged logger.go:42: 07:28:43 | es-from-aio-to-production/4-smoke-test | job.batch/check-span unchanged logger.go:42: 07:28:43 | es-from-aio-to-production/4-smoke-test | test step completed 4-smoke-test logger.go:42: 07:28:43 | es-from-aio-to-production | es-from-aio-to-production events from ns kuttl-test-climbing-lemming: logger.go:42: 07:28:43 | es-from-aio-to-production | 2024-09-12 07:27:31 +0000 UTC Normal Pod my-jaeger-5fb5bf7756-fg5pq Binding Scheduled Successfully assigned kuttl-test-climbing-lemming/my-jaeger-5fb5bf7756-fg5pq to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:28:43 | es-from-aio-to-production | 2024-09-12 07:27:31 +0000 UTC Normal ReplicaSet.apps my-jaeger-5fb5bf7756 SuccessfulCreate Created pod: my-jaeger-5fb5bf7756-fg5pq replicaset-controller logger.go:42: 07:28:43 | es-from-aio-to-production | 2024-09-12 07:27:31 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-5fb5bf7756 to 1 deployment-controller logger.go:42: 07:28:43 | es-from-aio-to-production | 2024-09-12 07:27:32 +0000 UTC Normal Pod my-jaeger-5fb5bf7756-fg5pq AddedInterface Add eth0 [10.130.0.69/23] from ovn-kubernetes logger.go:42: 07:28:43 | es-from-aio-to-production | 2024-09-12 07:27:32 +0000 UTC Normal Pod my-jaeger-5fb5bf7756-fg5pq.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" already present on machine kubelet logger.go:42: 07:28:43 | es-from-aio-to-production | 2024-09-12 07:27:32 +0000 UTC Normal Pod my-jaeger-5fb5bf7756-fg5pq.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:28:43 | es-from-aio-to-production | 2024-09-12 07:27:32 +0000 UTC Normal Pod my-jaeger-5fb5bf7756-fg5pq.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:28:43 | es-from-aio-to-production | 2024-09-12 07:27:32 +0000 UTC Normal Pod my-jaeger-5fb5bf7756-fg5pq.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:28:43 | es-from-aio-to-production | 2024-09-12 07:27:32 +0000 UTC Normal Pod my-jaeger-5fb5bf7756-fg5pq.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:28:43 | es-from-aio-to-production | 2024-09-12 07:27:32 +0000 UTC Normal Pod my-jaeger-5fb5bf7756-fg5pq.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:28:43 | es-from-aio-to-production | 2024-09-12 07:27:36 +0000 UTC Normal Pod my-jaeger-5fb5bf7756-fg5pq.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:28:43 | es-from-aio-to-production | 2024-09-12 07:27:36 +0000 UTC Normal Pod my-jaeger-5fb5bf7756-fg5pq.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:28:43 | es-from-aio-to-production | 2024-09-12 07:27:36 +0000 UTC Normal ReplicaSet.apps my-jaeger-5fb5bf7756 SuccessfulDelete Deleted pod: my-jaeger-5fb5bf7756-fg5pq replicaset-controller logger.go:42: 07:28:43 | es-from-aio-to-production | 2024-09-12 07:27:36 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled down replica set my-jaeger-5fb5bf7756 to 0 from 1 deployment-controller logger.go:42: 07:28:43 | es-from-aio-to-production | 2024-09-12 07:27:37 +0000 UTC Normal Pod my-jaeger-7b977bc878-765fb Binding Scheduled Successfully assigned kuttl-test-climbing-lemming/my-jaeger-7b977bc878-765fb to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:28:43 | es-from-aio-to-production | 2024-09-12 07:27:37 +0000 UTC Normal ReplicaSet.apps my-jaeger-7b977bc878 SuccessfulCreate Created pod: my-jaeger-7b977bc878-765fb replicaset-controller logger.go:42: 07:28:43 | es-from-aio-to-production | 2024-09-12 07:27:37 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-7b977bc878 to 1 deployment-controller logger.go:42: 07:28:43 | es-from-aio-to-production | 2024-09-12 07:27:38 +0000 UTC Normal Pod my-jaeger-7b977bc878-765fb AddedInterface Add eth0 [10.130.0.70/23] from ovn-kubernetes logger.go:42: 07:28:43 | es-from-aio-to-production | 2024-09-12 07:27:38 +0000 UTC Normal Pod my-jaeger-7b977bc878-765fb.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" already present on machine kubelet logger.go:42: 07:28:43 | es-from-aio-to-production | 2024-09-12 07:27:38 +0000 UTC Normal Pod my-jaeger-7b977bc878-765fb.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:28:43 | es-from-aio-to-production | 2024-09-12 07:27:38 +0000 UTC Normal Pod my-jaeger-7b977bc878-765fb.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:28:43 | es-from-aio-to-production | 2024-09-12 07:27:38 +0000 UTC Normal Pod my-jaeger-7b977bc878-765fb.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:28:43 | es-from-aio-to-production | 2024-09-12 07:27:38 +0000 UTC Normal Pod my-jaeger-7b977bc878-765fb.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:28:43 | es-from-aio-to-production | 2024-09-12 07:27:38 +0000 UTC Normal Pod my-jaeger-7b977bc878-765fb.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:28:43 | es-from-aio-to-production | 2024-09-12 07:27:44 +0000 UTC Normal Pod check-span-v9sv4 Binding Scheduled Successfully assigned kuttl-test-climbing-lemming/check-span-v9sv4 to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:28:43 | es-from-aio-to-production | 2024-09-12 07:27:44 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-v9sv4 job-controller logger.go:42: 07:28:43 | es-from-aio-to-production | 2024-09-12 07:27:44 +0000 UTC Normal Pod report-span-t2tvp Binding Scheduled Successfully assigned kuttl-test-climbing-lemming/report-span-t2tvp to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:28:43 | es-from-aio-to-production | 2024-09-12 07:27:44 +0000 UTC Normal Pod report-span-t2tvp AddedInterface Add eth0 [10.130.0.71/23] from ovn-kubernetes logger.go:42: 07:28:43 | es-from-aio-to-production | 2024-09-12 07:27:44 +0000 UTC Normal Pod report-span-t2tvp.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:28:43 | es-from-aio-to-production | 2024-09-12 07:27:44 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-t2tvp job-controller logger.go:42: 07:28:43 | es-from-aio-to-production | 2024-09-12 07:27:45 +0000 UTC Normal Pod check-span-v9sv4 AddedInterface Add eth0 [10.130.0.72/23] from ovn-kubernetes logger.go:42: 07:28:43 | es-from-aio-to-production | 2024-09-12 07:27:45 +0000 UTC Normal Pod check-span-v9sv4.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:28:43 | es-from-aio-to-production | 2024-09-12 07:27:45 +0000 UTC Normal Pod report-span-t2tvp.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 842ms (842ms including waiting) kubelet logger.go:42: 07:28:43 | es-from-aio-to-production | 2024-09-12 07:27:45 +0000 UTC Normal Pod report-span-t2tvp.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:28:43 | es-from-aio-to-production | 2024-09-12 07:27:45 +0000 UTC Normal Pod report-span-t2tvp.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:28:43 | es-from-aio-to-production | 2024-09-12 07:27:46 +0000 UTC Normal Pod check-span-v9sv4.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 839ms (839ms including waiting) kubelet logger.go:42: 07:28:43 | es-from-aio-to-production | 2024-09-12 07:27:46 +0000 UTC Normal Pod check-span-v9sv4.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:28:43 | es-from-aio-to-production | 2024-09-12 07:27:46 +0000 UTC Normal Pod check-span-v9sv4.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:28:43 | es-from-aio-to-production | 2024-09-12 07:27:56 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:28:43 | es-from-aio-to-production | 2024-09-12 07:28:03 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestclimbinglemmingmyjaeger-1-86f5c58d57 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestclimbinglemmingmyjaeger-1-86f5cz2bmx replicaset-controller logger.go:42: 07:28:43 | es-from-aio-to-production | 2024-09-12 07:28:03 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestclimbinglemmingmyjaeger-1-86f5cz2bmx Binding Scheduled Successfully assigned kuttl-test-climbing-lemming/elasticsearch-cdm-kuttltestclimbinglemmingmyjaeger-1-86f5cz2bmx to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:28:43 | es-from-aio-to-production | 2024-09-12 07:28:03 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestclimbinglemmingmyjaeger-1-86f5cz2bmx AddedInterface Add eth0 [10.130.0.73/23] from ovn-kubernetes logger.go:42: 07:28:43 | es-from-aio-to-production | 2024-09-12 07:28:03 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestclimbinglemmingmyjaeger-1-86f5cz2bmx.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:bc97b8e13087050dfb1a75b02c5b780cbb8fb12a849a655169d072ba8bbf42b4" already present on machine kubelet logger.go:42: 07:28:43 | es-from-aio-to-production | 2024-09-12 07:28:03 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestclimbinglemmingmyjaeger-1-86f5cz2bmx.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:28:43 | es-from-aio-to-production | 2024-09-12 07:28:03 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestclimbinglemmingmyjaeger-1-86f5cz2bmx.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:28:43 | es-from-aio-to-production | 2024-09-12 07:28:03 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestclimbinglemmingmyjaeger-1-86f5cz2bmx.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:1d1d7d47b616995d18692f3a2d6232d7e1a0f41bc1503d87100f31939e080a78" already present on machine kubelet logger.go:42: 07:28:43 | es-from-aio-to-production | 2024-09-12 07:28:03 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestclimbinglemmingmyjaeger-1-86f5cz2bmx.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:28:43 | es-from-aio-to-production | 2024-09-12 07:28:03 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestclimbinglemmingmyjaeger-1-86f5cz2bmx.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:28:43 | es-from-aio-to-production | 2024-09-12 07:28:03 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestclimbinglemmingmyjaeger-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestclimbinglemmingmyjaeger-1-86f5c58d57 to 1 deployment-controller logger.go:42: 07:28:43 | es-from-aio-to-production | 2024-09-12 07:28:13 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestclimbinglemmingmyjaeger-1-86f5cz2bmx.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 07:28:43 | es-from-aio-to-production | 2024-09-12 07:28:18 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestclimbinglemmingmyjaeger-1-86f5cz2bmx.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:28:43 | es-from-aio-to-production | 2024-09-12 07:28:18 +0000 UTC Normal Job.batch report-span Completed Job completed job-controller logger.go:42: 07:28:43 | es-from-aio-to-production | 2024-09-12 07:28:29 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-cd5669645 to 1 deployment-controller logger.go:42: 07:28:43 | es-from-aio-to-production | 2024-09-12 07:28:30 +0000 UTC Normal Pod my-jaeger-7b977bc878-765fb.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:28:43 | es-from-aio-to-production | 2024-09-12 07:28:30 +0000 UTC Normal Pod my-jaeger-7b977bc878-765fb.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:28:43 | es-from-aio-to-production | 2024-09-12 07:28:30 +0000 UTC Normal Pod my-jaeger-collector-cd5669645-jrrdn Binding Scheduled Successfully assigned kuttl-test-climbing-lemming/my-jaeger-collector-cd5669645-jrrdn to ip-10-0-54-110.us-west-2.compute.internal default-scheduler logger.go:42: 07:28:43 | es-from-aio-to-production | 2024-09-12 07:28:30 +0000 UTC Normal Pod my-jaeger-collector-cd5669645-jrrdn AddedInterface Add eth0 [10.129.0.66/23] from ovn-kubernetes logger.go:42: 07:28:43 | es-from-aio-to-production | 2024-09-12 07:28:30 +0000 UTC Normal Pod my-jaeger-collector-cd5669645-jrrdn.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:4bab63290ff52e4f6328009f9c8b2c4496b230e9f8a98eac01736a66a291ff6c" already present on machine kubelet logger.go:42: 07:28:43 | es-from-aio-to-production | 2024-09-12 07:28:30 +0000 UTC Normal Pod my-jaeger-collector-cd5669645-jrrdn.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:28:43 | es-from-aio-to-production | 2024-09-12 07:28:30 +0000 UTC Normal Pod my-jaeger-collector-cd5669645-jrrdn.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:28:43 | es-from-aio-to-production | 2024-09-12 07:28:30 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-cd5669645 SuccessfulCreate Created pod: my-jaeger-collector-cd5669645-jrrdn replicaset-controller logger.go:42: 07:28:43 | es-from-aio-to-production | 2024-09-12 07:28:30 +0000 UTC Normal Pod my-jaeger-query-54776c97f7-jvsg9 Binding Scheduled Successfully assigned kuttl-test-climbing-lemming/my-jaeger-query-54776c97f7-jvsg9 to ip-10-0-52-42.us-west-2.compute.internal default-scheduler logger.go:42: 07:28:43 | es-from-aio-to-production | 2024-09-12 07:28:30 +0000 UTC Normal Pod my-jaeger-query-54776c97f7-jvsg9 AddedInterface Add eth0 [10.128.0.64/23] from ovn-kubernetes logger.go:42: 07:28:43 | es-from-aio-to-production | 2024-09-12 07:28:30 +0000 UTC Normal Pod my-jaeger-query-54776c97f7-jvsg9.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a9186dcd910256c0f464b0a3928844a01de166a10c186c97ef4581bf288c23cb" already present on machine kubelet logger.go:42: 07:28:43 | es-from-aio-to-production | 2024-09-12 07:28:30 +0000 UTC Normal Pod my-jaeger-query-54776c97f7-jvsg9.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:28:43 | es-from-aio-to-production | 2024-09-12 07:28:30 +0000 UTC Normal Pod my-jaeger-query-54776c97f7-jvsg9.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:28:43 | es-from-aio-to-production | 2024-09-12 07:28:30 +0000 UTC Normal Pod my-jaeger-query-54776c97f7-jvsg9.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:28:43 | es-from-aio-to-production | 2024-09-12 07:28:30 +0000 UTC Normal Pod my-jaeger-query-54776c97f7-jvsg9.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:28:43 | es-from-aio-to-production | 2024-09-12 07:28:30 +0000 UTC Normal Pod my-jaeger-query-54776c97f7-jvsg9.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:28:43 | es-from-aio-to-production | 2024-09-12 07:28:30 +0000 UTC Normal Pod my-jaeger-query-54776c97f7-jvsg9.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" already present on machine kubelet logger.go:42: 07:28:43 | es-from-aio-to-production | 2024-09-12 07:28:30 +0000 UTC Normal Pod my-jaeger-query-54776c97f7-jvsg9.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:28:43 | es-from-aio-to-production | 2024-09-12 07:28:30 +0000 UTC Normal Pod my-jaeger-query-54776c97f7-jvsg9.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:28:43 | es-from-aio-to-production | 2024-09-12 07:28:30 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-54776c97f7 SuccessfulCreate Created pod: my-jaeger-query-54776c97f7-jvsg9 replicaset-controller logger.go:42: 07:28:43 | es-from-aio-to-production | 2024-09-12 07:28:30 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-54776c97f7 to 1 deployment-controller logger.go:42: 07:28:43 | es-from-aio-to-production | Deleting namespace: kuttl-test-climbing-lemming === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (626.00s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (6.09s) --- PASS: kuttl/harness/es-multiinstance (76.21s) --- PASS: kuttl/harness/es-simple-prod (6.09s) --- PASS: kuttl/harness/es-rollover-autoprov (211.89s) --- PASS: kuttl/harness/es-increasing-replicas (115.85s) --- PASS: kuttl/harness/es-index-cleaner-autoprov (129.73s) --- PASS: kuttl/harness/es-from-aio-to-production (79.56s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name elasticsearch --report --output /logs/artifacts/elasticsearch.xml ./artifacts/kuttl-report.xml time="2024-09-12T07:28:50Z" level=debug msg="Setting a new name for the test suites" time="2024-09-12T07:28:50Z" level=debug msg="Removing 'artifacts' TestCase" time="2024-09-12T07:28:50Z" level=debug msg="normalizing test case names" time="2024-09-12T07:28:50Z" level=debug msg="elasticsearch/artifacts -> elasticsearch_artifacts" time="2024-09-12T07:28:50Z" level=debug msg="elasticsearch/es-multiinstance -> elasticsearch_es_multiinstance" time="2024-09-12T07:28:50Z" level=debug msg="elasticsearch/es-simple-prod -> elasticsearch_es_simple_prod" time="2024-09-12T07:28:50Z" level=debug msg="elasticsearch/es-rollover-autoprov -> elasticsearch_es_rollover_autoprov" time="2024-09-12T07:28:50Z" level=debug msg="elasticsearch/es-increasing-replicas -> elasticsearch_es_increasing_replicas" time="2024-09-12T07:28:50Z" level=debug msg="elasticsearch/es-index-cleaner-autoprov -> elasticsearch_es_index_cleaner_autoprov" time="2024-09-12T07:28:50Z" level=debug msg="elasticsearch/es-from-aio-to-production -> elasticsearch_es_from_aio_to_production" +-----------------------------------------+--------+ | NAME | RESULT | +-----------------------------------------+--------+ | elasticsearch_artifacts | passed | | elasticsearch_es_multiinstance | passed | | elasticsearch_es_simple_prod | passed | | elasticsearch_es_rollover_autoprov | passed | | elasticsearch_es_increasing_replicas | passed | | elasticsearch_es_index_cleaner_autoprov | passed | | elasticsearch_es_from_aio_to_production | passed | +-----------------------------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh examples false true + '[' 3 -ne 3 ']' + test_suite_name=examples + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/examples.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-examples make[2]: Entering directory '/tmp/jaeger-tests' >>>> Elasticsearch image not loaded because SKIP_ES_EXTERNAL is true KAFKA_VERSION=3.6.0 \ SKIP_KAFKA=false \ VERTX_IMG=jaegertracing/vertx-create-span:operator-e2e-tests \ ./tests/e2e/examples/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.29 True False 40m Cluster version is 4.15.29' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.29 True False 40m Cluster version is 4.15.29' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/examples/render.sh ++ export SUITE_DIR=./tests/e2e/examples ++ SUITE_DIR=./tests/e2e/examples ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/examples ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + start_test examples-agent-with-priority-class + '[' 1 -ne 1 ']' + test_name=examples-agent-with-priority-class + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-agent-with-priority-class' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-agent-with-priority-class\e[0m' Rendering files for test examples-agent-with-priority-class + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build + '[' _build '!=' _build ']' + mkdir -p examples-agent-with-priority-class + cd examples-agent-with-priority-class + example_name=agent-with-priority-class + prepare_daemonset 00 + '[' 1 -ne 1 ']' + test_step=00 + '[' true = true ']' + cat /tmp/jaeger-tests/examples/openshift/hostport-scc-daemonset.yaml + echo --- + cat /tmp/jaeger-tests/examples/openshift/service_account_jaeger-agent-daemonset.yaml + '[' true '!=' true ']' + render_install_example agent-with-priority-class 02 + '[' 2 -ne 2 ']' + example_name=agent-with-priority-class + test_step=02 + install_file=./02-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/agent-with-priority-class.yaml -o ./02-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./02-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./02-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./02-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./02-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./02-install.yaml ++ jaeger_name=agent-as-daemonset ++ '[' -z agent-as-daemonset ']' ++ echo agent-as-daemonset ++ return 0 + JAEGER_NAME=agent-as-daemonset + local jaeger_strategy ++ get_jaeger_strategy ./02-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./02-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./02-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./02-install.yaml ++ strategy=DaemonSet ++ '[' DaemonSet = null ']' ++ echo DaemonSet ++ return 0 + jaeger_strategy=DaemonSet + '[' DaemonSet = DaemonSet ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./02-assert.yaml + render_smoke_test_example agent-with-priority-class 02 + '[' 2 -ne 2 ']' + example_name=agent-with-priority-class + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/agent-with-priority-class.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/agent-with-priority-class.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/agent-with-priority-class.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/agent-with-priority-class.yaml ++ jaeger_name=agent-as-daemonset ++ '[' -z agent-as-daemonset ']' ++ echo agent-as-daemonset ++ return 0 + jaeger_name=agent-as-daemonset + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test agent-as-daemonset true 02 + '[' 3 -ne 3 ']' + jaeger=agent-as-daemonset + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://agent-as-daemonset-query:443 + JAEGER_QUERY_ENDPOINT=https://agent-as-daemonset-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://agent-as-daemonset-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://agent-as-daemonset-collector-headless:14268 + export JAEGER_NAME=agent-as-daemonset + JAEGER_NAME=agent-as-daemonset + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-all-in-one-with-options + '[' 1 -ne 1 ']' + test_name=examples-all-in-one-with-options + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-all-in-one-with-options' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-all-in-one-with-options\e[0m' Rendering files for test examples-all-in-one-with-options + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-agent-with-priority-class + '[' examples-agent-with-priority-class '!=' _build ']' + cd .. + mkdir -p examples-all-in-one-with-options + cd examples-all-in-one-with-options + example_name=all-in-one-with-options + render_install_example all-in-one-with-options 00 + '[' 2 -ne 2 ']' + example_name=all-in-one-with-options + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/all-in-one-with-options.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=my-jaeger ++ '[' -z my-jaeger ']' ++ echo my-jaeger ++ return 0 + JAEGER_NAME=my-jaeger + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=allInOne ++ '[' allInOne = production ']' ++ '[' allInOne = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + /tmp/jaeger-tests/bin/yq e -i '.metadata.name="my-jaeger"' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i 'del(.spec.allInOne.image)' ./00-install.yaml + render_smoke_test_example all-in-one-with-options 01 + '[' 2 -ne 2 ']' + example_name=all-in-one-with-options + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/all-in-one-with-options.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/all-in-one-with-options.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/all-in-one-with-options.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/all-in-one-with-options.yaml ++ jaeger_name=my-jaeger ++ '[' -z my-jaeger ']' ++ echo my-jaeger ++ return 0 + jaeger_name=my-jaeger + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test my-jaeger true 01 + '[' 3 -ne 3 ']' + jaeger=my-jaeger + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + '[' true = true ']' + sed -i s~my-jaeger-query:443~my-jaeger-query:443/jaeger~gi ./01-smoke-test.yaml + start_test examples-business-application-injected-sidecar + '[' 1 -ne 1 ']' + test_name=examples-business-application-injected-sidecar + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-business-application-injected-sidecar' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-business-application-injected-sidecar\e[0m' Rendering files for test examples-business-application-injected-sidecar + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-all-in-one-with-options + '[' examples-all-in-one-with-options '!=' _build ']' + cd .. + mkdir -p examples-business-application-injected-sidecar + cd examples-business-application-injected-sidecar + example_name=simplest + cp /tmp/jaeger-tests/examples/business-application-injected-sidecar.yaml ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].image=strenv(VERTX_IMG)' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.httpGet.path="/"' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.httpGet.port=8080' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.initialDelaySeconds=1' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.failureThreshold=3' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.periodSeconds=10' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.successThreshold=1' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.timeoutSeconds=1' ./00-install.yaml + render_install_example simplest 01 + '[' 2 -ne 2 ']' + example_name=simplest + test_step=01 + install_file=./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/simplest.yaml -o ./01-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./01-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./01-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./01-install.yaml ++ jaeger_name=simplest ++ '[' -z simplest ']' ++ echo simplest ++ return 0 + JAEGER_NAME=simplest + local jaeger_strategy ++ get_jaeger_strategy ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./01-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./01-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./01-assert.yaml + render_smoke_test_example simplest 02 + '[' 2 -ne 2 ']' + example_name=simplest + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/simplest.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/simplest.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/simplest.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/simplest.yaml ++ jaeger_name=simplest ++ '[' -z simplest ']' ++ echo simplest ++ return 0 + jaeger_name=simplest + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test simplest true 02 + '[' 3 -ne 3 ']' + jaeger=simplest + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simplest-query:443 + JAEGER_QUERY_ENDPOINT=https://simplest-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 + export JAEGER_NAME=simplest + JAEGER_NAME=simplest + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-collector-with-priority-class + '[' 1 -ne 1 ']' + test_name=examples-collector-with-priority-class + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-collector-with-priority-class' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-collector-with-priority-class\e[0m' Rendering files for test examples-collector-with-priority-class + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-business-application-injected-sidecar + '[' examples-business-application-injected-sidecar '!=' _build ']' + cd .. + mkdir -p examples-collector-with-priority-class + cd examples-collector-with-priority-class + example_name=collector-with-priority-class + render_install_example collector-with-priority-class 00 + '[' 2 -ne 2 ']' + example_name=collector-with-priority-class + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/collector-with-priority-class.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=collector-with-high-priority ++ '[' -z collector-with-high-priority ']' ++ echo collector-with-high-priority ++ return 0 + JAEGER_NAME=collector-with-high-priority + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test_example collector-with-priority-class 01 + '[' 2 -ne 2 ']' + example_name=collector-with-priority-class + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/collector-with-priority-class.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/collector-with-priority-class.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/collector-with-priority-class.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/collector-with-priority-class.yaml ++ jaeger_name=collector-with-high-priority ++ '[' -z collector-with-high-priority ']' ++ echo collector-with-high-priority ++ return 0 + jaeger_name=collector-with-high-priority + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test collector-with-high-priority true 01 + '[' 3 -ne 3 ']' + jaeger=collector-with-high-priority + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://collector-with-high-priority-query:443 + JAEGER_QUERY_ENDPOINT=https://collector-with-high-priority-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://collector-with-high-priority-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://collector-with-high-priority-collector-headless:14268 + export JAEGER_NAME=collector-with-high-priority + JAEGER_NAME=collector-with-high-priority + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-service-types + '[' 1 -ne 1 ']' + test_name=examples-service-types + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-service-types' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-service-types\e[0m' Rendering files for test examples-service-types + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-collector-with-priority-class + '[' examples-collector-with-priority-class '!=' _build ']' + cd .. + mkdir -p examples-service-types + cd examples-service-types + example_name=service-types + render_install_example service-types 00 + '[' 2 -ne 2 ']' + example_name=service-types + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/service-types.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=service-types ++ '[' -z service-types ']' ++ echo service-types ++ return 0 + JAEGER_NAME=service-types + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test_example service-types 01 + '[' 2 -ne 2 ']' + example_name=service-types + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/service-types.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/service-types.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/service-types.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/service-types.yaml ++ jaeger_name=service-types ++ '[' -z service-types ']' ++ echo service-types ++ return 0 + jaeger_name=service-types + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test service-types true 01 + '[' 3 -ne 3 ']' + jaeger=service-types + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://service-types-query:443 + JAEGER_QUERY_ENDPOINT=https://service-types-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://service-types-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://service-types-collector-headless:14268 + export JAEGER_NAME=service-types + JAEGER_NAME=service-types + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-simple-prod + '[' 1 -ne 1 ']' + test_name=examples-simple-prod + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-simple-prod' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-simple-prod\e[0m' Rendering files for test examples-simple-prod + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-service-types + '[' examples-service-types '!=' _build ']' + cd .. + mkdir -p examples-simple-prod + cd examples-simple-prod + example_name=simple-prod + render_install_example simple-prod 01 + '[' 2 -ne 2 ']' + example_name=simple-prod + test_step=01 + install_file=./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/simple-prod.yaml -o ./01-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./01-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./01-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./01-install.yaml ++ jaeger_name=simple-prod ++ '[' -z simple-prod ']' ++ echo simple-prod ++ return 0 + JAEGER_NAME=simple-prod + local jaeger_strategy ++ get_jaeger_strategy ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./01-install.yaml ++ strategy=production ++ '[' production = production ']' ++ echo production ++ return 0 + jaeger_strategy=production + '[' production = DaemonSet ']' + '[' production = allInOne ']' + '[' production = production ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + [[ true = true ]] + [[ true = true ]] + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.options={}' ./01-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.elasticsearch={"nodeCount":1,"resources":{"limits":{"memory":"2Gi"}}}' ./01-install.yaml + render_smoke_test_example simple-prod 02 + '[' 2 -ne 2 ']' + example_name=simple-prod + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/simple-prod.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/simple-prod.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/simple-prod.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/simple-prod.yaml ++ jaeger_name=simple-prod ++ '[' -z simple-prod ']' ++ echo simple-prod ++ return 0 + jaeger_name=simple-prod + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test simple-prod true 02 + '[' 3 -ne 3 ']' + jaeger=simple-prod + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + export JAEGER_NAME=simple-prod + JAEGER_NAME=simple-prod + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-simple-prod-with-volumes + '[' 1 -ne 1 ']' + test_name=examples-simple-prod-with-volumes + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-simple-prod-with-volumes' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-simple-prod-with-volumes\e[0m' Rendering files for test examples-simple-prod-with-volumes + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-simple-prod + '[' examples-simple-prod '!=' _build ']' + cd .. + mkdir -p examples-simple-prod-with-volumes + cd examples-simple-prod-with-volumes + example_name=simple-prod-with-volumes + render_install_example simple-prod-with-volumes 01 + '[' 2 -ne 2 ']' + example_name=simple-prod-with-volumes + test_step=01 + install_file=./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/simple-prod-with-volumes.yaml -o ./01-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./01-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./01-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./01-install.yaml ++ jaeger_name=simple-prod ++ '[' -z simple-prod ']' ++ echo simple-prod ++ return 0 + JAEGER_NAME=simple-prod + local jaeger_strategy ++ get_jaeger_strategy ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./01-install.yaml ++ strategy=production ++ '[' production = production ']' ++ echo production ++ return 0 + jaeger_strategy=production + '[' production = DaemonSet ']' + '[' production = allInOne ']' + '[' production = production ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + [[ true = true ]] + [[ true = true ]] + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.options={}' ./01-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.elasticsearch={"nodeCount":1,"resources":{"limits":{"memory":"2Gi"}}}' ./01-install.yaml + render_smoke_test_example simple-prod-with-volumes 02 + '[' 2 -ne 2 ']' + example_name=simple-prod-with-volumes + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/simple-prod-with-volumes.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/simple-prod-with-volumes.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/simple-prod-with-volumes.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/simple-prod-with-volumes.yaml ++ jaeger_name=simple-prod ++ '[' -z simple-prod ']' ++ echo simple-prod ++ return 0 + jaeger_name=simple-prod + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test simple-prod true 02 + '[' 3 -ne 3 ']' + jaeger=simple-prod + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + export JAEGER_NAME=simple-prod + JAEGER_NAME=simple-prod + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + /tmp/jaeger-tests/bin/gomplate -f ./03-check-volume.yaml.template -o 03-check-volume.yaml + start_test examples-simplest + '[' 1 -ne 1 ']' + test_name=examples-simplest + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-simplest' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-simplest\e[0m' Rendering files for test examples-simplest + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-simple-prod-with-volumes + '[' examples-simple-prod-with-volumes '!=' _build ']' + cd .. + mkdir -p examples-simplest + cd examples-simplest + example_name=simplest + render_install_example simplest 00 + '[' 2 -ne 2 ']' + example_name=simplest + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/simplest.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=simplest ++ '[' -z simplest ']' ++ echo simplest ++ return 0 + JAEGER_NAME=simplest + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test_example simplest 01 + '[' 2 -ne 2 ']' + example_name=simplest + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/simplest.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/simplest.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/simplest.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/simplest.yaml ++ jaeger_name=simplest ++ '[' -z simplest ']' ++ echo simplest ++ return 0 + jaeger_name=simplest + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test simplest true 01 + '[' 3 -ne 3 ']' + jaeger=simplest + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simplest-query:443 + JAEGER_QUERY_ENDPOINT=https://simplest-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 + export JAEGER_NAME=simplest + JAEGER_NAME=simplest + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-with-badger + '[' 1 -ne 1 ']' + test_name=examples-with-badger + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-with-badger' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-with-badger\e[0m' Rendering files for test examples-with-badger + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-simplest + '[' examples-simplest '!=' _build ']' + cd .. + mkdir -p examples-with-badger + cd examples-with-badger + example_name=with-badger + render_install_example with-badger 00 + '[' 2 -ne 2 ']' + example_name=with-badger + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/with-badger.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=with-badger ++ '[' -z with-badger ']' ++ echo with-badger ++ return 0 + JAEGER_NAME=with-badger + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test_example with-badger 01 + '[' 2 -ne 2 ']' + example_name=with-badger + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/with-badger.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/with-badger.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/with-badger.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/with-badger.yaml ++ jaeger_name=with-badger ++ '[' -z with-badger ']' ++ echo with-badger ++ return 0 + jaeger_name=with-badger + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test with-badger true 01 + '[' 3 -ne 3 ']' + jaeger=with-badger + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://with-badger-query:443 + JAEGER_QUERY_ENDPOINT=https://with-badger-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://with-badger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://with-badger-collector-headless:14268 + export JAEGER_NAME=with-badger + JAEGER_NAME=with-badger + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-with-badger-and-volume + '[' 1 -ne 1 ']' + test_name=examples-with-badger-and-volume + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-with-badger-and-volume' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-with-badger-and-volume\e[0m' Rendering files for test examples-with-badger-and-volume + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-with-badger + '[' examples-with-badger '!=' _build ']' + cd .. + mkdir -p examples-with-badger-and-volume + cd examples-with-badger-and-volume + example_name=with-badger-and-volume + render_install_example with-badger-and-volume 00 + '[' 2 -ne 2 ']' + example_name=with-badger-and-volume + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/with-badger-and-volume.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=with-badger-and-volume ++ '[' -z with-badger-and-volume ']' ++ echo with-badger-and-volume ++ return 0 + JAEGER_NAME=with-badger-and-volume + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test_example with-badger-and-volume 01 + '[' 2 -ne 2 ']' + example_name=with-badger-and-volume + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/with-badger-and-volume.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/with-badger-and-volume.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/with-badger-and-volume.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/with-badger-and-volume.yaml ++ jaeger_name=with-badger-and-volume ++ '[' -z with-badger-and-volume ']' ++ echo with-badger-and-volume ++ return 0 + jaeger_name=with-badger-and-volume + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test with-badger-and-volume true 01 + '[' 3 -ne 3 ']' + jaeger=with-badger-and-volume + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://with-badger-and-volume-query:443 + JAEGER_QUERY_ENDPOINT=https://with-badger-and-volume-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://with-badger-and-volume-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://with-badger-and-volume-collector-headless:14268 + export JAEGER_NAME=with-badger-and-volume + JAEGER_NAME=with-badger-and-volume + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-with-cassandra + '[' 1 -ne 1 ']' + test_name=examples-with-cassandra + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-with-cassandra' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-with-cassandra\e[0m' Rendering files for test examples-with-cassandra + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-with-badger-and-volume + '[' examples-with-badger-and-volume '!=' _build ']' + cd .. + mkdir -p examples-with-cassandra + cd examples-with-cassandra + example_name=with-cassandra + render_install_cassandra 00 + '[' 1 -ne 1 ']' + test_step=00 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/cassandra-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/cassandra-assert.yaml.template -o ./00-assert.yaml + render_install_example with-cassandra 01 + '[' 2 -ne 2 ']' + example_name=with-cassandra + test_step=01 + install_file=./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/with-cassandra.yaml -o ./01-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./01-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./01-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./01-install.yaml ++ jaeger_name=with-cassandra ++ '[' -z with-cassandra ']' ++ echo with-cassandra ++ return 0 + JAEGER_NAME=with-cassandra + local jaeger_strategy ++ get_jaeger_strategy ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./01-install.yaml ++ strategy=allInOne ++ '[' allInOne = production ']' ++ '[' allInOne = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./01-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./01-assert.yaml + render_smoke_test_example with-cassandra 02 + '[' 2 -ne 2 ']' + example_name=with-cassandra + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/with-cassandra.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/with-cassandra.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/with-cassandra.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/with-cassandra.yaml ++ jaeger_name=with-cassandra ++ '[' -z with-cassandra ']' ++ echo with-cassandra ++ return 0 + jaeger_name=with-cassandra + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test with-cassandra true 02 + '[' 3 -ne 3 ']' + jaeger=with-cassandra + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://with-cassandra-query:443 + JAEGER_QUERY_ENDPOINT=https://with-cassandra-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://with-cassandra-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://with-cassandra-collector-headless:14268 + export JAEGER_NAME=with-cassandra + JAEGER_NAME=with-cassandra + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-with-sampling + '[' 1 -ne 1 ']' + test_name=examples-with-sampling + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-with-sampling' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-with-sampling\e[0m' Rendering files for test examples-with-sampling + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-with-cassandra + '[' examples-with-cassandra '!=' _build ']' + cd .. + mkdir -p examples-with-sampling + cd examples-with-sampling + export example_name=with-sampling + example_name=with-sampling + render_install_cassandra 00 + '[' 1 -ne 1 ']' + test_step=00 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/cassandra-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/cassandra-assert.yaml.template -o ./00-assert.yaml + render_install_example with-sampling 01 + '[' 2 -ne 2 ']' + example_name=with-sampling + test_step=01 + install_file=./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/with-sampling.yaml -o ./01-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./01-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./01-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./01-install.yaml ++ jaeger_name=with-sampling ++ '[' -z with-sampling ']' ++ echo with-sampling ++ return 0 + JAEGER_NAME=with-sampling + local jaeger_strategy ++ get_jaeger_strategy ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./01-install.yaml ++ strategy=allInOne ++ '[' allInOne = production ']' ++ '[' allInOne = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./01-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./01-assert.yaml + render_smoke_test_example with-sampling 02 + '[' 2 -ne 2 ']' + example_name=with-sampling + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/with-sampling.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/with-sampling.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/with-sampling.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/with-sampling.yaml ++ jaeger_name=with-sampling ++ '[' -z with-sampling ']' ++ echo with-sampling ++ return 0 + jaeger_name=with-sampling + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test with-sampling true 02 + '[' 3 -ne 3 ']' + jaeger=with-sampling + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://with-sampling-query:443 + JAEGER_QUERY_ENDPOINT=https://with-sampling-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://with-sampling-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://with-sampling-collector-headless:14268 + export JAEGER_NAME=with-sampling + JAEGER_NAME=with-sampling + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-agent-as-daemonset + '[' 1 -ne 1 ']' + test_name=examples-agent-as-daemonset + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-agent-as-daemonset' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-agent-as-daemonset\e[0m' Rendering files for test examples-agent-as-daemonset + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-with-sampling + '[' examples-with-sampling '!=' _build ']' + cd .. + mkdir -p examples-agent-as-daemonset + cd examples-agent-as-daemonset + '[' true = true ']' + prepare_daemonset 00 + '[' 1 -ne 1 ']' + test_step=00 + '[' true = true ']' + cat /tmp/jaeger-tests/examples/openshift/hostport-scc-daemonset.yaml + echo --- + cat /tmp/jaeger-tests/examples/openshift/service_account_jaeger-agent-daemonset.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/openshift/agent-as-daemonset.yaml -o 02-install.yaml + '[' true = true ']' + start_test examples-openshift-with-htpasswd + '[' 1 -ne 1 ']' + test_name=examples-openshift-with-htpasswd + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-openshift-with-htpasswd' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-openshift-with-htpasswd\e[0m' Rendering files for test examples-openshift-with-htpasswd + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-agent-as-daemonset + '[' examples-agent-as-daemonset '!=' _build ']' + cd .. + mkdir -p examples-openshift-with-htpasswd + cd examples-openshift-with-htpasswd + export JAEGER_NAME=with-htpasswd + JAEGER_NAME=with-htpasswd + export JAEGER_USERNAME=awesomeuser + JAEGER_USERNAME=awesomeuser + export JAEGER_PASSWORD=awesomepassword + JAEGER_PASSWORD=awesomepassword + export 'JAEGER_USER_PASSWORD_HASH=awesomeuser:{SHA}uUdqPVUyqNBmERU0Qxj3KFaZnjw=' + JAEGER_USER_PASSWORD_HASH='awesomeuser:{SHA}uUdqPVUyqNBmERU0Qxj3KFaZnjw=' ++ echo 'awesomeuser:{SHA}uUdqPVUyqNBmERU0Qxj3KFaZnjw=' ++ base64 + SECRET=YXdlc29tZXVzZXI6e1NIQX11VWRxUFZVeXFOQm1FUlUwUXhqM0tGYVpuanc9Cg== + /tmp/jaeger-tests/bin/gomplate -f ./00-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/openshift/with-htpasswd.yaml -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./01-assert.yaml + export 'GET_URL_COMMAND=kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE' + GET_URL_COMMAND='kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE' + export 'URL=https://$(kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE)/search' + URL='https://$(kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE)/search' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/ensure-ingress-host.sh.template -o ./ensure-ingress-host.sh + chmod +x ./ensure-ingress-host.sh + INSECURE=true + JAEGER_USERNAME= + JAEGER_PASSWORD= + EXPECTED_CODE=403 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./02-check-unsecured.yaml + JAEGER_USERNAME=wronguser + JAEGER_PASSWORD=wrongpassword + EXPECTED_CODE=403 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./03-check-unauthorized.yaml + EXPECTED_CODE=200 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./04-check-authorized.yaml + skip_test examples-agent-as-daemonset 'This test is flaky in Prow CI' + '[' 2 -ne 2 ']' + test_name=examples-agent-as-daemonset + message='This test is flaky in Prow CI' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-openshift-with-htpasswd + '[' examples-openshift-with-htpasswd '!=' _build ']' + cd .. + rm -rf examples-agent-as-daemonset + warning 'examples-agent-as-daemonset: This test is flaky in Prow CI' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: examples-agent-as-daemonset: This test is flaky in Prow CI\e[0m' WAR: examples-agent-as-daemonset: This test is flaky in Prow CI + skip_test examples-with-badger-and-volume 'This test is flaky in Prow CI' + '[' 2 -ne 2 ']' + test_name=examples-with-badger-and-volume + message='This test is flaky in Prow CI' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build + '[' _build '!=' _build ']' + rm -rf examples-with-badger-and-volume + warning 'examples-with-badger-and-volume: This test is flaky in Prow CI' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: examples-with-badger-and-volume: This test is flaky in Prow CI\e[0m' WAR: examples-with-badger-and-volume: This test is flaky in Prow CI + skip_test examples-collector-with-priority-class 'This test is flaky in Prow CI' + '[' 2 -ne 2 ']' + test_name=examples-collector-with-priority-class + message='This test is flaky in Prow CI' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build + '[' _build '!=' _build ']' + rm -rf examples-collector-with-priority-class + warning 'examples-collector-with-priority-class: This test is flaky in Prow CI' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: examples-collector-with-priority-class: This test is flaky in Prow CI\e[0m' WAR: examples-collector-with-priority-class: This test is flaky in Prow CI make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running examples E2E tests' Running examples E2E tests + cd tests/e2e/examples/_build + set +e + KUBECONFIG=/tmp/kubeconfig-2486720052 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-rosa-h-u0tm.f61o.s3.devshift.org:443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 12 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/examples-agent-with-priority-class === PAUSE kuttl/harness/examples-agent-with-priority-class === RUN kuttl/harness/examples-all-in-one-with-options === PAUSE kuttl/harness/examples-all-in-one-with-options === RUN kuttl/harness/examples-business-application-injected-sidecar === PAUSE kuttl/harness/examples-business-application-injected-sidecar === RUN kuttl/harness/examples-openshift-with-htpasswd === PAUSE kuttl/harness/examples-openshift-with-htpasswd === RUN kuttl/harness/examples-service-types === PAUSE kuttl/harness/examples-service-types === RUN kuttl/harness/examples-simple-prod === PAUSE kuttl/harness/examples-simple-prod === RUN kuttl/harness/examples-simple-prod-with-volumes === PAUSE kuttl/harness/examples-simple-prod-with-volumes === RUN kuttl/harness/examples-simplest === PAUSE kuttl/harness/examples-simplest === RUN kuttl/harness/examples-with-badger === PAUSE kuttl/harness/examples-with-badger === RUN kuttl/harness/examples-with-cassandra === PAUSE kuttl/harness/examples-with-cassandra === RUN kuttl/harness/examples-with-sampling === PAUSE kuttl/harness/examples-with-sampling === CONT kuttl/harness/artifacts logger.go:42: 07:29:21 | artifacts | Creating namespace: kuttl-test-assuring-lizard logger.go:42: 07:29:22 | artifacts | artifacts events from ns kuttl-test-assuring-lizard: logger.go:42: 07:29:22 | artifacts | Deleting namespace: kuttl-test-assuring-lizard === CONT kuttl/harness/examples-simple-prod logger.go:42: 07:29:28 | examples-simple-prod | Creating namespace: kuttl-test-inspired-monarch logger.go:42: 07:29:28 | examples-simple-prod/1-install | starting test step 1-install logger.go:42: 07:29:28 | examples-simple-prod/1-install | Jaeger:kuttl-test-inspired-monarch/simple-prod created logger.go:42: 07:30:04 | examples-simple-prod/1-install | test step completed 1-install logger.go:42: 07:30:04 | examples-simple-prod/2-smoke-test | starting test step 2-smoke-test logger.go:42: 07:30:04 | examples-simple-prod/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE simple-prod /dev/null] logger.go:42: 07:30:05 | examples-simple-prod/2-smoke-test | Warning: resource jaegers/simple-prod is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:30:13 | examples-simple-prod/2-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:30:13 | examples-simple-prod/2-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:30:13 | examples-simple-prod/2-smoke-test | job.batch/report-span created logger.go:42: 07:30:14 | examples-simple-prod/2-smoke-test | job.batch/check-span created logger.go:42: 07:30:27 | examples-simple-prod/2-smoke-test | test step completed 2-smoke-test logger.go:42: 07:30:27 | examples-simple-prod | examples-simple-prod events from ns kuttl-test-inspired-monarch: logger.go:42: 07:30:27 | examples-simple-prod | 2024-09-12 07:29:34 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestinspiredmonarchsimpleprod-1-58b79dcbbc SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestinspiredmonarchsimpleprod-1-58bwslgn replicaset-controller logger.go:42: 07:30:27 | examples-simple-prod | 2024-09-12 07:29:34 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestinspiredmonarchsimpleprod-1-58bwslgn Binding Scheduled Successfully assigned kuttl-test-inspired-monarch/elasticsearch-cdm-kuttltestinspiredmonarchsimpleprod-1-58bwslgn to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:30:27 | examples-simple-prod | 2024-09-12 07:29:34 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestinspiredmonarchsimpleprod-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestinspiredmonarchsimpleprod-1-58b79dcbbc to 1 deployment-controller logger.go:42: 07:30:27 | examples-simple-prod | 2024-09-12 07:29:35 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestinspiredmonarchsimpleprod-1-58bwslgn AddedInterface Add eth0 [10.130.0.74/23] from ovn-kubernetes logger.go:42: 07:30:27 | examples-simple-prod | 2024-09-12 07:29:35 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestinspiredmonarchsimpleprod-1-58bwslgn.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:bc97b8e13087050dfb1a75b02c5b780cbb8fb12a849a655169d072ba8bbf42b4" already present on machine kubelet logger.go:42: 07:30:27 | examples-simple-prod | 2024-09-12 07:29:35 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestinspiredmonarchsimpleprod-1-58bwslgn.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:30:27 | examples-simple-prod | 2024-09-12 07:29:35 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestinspiredmonarchsimpleprod-1-58bwslgn.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:30:27 | examples-simple-prod | 2024-09-12 07:29:35 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestinspiredmonarchsimpleprod-1-58bwslgn.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:1d1d7d47b616995d18692f3a2d6232d7e1a0f41bc1503d87100f31939e080a78" already present on machine kubelet logger.go:42: 07:30:27 | examples-simple-prod | 2024-09-12 07:29:35 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestinspiredmonarchsimpleprod-1-58bwslgn.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:30:27 | examples-simple-prod | 2024-09-12 07:29:35 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestinspiredmonarchsimpleprod-1-58bwslgn.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:30:27 | examples-simple-prod | 2024-09-12 07:29:45 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestinspiredmonarchsimpleprod-1-58bwslgn.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 07:30:27 | examples-simple-prod | 2024-09-12 07:29:50 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestinspiredmonarchsimpleprod-1-58bwslgn.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:30:27 | examples-simple-prod | 2024-09-12 07:30:01 +0000 UTC Normal Pod simple-prod-collector-64885c9dc9-nhvsf Binding Scheduled Successfully assigned kuttl-test-inspired-monarch/simple-prod-collector-64885c9dc9-nhvsf to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:30:27 | examples-simple-prod | 2024-09-12 07:30:01 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-64885c9dc9 SuccessfulCreate Created pod: simple-prod-collector-64885c9dc9-nhvsf replicaset-controller logger.go:42: 07:30:27 | examples-simple-prod | 2024-09-12 07:30:01 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-64885c9dc9 to 1 deployment-controller logger.go:42: 07:30:27 | examples-simple-prod | 2024-09-12 07:30:01 +0000 UTC Normal Pod simple-prod-query-86c9f84b8-t44lb Binding Scheduled Successfully assigned kuttl-test-inspired-monarch/simple-prod-query-86c9f84b8-t44lb to ip-10-0-52-42.us-west-2.compute.internal default-scheduler logger.go:42: 07:30:27 | examples-simple-prod | 2024-09-12 07:30:01 +0000 UTC Normal ReplicaSet.apps simple-prod-query-86c9f84b8 SuccessfulCreate Created pod: simple-prod-query-86c9f84b8-t44lb replicaset-controller logger.go:42: 07:30:27 | examples-simple-prod | 2024-09-12 07:30:01 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-86c9f84b8 to 1 deployment-controller logger.go:42: 07:30:27 | examples-simple-prod | 2024-09-12 07:30:02 +0000 UTC Normal Pod simple-prod-collector-64885c9dc9-nhvsf AddedInterface Add eth0 [10.130.0.75/23] from ovn-kubernetes logger.go:42: 07:30:27 | examples-simple-prod | 2024-09-12 07:30:02 +0000 UTC Normal Pod simple-prod-collector-64885c9dc9-nhvsf.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:4bab63290ff52e4f6328009f9c8b2c4496b230e9f8a98eac01736a66a291ff6c" already present on machine kubelet logger.go:42: 07:30:27 | examples-simple-prod | 2024-09-12 07:30:02 +0000 UTC Normal Pod simple-prod-collector-64885c9dc9-nhvsf.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:30:27 | examples-simple-prod | 2024-09-12 07:30:02 +0000 UTC Normal Pod simple-prod-collector-64885c9dc9-nhvsf.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:30:27 | examples-simple-prod | 2024-09-12 07:30:02 +0000 UTC Normal Pod simple-prod-query-86c9f84b8-t44lb AddedInterface Add eth0 [10.128.0.65/23] from ovn-kubernetes logger.go:42: 07:30:27 | examples-simple-prod | 2024-09-12 07:30:02 +0000 UTC Normal Pod simple-prod-query-86c9f84b8-t44lb.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a9186dcd910256c0f464b0a3928844a01de166a10c186c97ef4581bf288c23cb" already present on machine kubelet logger.go:42: 07:30:27 | examples-simple-prod | 2024-09-12 07:30:02 +0000 UTC Normal Pod simple-prod-query-86c9f84b8-t44lb.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:30:27 | examples-simple-prod | 2024-09-12 07:30:02 +0000 UTC Normal Pod simple-prod-query-86c9f84b8-t44lb.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:30:27 | examples-simple-prod | 2024-09-12 07:30:02 +0000 UTC Normal Pod simple-prod-query-86c9f84b8-t44lb.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:30:27 | examples-simple-prod | 2024-09-12 07:30:02 +0000 UTC Normal Pod simple-prod-query-86c9f84b8-t44lb.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:30:27 | examples-simple-prod | 2024-09-12 07:30:02 +0000 UTC Normal Pod simple-prod-query-86c9f84b8-t44lb.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:30:27 | examples-simple-prod | 2024-09-12 07:30:02 +0000 UTC Normal Pod simple-prod-query-86c9f84b8-t44lb.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" already present on machine kubelet logger.go:42: 07:30:27 | examples-simple-prod | 2024-09-12 07:30:02 +0000 UTC Normal Pod simple-prod-query-86c9f84b8-t44lb.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:30:27 | examples-simple-prod | 2024-09-12 07:30:02 +0000 UTC Normal Pod simple-prod-query-86c9f84b8-t44lb.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:30:27 | examples-simple-prod | 2024-09-12 07:30:06 +0000 UTC Normal Pod simple-prod-query-86c9f84b8-t44lb.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:30:27 | examples-simple-prod | 2024-09-12 07:30:06 +0000 UTC Normal Pod simple-prod-query-86c9f84b8-t44lb.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:30:27 | examples-simple-prod | 2024-09-12 07:30:06 +0000 UTC Normal ReplicaSet.apps simple-prod-query-86c9f84b8 SuccessfulDelete Deleted pod: simple-prod-query-86c9f84b8-t44lb replicaset-controller logger.go:42: 07:30:27 | examples-simple-prod | 2024-09-12 07:30:06 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled down replica set simple-prod-query-86c9f84b8 to 0 from 1 deployment-controller logger.go:42: 07:30:27 | examples-simple-prod | 2024-09-12 07:30:07 +0000 UTC Normal Pod simple-prod-query-86c9f84b8-t44lb.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:30:27 | examples-simple-prod | 2024-09-12 07:30:08 +0000 UTC Normal Pod simple-prod-query-cd47cf78c-97mnq Binding Scheduled Successfully assigned kuttl-test-inspired-monarch/simple-prod-query-cd47cf78c-97mnq to ip-10-0-52-42.us-west-2.compute.internal default-scheduler logger.go:42: 07:30:27 | examples-simple-prod | 2024-09-12 07:30:08 +0000 UTC Normal Pod simple-prod-query-cd47cf78c-97mnq AddedInterface Add eth0 [10.128.0.66/23] from ovn-kubernetes logger.go:42: 07:30:27 | examples-simple-prod | 2024-09-12 07:30:08 +0000 UTC Normal Pod simple-prod-query-cd47cf78c-97mnq.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a9186dcd910256c0f464b0a3928844a01de166a10c186c97ef4581bf288c23cb" already present on machine kubelet logger.go:42: 07:30:27 | examples-simple-prod | 2024-09-12 07:30:08 +0000 UTC Normal Pod simple-prod-query-cd47cf78c-97mnq.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:30:27 | examples-simple-prod | 2024-09-12 07:30:08 +0000 UTC Normal Pod simple-prod-query-cd47cf78c-97mnq.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:30:27 | examples-simple-prod | 2024-09-12 07:30:08 +0000 UTC Normal Pod simple-prod-query-cd47cf78c-97mnq.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:30:27 | examples-simple-prod | 2024-09-12 07:30:08 +0000 UTC Normal Pod simple-prod-query-cd47cf78c-97mnq.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:30:27 | examples-simple-prod | 2024-09-12 07:30:08 +0000 UTC Normal Pod simple-prod-query-cd47cf78c-97mnq.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:30:27 | examples-simple-prod | 2024-09-12 07:30:08 +0000 UTC Normal Pod simple-prod-query-cd47cf78c-97mnq.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" already present on machine kubelet logger.go:42: 07:30:27 | examples-simple-prod | 2024-09-12 07:30:08 +0000 UTC Normal Pod simple-prod-query-cd47cf78c-97mnq.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:30:27 | examples-simple-prod | 2024-09-12 07:30:08 +0000 UTC Normal Pod simple-prod-query-cd47cf78c-97mnq.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:30:27 | examples-simple-prod | 2024-09-12 07:30:08 +0000 UTC Normal ReplicaSet.apps simple-prod-query-cd47cf78c SuccessfulCreate Created pod: simple-prod-query-cd47cf78c-97mnq replicaset-controller logger.go:42: 07:30:27 | examples-simple-prod | 2024-09-12 07:30:08 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-cd47cf78c to 1 deployment-controller logger.go:42: 07:30:27 | examples-simple-prod | 2024-09-12 07:30:13 +0000 UTC Normal Pod report-span-tbc85 Binding Scheduled Successfully assigned kuttl-test-inspired-monarch/report-span-tbc85 to ip-10-0-54-110.us-west-2.compute.internal default-scheduler logger.go:42: 07:30:27 | examples-simple-prod | 2024-09-12 07:30:13 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-tbc85 job-controller logger.go:42: 07:30:27 | examples-simple-prod | 2024-09-12 07:30:14 +0000 UTC Normal Pod check-span-9lrrn Binding Scheduled Successfully assigned kuttl-test-inspired-monarch/check-span-9lrrn to ip-10-0-54-110.us-west-2.compute.internal default-scheduler logger.go:42: 07:30:27 | examples-simple-prod | 2024-09-12 07:30:14 +0000 UTC Normal Pod check-span-9lrrn AddedInterface Add eth0 [10.129.0.69/23] from ovn-kubernetes logger.go:42: 07:30:27 | examples-simple-prod | 2024-09-12 07:30:14 +0000 UTC Normal Pod check-span-9lrrn.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:30:27 | examples-simple-prod | 2024-09-12 07:30:14 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-9lrrn job-controller logger.go:42: 07:30:27 | examples-simple-prod | 2024-09-12 07:30:14 +0000 UTC Normal Pod report-span-tbc85 AddedInterface Add eth0 [10.129.0.68/23] from ovn-kubernetes logger.go:42: 07:30:27 | examples-simple-prod | 2024-09-12 07:30:14 +0000 UTC Normal Pod report-span-tbc85.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:30:27 | examples-simple-prod | 2024-09-12 07:30:15 +0000 UTC Normal Pod check-span-9lrrn.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 815ms (815ms including waiting) kubelet logger.go:42: 07:30:27 | examples-simple-prod | 2024-09-12 07:30:15 +0000 UTC Normal Pod check-span-9lrrn.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:30:27 | examples-simple-prod | 2024-09-12 07:30:15 +0000 UTC Normal Pod check-span-9lrrn.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:30:27 | examples-simple-prod | 2024-09-12 07:30:15 +0000 UTC Normal Pod report-span-tbc85.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 848ms (848ms including waiting) kubelet logger.go:42: 07:30:27 | examples-simple-prod | 2024-09-12 07:30:15 +0000 UTC Normal Pod report-span-tbc85.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:30:27 | examples-simple-prod | 2024-09-12 07:30:15 +0000 UTC Normal Pod report-span-tbc85.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:30:27 | examples-simple-prod | 2024-09-12 07:30:16 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:30:27 | examples-simple-prod | 2024-09-12 07:30:16 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:30:27 | examples-simple-prod | 2024-09-12 07:30:16 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:30:27 | examples-simple-prod | 2024-09-12 07:30:26 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:30:27 | examples-simple-prod | Deleting namespace: kuttl-test-inspired-monarch === CONT kuttl/harness/examples-with-sampling logger.go:42: 07:30:34 | examples-with-sampling | Creating namespace: kuttl-test-gorgeous-dingo logger.go:42: 07:30:34 | examples-with-sampling/0-install | starting test step 0-install logger.go:42: 07:30:34 | examples-with-sampling/0-install | running command: [sh -c cd /tmp/jaeger-tests && make cassandra STORAGE_NAMESPACE=$NAMESPACE] logger.go:42: 07:30:34 | examples-with-sampling/0-install | make[2]: Entering directory '/tmp/jaeger-tests' logger.go:42: 07:30:34 | examples-with-sampling/0-install | >>>> Creating namespace kuttl-test-gorgeous-dingo logger.go:42: 07:30:34 | examples-with-sampling/0-install | kubectl create namespace kuttl-test-gorgeous-dingo 2>&1 | grep -v "already exists" || true logger.go:42: 07:30:35 | examples-with-sampling/0-install | kubectl create -f ./tests/cassandra.yml --namespace kuttl-test-gorgeous-dingo 2>&1 | grep -v "already exists" || true logger.go:42: 07:30:35 | examples-with-sampling/0-install | service/cassandra created logger.go:42: 07:30:35 | examples-with-sampling/0-install | statefulset.apps/cassandra created logger.go:42: 07:30:35 | examples-with-sampling/0-install | make[2]: Leaving directory '/tmp/jaeger-tests' logger.go:42: 07:30:49 | examples-with-sampling/0-install | test step completed 0-install logger.go:42: 07:30:49 | examples-with-sampling/1-install | starting test step 1-install logger.go:42: 07:30:50 | examples-with-sampling/1-install | Jaeger:kuttl-test-gorgeous-dingo/with-sampling created logger.go:42: 07:30:53 | examples-with-sampling/1-install | test step completed 1-install logger.go:42: 07:30:53 | examples-with-sampling/2-smoke-test | starting test step 2-smoke-test logger.go:42: 07:30:53 | examples-with-sampling/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE with-sampling /dev/null] logger.go:42: 07:30:55 | examples-with-sampling/2-smoke-test | Warning: resource jaegers/with-sampling is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:31:02 | examples-with-sampling/2-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://with-sampling-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://with-sampling-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:31:03 | examples-with-sampling/2-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:31:03 | examples-with-sampling/2-smoke-test | job.batch/report-span created logger.go:42: 07:31:03 | examples-with-sampling/2-smoke-test | job.batch/check-span created logger.go:42: 07:31:15 | examples-with-sampling/2-smoke-test | test step completed 2-smoke-test logger.go:42: 07:31:15 | examples-with-sampling/3- | starting test step 3- logger.go:42: 07:31:16 | examples-with-sampling/3- | test step completed 3- logger.go:42: 07:31:16 | examples-with-sampling | examples-with-sampling events from ns kuttl-test-gorgeous-dingo: logger.go:42: 07:31:16 | examples-with-sampling | 2024-09-12 07:30:35 +0000 UTC Normal Pod cassandra-0 Binding Scheduled Successfully assigned kuttl-test-gorgeous-dingo/cassandra-0 to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:31:16 | examples-with-sampling | 2024-09-12 07:30:35 +0000 UTC Normal StatefulSet.apps cassandra SuccessfulCreate create Pod cassandra-0 in StatefulSet cassandra successful statefulset-controller logger.go:42: 07:31:16 | examples-with-sampling | 2024-09-12 07:30:36 +0000 UTC Normal Pod cassandra-0 AddedInterface Add eth0 [10.130.0.76/23] from ovn-kubernetes logger.go:42: 07:31:16 | examples-with-sampling | 2024-09-12 07:30:36 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Pulling Pulling image "cassandra:3.11" kubelet logger.go:42: 07:31:16 | examples-with-sampling | 2024-09-12 07:30:41 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Pulled Successfully pulled image "cassandra:3.11" in 5.558s (5.558s including waiting) kubelet logger.go:42: 07:31:16 | examples-with-sampling | 2024-09-12 07:30:41 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Created Created container cassandra kubelet logger.go:42: 07:31:16 | examples-with-sampling | 2024-09-12 07:30:41 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Started Started container cassandra kubelet logger.go:42: 07:31:16 | examples-with-sampling | 2024-09-12 07:30:42 +0000 UTC Normal Pod cassandra-1 Binding Scheduled Successfully assigned kuttl-test-gorgeous-dingo/cassandra-1 to ip-10-0-54-110.us-west-2.compute.internal default-scheduler logger.go:42: 07:31:16 | examples-with-sampling | 2024-09-12 07:30:42 +0000 UTC Normal Pod cassandra-1 AddedInterface Add eth0 [10.129.0.70/23] from ovn-kubernetes logger.go:42: 07:31:16 | examples-with-sampling | 2024-09-12 07:30:42 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Pulling Pulling image "cassandra:3.11" kubelet logger.go:42: 07:31:16 | examples-with-sampling | 2024-09-12 07:30:42 +0000 UTC Normal StatefulSet.apps cassandra SuccessfulCreate create Pod cassandra-1 in StatefulSet cassandra successful statefulset-controller logger.go:42: 07:31:16 | examples-with-sampling | 2024-09-12 07:30:48 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Pulled Successfully pulled image "cassandra:3.11" in 5.905s (5.905s including waiting) kubelet logger.go:42: 07:31:16 | examples-with-sampling | 2024-09-12 07:30:48 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Created Created container cassandra kubelet logger.go:42: 07:31:16 | examples-with-sampling | 2024-09-12 07:30:48 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Started Started container cassandra kubelet logger.go:42: 07:31:16 | examples-with-sampling | 2024-09-12 07:30:50 +0000 UTC Normal Pod with-sampling-6d468c4558-nq978 Binding Scheduled Successfully assigned kuttl-test-gorgeous-dingo/with-sampling-6d468c4558-nq978 to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:31:16 | examples-with-sampling | 2024-09-12 07:30:50 +0000 UTC Normal Pod with-sampling-6d468c4558-nq978 AddedInterface Add eth0 [10.130.0.77/23] from ovn-kubernetes logger.go:42: 07:31:16 | examples-with-sampling | 2024-09-12 07:30:50 +0000 UTC Normal Pod with-sampling-6d468c4558-nq978.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" already present on machine kubelet logger.go:42: 07:31:16 | examples-with-sampling | 2024-09-12 07:30:50 +0000 UTC Normal ReplicaSet.apps with-sampling-6d468c4558 SuccessfulCreate Created pod: with-sampling-6d468c4558-nq978 replicaset-controller logger.go:42: 07:31:16 | examples-with-sampling | 2024-09-12 07:30:50 +0000 UTC Normal Deployment.apps with-sampling ScalingReplicaSet Scaled up replica set with-sampling-6d468c4558 to 1 deployment-controller logger.go:42: 07:31:16 | examples-with-sampling | 2024-09-12 07:30:51 +0000 UTC Normal Pod with-sampling-6d468c4558-nq978.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:31:16 | examples-with-sampling | 2024-09-12 07:30:51 +0000 UTC Normal Pod with-sampling-6d468c4558-nq978.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:31:16 | examples-with-sampling | 2024-09-12 07:30:51 +0000 UTC Normal Pod with-sampling-6d468c4558-nq978.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:31:16 | examples-with-sampling | 2024-09-12 07:30:51 +0000 UTC Normal Pod with-sampling-6d468c4558-nq978.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:31:16 | examples-with-sampling | 2024-09-12 07:30:51 +0000 UTC Normal Pod with-sampling-6d468c4558-nq978.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:31:16 | examples-with-sampling | 2024-09-12 07:30:55 +0000 UTC Normal Pod with-sampling-6d468c4558-nq978.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:31:16 | examples-with-sampling | 2024-09-12 07:30:55 +0000 UTC Normal Pod with-sampling-6d468c4558-nq978.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:31:16 | examples-with-sampling | 2024-09-12 07:30:55 +0000 UTC Normal ReplicaSet.apps with-sampling-6d468c4558 SuccessfulDelete Deleted pod: with-sampling-6d468c4558-nq978 replicaset-controller logger.go:42: 07:31:16 | examples-with-sampling | 2024-09-12 07:30:55 +0000 UTC Normal Deployment.apps with-sampling ScalingReplicaSet Scaled down replica set with-sampling-6d468c4558 to 0 from 1 deployment-controller logger.go:42: 07:31:16 | examples-with-sampling | 2024-09-12 07:30:57 +0000 UTC Normal Pod with-sampling-ff89dfcdd-zchd2 Binding Scheduled Successfully assigned kuttl-test-gorgeous-dingo/with-sampling-ff89dfcdd-zchd2 to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:31:16 | examples-with-sampling | 2024-09-12 07:30:57 +0000 UTC Normal ReplicaSet.apps with-sampling-ff89dfcdd SuccessfulCreate Created pod: with-sampling-ff89dfcdd-zchd2 replicaset-controller logger.go:42: 07:31:16 | examples-with-sampling | 2024-09-12 07:30:57 +0000 UTC Normal Deployment.apps with-sampling ScalingReplicaSet Scaled up replica set with-sampling-ff89dfcdd to 1 deployment-controller logger.go:42: 07:31:16 | examples-with-sampling | 2024-09-12 07:30:58 +0000 UTC Normal Pod with-sampling-ff89dfcdd-zchd2 AddedInterface Add eth0 [10.130.0.78/23] from ovn-kubernetes logger.go:42: 07:31:16 | examples-with-sampling | 2024-09-12 07:30:58 +0000 UTC Normal Pod with-sampling-ff89dfcdd-zchd2.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" already present on machine kubelet logger.go:42: 07:31:16 | examples-with-sampling | 2024-09-12 07:30:58 +0000 UTC Normal Pod with-sampling-ff89dfcdd-zchd2.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:31:16 | examples-with-sampling | 2024-09-12 07:30:58 +0000 UTC Normal Pod with-sampling-ff89dfcdd-zchd2.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:31:16 | examples-with-sampling | 2024-09-12 07:30:58 +0000 UTC Normal Pod with-sampling-ff89dfcdd-zchd2.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:31:16 | examples-with-sampling | 2024-09-12 07:30:58 +0000 UTC Normal Pod with-sampling-ff89dfcdd-zchd2.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:31:16 | examples-with-sampling | 2024-09-12 07:30:58 +0000 UTC Normal Pod with-sampling-ff89dfcdd-zchd2.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:31:16 | examples-with-sampling | 2024-09-12 07:31:03 +0000 UTC Normal Pod check-span-l7d8l Binding Scheduled Successfully assigned kuttl-test-gorgeous-dingo/check-span-l7d8l to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:31:16 | examples-with-sampling | 2024-09-12 07:31:03 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-l7d8l job-controller logger.go:42: 07:31:16 | examples-with-sampling | 2024-09-12 07:31:03 +0000 UTC Normal Pod report-span-7vs9d Binding Scheduled Successfully assigned kuttl-test-gorgeous-dingo/report-span-7vs9d to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:31:16 | examples-with-sampling | 2024-09-12 07:31:03 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-7vs9d job-controller logger.go:42: 07:31:16 | examples-with-sampling | 2024-09-12 07:31:04 +0000 UTC Normal Pod check-span-l7d8l AddedInterface Add eth0 [10.130.0.80/23] from ovn-kubernetes logger.go:42: 07:31:16 | examples-with-sampling | 2024-09-12 07:31:04 +0000 UTC Normal Pod check-span-l7d8l.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:31:16 | examples-with-sampling | 2024-09-12 07:31:04 +0000 UTC Normal Pod report-span-7vs9d AddedInterface Add eth0 [10.130.0.79/23] from ovn-kubernetes logger.go:42: 07:31:16 | examples-with-sampling | 2024-09-12 07:31:04 +0000 UTC Normal Pod report-span-7vs9d.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:31:16 | examples-with-sampling | 2024-09-12 07:31:05 +0000 UTC Normal Pod check-span-l7d8l.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 819ms (819ms including waiting) kubelet logger.go:42: 07:31:16 | examples-with-sampling | 2024-09-12 07:31:05 +0000 UTC Normal Pod check-span-l7d8l.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:31:16 | examples-with-sampling | 2024-09-12 07:31:05 +0000 UTC Normal Pod check-span-l7d8l.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:31:16 | examples-with-sampling | 2024-09-12 07:31:05 +0000 UTC Normal Pod report-span-7vs9d.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 921ms (921ms including waiting) kubelet logger.go:42: 07:31:16 | examples-with-sampling | 2024-09-12 07:31:05 +0000 UTC Normal Pod report-span-7vs9d.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:31:16 | examples-with-sampling | 2024-09-12 07:31:05 +0000 UTC Normal Pod report-span-7vs9d.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:31:16 | examples-with-sampling | 2024-09-12 07:31:15 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:31:16 | examples-with-sampling | Deleting namespace: kuttl-test-gorgeous-dingo === CONT kuttl/harness/examples-with-cassandra logger.go:42: 07:33:15 | examples-with-cassandra | Creating namespace: kuttl-test-valued-kite logger.go:42: 07:33:16 | examples-with-cassandra/0-install | starting test step 0-install logger.go:42: 07:33:16 | examples-with-cassandra/0-install | running command: [sh -c cd /tmp/jaeger-tests && make cassandra STORAGE_NAMESPACE=$NAMESPACE] logger.go:42: 07:33:16 | examples-with-cassandra/0-install | make[2]: Entering directory '/tmp/jaeger-tests' logger.go:42: 07:33:16 | examples-with-cassandra/0-install | >>>> Creating namespace kuttl-test-valued-kite logger.go:42: 07:33:16 | examples-with-cassandra/0-install | kubectl create namespace kuttl-test-valued-kite 2>&1 | grep -v "already exists" || true logger.go:42: 07:33:16 | examples-with-cassandra/0-install | kubectl create -f ./tests/cassandra.yml --namespace kuttl-test-valued-kite 2>&1 | grep -v "already exists" || true logger.go:42: 07:33:16 | examples-with-cassandra/0-install | service/cassandra created logger.go:42: 07:33:16 | examples-with-cassandra/0-install | statefulset.apps/cassandra created logger.go:42: 07:33:16 | examples-with-cassandra/0-install | make[2]: Leaving directory '/tmp/jaeger-tests' logger.go:42: 07:33:19 | examples-with-cassandra/0-install | test step completed 0-install logger.go:42: 07:33:19 | examples-with-cassandra/1-install | starting test step 1-install logger.go:42: 07:33:19 | examples-with-cassandra/1-install | Jaeger:kuttl-test-valued-kite/with-cassandra created logger.go:42: 07:33:38 | examples-with-cassandra/1-install | test step completed 1-install logger.go:42: 07:33:38 | examples-with-cassandra/2-smoke-test | starting test step 2-smoke-test logger.go:42: 07:33:38 | examples-with-cassandra/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE with-cassandra /dev/null] logger.go:42: 07:33:40 | examples-with-cassandra/2-smoke-test | Warning: resource jaegers/with-cassandra is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:33:47 | examples-with-cassandra/2-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://with-cassandra-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://with-cassandra-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:33:48 | examples-with-cassandra/2-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:33:48 | examples-with-cassandra/2-smoke-test | job.batch/report-span created logger.go:42: 07:33:48 | examples-with-cassandra/2-smoke-test | job.batch/check-span created logger.go:42: 07:34:00 | examples-with-cassandra/2-smoke-test | test step completed 2-smoke-test logger.go:42: 07:34:00 | examples-with-cassandra | examples-with-cassandra events from ns kuttl-test-valued-kite: logger.go:42: 07:34:00 | examples-with-cassandra | 2024-09-12 07:33:16 +0000 UTC Normal Pod cassandra-0 Binding Scheduled Successfully assigned kuttl-test-valued-kite/cassandra-0 to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:34:00 | examples-with-cassandra | 2024-09-12 07:33:16 +0000 UTC Normal StatefulSet.apps cassandra SuccessfulCreate create Pod cassandra-0 in StatefulSet cassandra successful statefulset-controller logger.go:42: 07:34:00 | examples-with-cassandra | 2024-09-12 07:33:17 +0000 UTC Normal Pod cassandra-0 AddedInterface Add eth0 [10.130.0.81/23] from ovn-kubernetes logger.go:42: 07:34:00 | examples-with-cassandra | 2024-09-12 07:33:17 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Pulled Container image "cassandra:3.11" already present on machine kubelet logger.go:42: 07:34:00 | examples-with-cassandra | 2024-09-12 07:33:17 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Created Created container cassandra kubelet logger.go:42: 07:34:00 | examples-with-cassandra | 2024-09-12 07:33:17 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Started Started container cassandra kubelet logger.go:42: 07:34:00 | examples-with-cassandra | 2024-09-12 07:33:17 +0000 UTC Normal Pod cassandra-1 Binding Scheduled Successfully assigned kuttl-test-valued-kite/cassandra-1 to ip-10-0-54-110.us-west-2.compute.internal default-scheduler logger.go:42: 07:34:00 | examples-with-cassandra | 2024-09-12 07:33:17 +0000 UTC Normal StatefulSet.apps cassandra SuccessfulCreate create Pod cassandra-1 in StatefulSet cassandra successful statefulset-controller logger.go:42: 07:34:00 | examples-with-cassandra | 2024-09-12 07:33:18 +0000 UTC Normal Pod cassandra-1 AddedInterface Add eth0 [10.129.0.71/23] from ovn-kubernetes logger.go:42: 07:34:00 | examples-with-cassandra | 2024-09-12 07:33:18 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Pulled Container image "cassandra:3.11" already present on machine kubelet logger.go:42: 07:34:00 | examples-with-cassandra | 2024-09-12 07:33:18 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Created Created container cassandra kubelet logger.go:42: 07:34:00 | examples-with-cassandra | 2024-09-12 07:33:18 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Started Started container cassandra kubelet logger.go:42: 07:34:00 | examples-with-cassandra | 2024-09-12 07:33:20 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-zvglk Binding Scheduled Successfully assigned kuttl-test-valued-kite/with-cassandra-cassandra-schema-job-zvglk to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:34:00 | examples-with-cassandra | 2024-09-12 07:33:20 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-zvglk AddedInterface Add eth0 [10.130.0.82/23] from ovn-kubernetes logger.go:42: 07:34:00 | examples-with-cassandra | 2024-09-12 07:33:20 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-zvglk.spec.containers{with-cassandra-cassandra-schema-job} Pulling Pulling image "jaegertracing/jaeger-cassandra-schema:1.57.0" kubelet logger.go:42: 07:34:00 | examples-with-cassandra | 2024-09-12 07:33:20 +0000 UTC Normal Job.batch with-cassandra-cassandra-schema-job SuccessfulCreate Created pod: with-cassandra-cassandra-schema-job-zvglk job-controller logger.go:42: 07:34:00 | examples-with-cassandra | 2024-09-12 07:33:26 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-zvglk.spec.containers{with-cassandra-cassandra-schema-job} Pulled Successfully pulled image "jaegertracing/jaeger-cassandra-schema:1.57.0" in 6.441s (6.441s including waiting) kubelet logger.go:42: 07:34:00 | examples-with-cassandra | 2024-09-12 07:33:27 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-zvglk.spec.containers{with-cassandra-cassandra-schema-job} Created Created container with-cassandra-cassandra-schema-job kubelet logger.go:42: 07:34:00 | examples-with-cassandra | 2024-09-12 07:33:27 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-zvglk.spec.containers{with-cassandra-cassandra-schema-job} Started Started container with-cassandra-cassandra-schema-job kubelet logger.go:42: 07:34:00 | examples-with-cassandra | 2024-09-12 07:33:35 +0000 UTC Normal Job.batch with-cassandra-cassandra-schema-job Completed Job completed job-controller logger.go:42: 07:34:00 | examples-with-cassandra | 2024-09-12 07:33:36 +0000 UTC Normal Pod with-cassandra-75866d478c-pqfqt Binding Scheduled Successfully assigned kuttl-test-valued-kite/with-cassandra-75866d478c-pqfqt to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:34:00 | examples-with-cassandra | 2024-09-12 07:33:36 +0000 UTC Warning Pod with-cassandra-75866d478c-pqfqt FailedMount MountVolume.SetUp failed for volume "with-cassandra-collector-tls-config-volume" : secret "with-cassandra-collector-headless-tls" not found kubelet logger.go:42: 07:34:00 | examples-with-cassandra | 2024-09-12 07:33:36 +0000 UTC Normal ReplicaSet.apps with-cassandra-75866d478c SuccessfulCreate Created pod: with-cassandra-75866d478c-pqfqt replicaset-controller logger.go:42: 07:34:00 | examples-with-cassandra | 2024-09-12 07:33:36 +0000 UTC Normal Deployment.apps with-cassandra ScalingReplicaSet Scaled up replica set with-cassandra-75866d478c to 1 deployment-controller logger.go:42: 07:34:00 | examples-with-cassandra | 2024-09-12 07:33:37 +0000 UTC Normal Pod with-cassandra-75866d478c-pqfqt AddedInterface Add eth0 [10.130.0.83/23] from ovn-kubernetes logger.go:42: 07:34:00 | examples-with-cassandra | 2024-09-12 07:33:37 +0000 UTC Normal Pod with-cassandra-75866d478c-pqfqt.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" already present on machine kubelet logger.go:42: 07:34:00 | examples-with-cassandra | 2024-09-12 07:33:37 +0000 UTC Normal Pod with-cassandra-75866d478c-pqfqt.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:34:00 | examples-with-cassandra | 2024-09-12 07:33:37 +0000 UTC Normal Pod with-cassandra-75866d478c-pqfqt.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:34:00 | examples-with-cassandra | 2024-09-12 07:33:37 +0000 UTC Normal Pod with-cassandra-75866d478c-pqfqt.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:34:00 | examples-with-cassandra | 2024-09-12 07:33:37 +0000 UTC Normal Pod with-cassandra-75866d478c-pqfqt.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:34:00 | examples-with-cassandra | 2024-09-12 07:33:37 +0000 UTC Normal Pod with-cassandra-75866d478c-pqfqt.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:34:00 | examples-with-cassandra | 2024-09-12 07:33:41 +0000 UTC Normal Pod with-cassandra-75866d478c-pqfqt.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:34:00 | examples-with-cassandra | 2024-09-12 07:33:41 +0000 UTC Normal Pod with-cassandra-75866d478c-pqfqt.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:34:00 | examples-with-cassandra | 2024-09-12 07:33:41 +0000 UTC Normal ReplicaSet.apps with-cassandra-75866d478c SuccessfulDelete Deleted pod: with-cassandra-75866d478c-pqfqt replicaset-controller logger.go:42: 07:34:00 | examples-with-cassandra | 2024-09-12 07:33:41 +0000 UTC Normal Deployment.apps with-cassandra ScalingReplicaSet Scaled down replica set with-cassandra-75866d478c to 0 from 1 deployment-controller logger.go:42: 07:34:00 | examples-with-cassandra | 2024-09-12 07:33:42 +0000 UTC Normal Pod with-cassandra-7b9fd5cf55-5dcvx Binding Scheduled Successfully assigned kuttl-test-valued-kite/with-cassandra-7b9fd5cf55-5dcvx to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:34:00 | examples-with-cassandra | 2024-09-12 07:33:42 +0000 UTC Normal ReplicaSet.apps with-cassandra-7b9fd5cf55 SuccessfulCreate Created pod: with-cassandra-7b9fd5cf55-5dcvx replicaset-controller logger.go:42: 07:34:00 | examples-with-cassandra | 2024-09-12 07:33:42 +0000 UTC Normal Deployment.apps with-cassandra ScalingReplicaSet Scaled up replica set with-cassandra-7b9fd5cf55 to 1 deployment-controller logger.go:42: 07:34:00 | examples-with-cassandra | 2024-09-12 07:33:43 +0000 UTC Normal Pod with-cassandra-7b9fd5cf55-5dcvx AddedInterface Add eth0 [10.130.0.84/23] from ovn-kubernetes logger.go:42: 07:34:00 | examples-with-cassandra | 2024-09-12 07:33:43 +0000 UTC Normal Pod with-cassandra-7b9fd5cf55-5dcvx.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" already present on machine kubelet logger.go:42: 07:34:00 | examples-with-cassandra | 2024-09-12 07:33:43 +0000 UTC Normal Pod with-cassandra-7b9fd5cf55-5dcvx.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:34:00 | examples-with-cassandra | 2024-09-12 07:33:43 +0000 UTC Normal Pod with-cassandra-7b9fd5cf55-5dcvx.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:34:00 | examples-with-cassandra | 2024-09-12 07:33:43 +0000 UTC Normal Pod with-cassandra-7b9fd5cf55-5dcvx.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:34:00 | examples-with-cassandra | 2024-09-12 07:33:43 +0000 UTC Normal Pod with-cassandra-7b9fd5cf55-5dcvx.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:34:00 | examples-with-cassandra | 2024-09-12 07:33:43 +0000 UTC Normal Pod with-cassandra-7b9fd5cf55-5dcvx.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:34:00 | examples-with-cassandra | 2024-09-12 07:33:48 +0000 UTC Normal Pod check-span-457dt Binding Scheduled Successfully assigned kuttl-test-valued-kite/check-span-457dt to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:34:00 | examples-with-cassandra | 2024-09-12 07:33:48 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-457dt job-controller logger.go:42: 07:34:00 | examples-with-cassandra | 2024-09-12 07:33:48 +0000 UTC Normal Pod report-span-qmd4x Binding Scheduled Successfully assigned kuttl-test-valued-kite/report-span-qmd4x to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:34:00 | examples-with-cassandra | 2024-09-12 07:33:48 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-qmd4x job-controller logger.go:42: 07:34:00 | examples-with-cassandra | 2024-09-12 07:33:49 +0000 UTC Normal Pod check-span-457dt AddedInterface Add eth0 [10.130.0.86/23] from ovn-kubernetes logger.go:42: 07:34:00 | examples-with-cassandra | 2024-09-12 07:33:49 +0000 UTC Normal Pod check-span-457dt.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:34:00 | examples-with-cassandra | 2024-09-12 07:33:49 +0000 UTC Normal Pod report-span-qmd4x AddedInterface Add eth0 [10.130.0.85/23] from ovn-kubernetes logger.go:42: 07:34:00 | examples-with-cassandra | 2024-09-12 07:33:49 +0000 UTC Normal Pod report-span-qmd4x.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:34:00 | examples-with-cassandra | 2024-09-12 07:33:50 +0000 UTC Normal Pod check-span-457dt.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 825ms (825ms including waiting) kubelet logger.go:42: 07:34:00 | examples-with-cassandra | 2024-09-12 07:33:50 +0000 UTC Normal Pod check-span-457dt.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:34:00 | examples-with-cassandra | 2024-09-12 07:33:50 +0000 UTC Normal Pod check-span-457dt.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:34:00 | examples-with-cassandra | 2024-09-12 07:33:50 +0000 UTC Normal Pod report-span-qmd4x.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 867ms (867ms including waiting) kubelet logger.go:42: 07:34:00 | examples-with-cassandra | 2024-09-12 07:33:50 +0000 UTC Normal Pod report-span-qmd4x.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:34:00 | examples-with-cassandra | 2024-09-12 07:33:50 +0000 UTC Normal Pod report-span-qmd4x.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:34:00 | examples-with-cassandra | 2024-09-12 07:34:00 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:34:01 | examples-with-cassandra | Deleting namespace: kuttl-test-valued-kite === CONT kuttl/harness/examples-with-badger logger.go:42: 07:34:13 | examples-with-badger | Creating namespace: kuttl-test-giving-flea logger.go:42: 07:34:13 | examples-with-badger/0-install | starting test step 0-install logger.go:42: 07:34:13 | examples-with-badger/0-install | Jaeger:kuttl-test-giving-flea/with-badger created logger.go:42: 07:34:16 | examples-with-badger/0-install | test step completed 0-install logger.go:42: 07:34:16 | examples-with-badger/1-smoke-test | starting test step 1-smoke-test logger.go:42: 07:34:16 | examples-with-badger/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE with-badger /dev/null] logger.go:42: 07:34:17 | examples-with-badger/1-smoke-test | Warning: resource jaegers/with-badger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:34:25 | examples-with-badger/1-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://with-badger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://with-badger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:34:25 | examples-with-badger/1-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:34:26 | examples-with-badger/1-smoke-test | job.batch/report-span created logger.go:42: 07:34:26 | examples-with-badger/1-smoke-test | job.batch/check-span created logger.go:42: 07:34:38 | examples-with-badger/1-smoke-test | test step completed 1-smoke-test logger.go:42: 07:34:38 | examples-with-badger | examples-with-badger events from ns kuttl-test-giving-flea: logger.go:42: 07:34:38 | examples-with-badger | 2024-09-12 07:34:14 +0000 UTC Normal Pod with-badger-6f78957695-rqgpb Binding Scheduled Successfully assigned kuttl-test-giving-flea/with-badger-6f78957695-rqgpb to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:34:38 | examples-with-badger | 2024-09-12 07:34:14 +0000 UTC Normal Pod with-badger-6f78957695-rqgpb AddedInterface Add eth0 [10.130.0.87/23] from ovn-kubernetes logger.go:42: 07:34:38 | examples-with-badger | 2024-09-12 07:34:14 +0000 UTC Normal Pod with-badger-6f78957695-rqgpb.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" already present on machine kubelet logger.go:42: 07:34:38 | examples-with-badger | 2024-09-12 07:34:14 +0000 UTC Normal Pod with-badger-6f78957695-rqgpb.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:34:38 | examples-with-badger | 2024-09-12 07:34:14 +0000 UTC Normal Pod with-badger-6f78957695-rqgpb.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:34:38 | examples-with-badger | 2024-09-12 07:34:14 +0000 UTC Normal Pod with-badger-6f78957695-rqgpb.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:34:38 | examples-with-badger | 2024-09-12 07:34:14 +0000 UTC Normal ReplicaSet.apps with-badger-6f78957695 SuccessfulCreate Created pod: with-badger-6f78957695-rqgpb replicaset-controller logger.go:42: 07:34:38 | examples-with-badger | 2024-09-12 07:34:14 +0000 UTC Normal Deployment.apps with-badger ScalingReplicaSet Scaled up replica set with-badger-6f78957695 to 1 deployment-controller logger.go:42: 07:34:38 | examples-with-badger | 2024-09-12 07:34:15 +0000 UTC Normal Pod with-badger-6f78957695-rqgpb.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:34:38 | examples-with-badger | 2024-09-12 07:34:15 +0000 UTC Normal Pod with-badger-6f78957695-rqgpb.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:34:38 | examples-with-badger | 2024-09-12 07:34:18 +0000 UTC Normal Pod with-badger-6f78957695-rqgpb.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:34:38 | examples-with-badger | 2024-09-12 07:34:18 +0000 UTC Normal Pod with-badger-6f78957695-rqgpb.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:34:38 | examples-with-badger | 2024-09-12 07:34:18 +0000 UTC Normal ReplicaSet.apps with-badger-6f78957695 SuccessfulDelete Deleted pod: with-badger-6f78957695-rqgpb replicaset-controller logger.go:42: 07:34:38 | examples-with-badger | 2024-09-12 07:34:18 +0000 UTC Normal Deployment.apps with-badger ScalingReplicaSet Scaled down replica set with-badger-6f78957695 to 0 from 1 deployment-controller logger.go:42: 07:34:38 | examples-with-badger | 2024-09-12 07:34:19 +0000 UTC Normal Pod with-badger-69596bffc-bkx7c Binding Scheduled Successfully assigned kuttl-test-giving-flea/with-badger-69596bffc-bkx7c to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:34:38 | examples-with-badger | 2024-09-12 07:34:19 +0000 UTC Normal ReplicaSet.apps with-badger-69596bffc SuccessfulCreate Created pod: with-badger-69596bffc-bkx7c replicaset-controller logger.go:42: 07:34:38 | examples-with-badger | 2024-09-12 07:34:19 +0000 UTC Normal Deployment.apps with-badger ScalingReplicaSet Scaled up replica set with-badger-69596bffc to 1 deployment-controller logger.go:42: 07:34:38 | examples-with-badger | 2024-09-12 07:34:20 +0000 UTC Normal Pod with-badger-69596bffc-bkx7c AddedInterface Add eth0 [10.130.0.88/23] from ovn-kubernetes logger.go:42: 07:34:38 | examples-with-badger | 2024-09-12 07:34:20 +0000 UTC Normal Pod with-badger-69596bffc-bkx7c.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" already present on machine kubelet logger.go:42: 07:34:38 | examples-with-badger | 2024-09-12 07:34:20 +0000 UTC Normal Pod with-badger-69596bffc-bkx7c.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:34:38 | examples-with-badger | 2024-09-12 07:34:20 +0000 UTC Normal Pod with-badger-69596bffc-bkx7c.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:34:38 | examples-with-badger | 2024-09-12 07:34:20 +0000 UTC Normal Pod with-badger-69596bffc-bkx7c.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:34:38 | examples-with-badger | 2024-09-12 07:34:20 +0000 UTC Normal Pod with-badger-69596bffc-bkx7c.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:34:38 | examples-with-badger | 2024-09-12 07:34:20 +0000 UTC Normal Pod with-badger-69596bffc-bkx7c.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:34:38 | examples-with-badger | 2024-09-12 07:34:26 +0000 UTC Normal Pod check-span-jj58h Binding Scheduled Successfully assigned kuttl-test-giving-flea/check-span-jj58h to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:34:38 | examples-with-badger | 2024-09-12 07:34:26 +0000 UTC Normal Pod check-span-jj58h AddedInterface Add eth0 [10.130.0.90/23] from ovn-kubernetes logger.go:42: 07:34:38 | examples-with-badger | 2024-09-12 07:34:26 +0000 UTC Normal Pod check-span-jj58h.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:34:38 | examples-with-badger | 2024-09-12 07:34:26 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-jj58h job-controller logger.go:42: 07:34:38 | examples-with-badger | 2024-09-12 07:34:26 +0000 UTC Normal Pod report-span-bj5bj Binding Scheduled Successfully assigned kuttl-test-giving-flea/report-span-bj5bj to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:34:38 | examples-with-badger | 2024-09-12 07:34:26 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-bj5bj job-controller logger.go:42: 07:34:38 | examples-with-badger | 2024-09-12 07:34:27 +0000 UTC Normal Pod check-span-jj58h.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 810ms (810ms including waiting) kubelet logger.go:42: 07:34:38 | examples-with-badger | 2024-09-12 07:34:27 +0000 UTC Normal Pod check-span-jj58h.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:34:38 | examples-with-badger | 2024-09-12 07:34:27 +0000 UTC Normal Pod check-span-jj58h.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:34:38 | examples-with-badger | 2024-09-12 07:34:27 +0000 UTC Normal Pod report-span-bj5bj AddedInterface Add eth0 [10.130.0.89/23] from ovn-kubernetes logger.go:42: 07:34:38 | examples-with-badger | 2024-09-12 07:34:27 +0000 UTC Normal Pod report-span-bj5bj.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:34:38 | examples-with-badger | 2024-09-12 07:34:27 +0000 UTC Normal Pod report-span-bj5bj.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 840ms (840ms including waiting) kubelet logger.go:42: 07:34:38 | examples-with-badger | 2024-09-12 07:34:27 +0000 UTC Normal Pod report-span-bj5bj.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:34:38 | examples-with-badger | 2024-09-12 07:34:27 +0000 UTC Normal Pod report-span-bj5bj.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:34:38 | examples-with-badger | 2024-09-12 07:34:37 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:34:38 | examples-with-badger | Deleting namespace: kuttl-test-giving-flea === CONT kuttl/harness/examples-simplest logger.go:42: 07:34:50 | examples-simplest | Creating namespace: kuttl-test-unbiased-arachnid logger.go:42: 07:34:50 | examples-simplest/0-install | starting test step 0-install logger.go:42: 07:34:51 | examples-simplest/0-install | Jaeger:kuttl-test-unbiased-arachnid/simplest created logger.go:42: 07:34:54 | examples-simplest/0-install | test step completed 0-install logger.go:42: 07:34:54 | examples-simplest/1-smoke-test | starting test step 1-smoke-test logger.go:42: 07:34:54 | examples-simplest/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE simplest /dev/null] logger.go:42: 07:34:56 | examples-simplest/1-smoke-test | Warning: resource jaegers/simplest is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:35:03 | examples-simplest/1-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://simplest-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:35:04 | examples-simplest/1-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:35:04 | examples-simplest/1-smoke-test | job.batch/report-span created logger.go:42: 07:35:04 | examples-simplest/1-smoke-test | job.batch/check-span created logger.go:42: 07:35:17 | examples-simplest/1-smoke-test | test step completed 1-smoke-test logger.go:42: 07:35:17 | examples-simplest | examples-simplest events from ns kuttl-test-unbiased-arachnid: logger.go:42: 07:35:17 | examples-simplest | 2024-09-12 07:34:51 +0000 UTC Normal Pod simplest-7478f7698d-6r8n2 Binding Scheduled Successfully assigned kuttl-test-unbiased-arachnid/simplest-7478f7698d-6r8n2 to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:35:17 | examples-simplest | 2024-09-12 07:34:51 +0000 UTC Normal ReplicaSet.apps simplest-7478f7698d SuccessfulCreate Created pod: simplest-7478f7698d-6r8n2 replicaset-controller logger.go:42: 07:35:17 | examples-simplest | 2024-09-12 07:34:51 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled up replica set simplest-7478f7698d to 1 deployment-controller logger.go:42: 07:35:17 | examples-simplest | 2024-09-12 07:34:52 +0000 UTC Normal Pod simplest-7478f7698d-6r8n2 AddedInterface Add eth0 [10.130.0.91/23] from ovn-kubernetes logger.go:42: 07:35:17 | examples-simplest | 2024-09-12 07:34:52 +0000 UTC Normal Pod simplest-7478f7698d-6r8n2.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" already present on machine kubelet logger.go:42: 07:35:17 | examples-simplest | 2024-09-12 07:34:52 +0000 UTC Normal Pod simplest-7478f7698d-6r8n2.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:35:17 | examples-simplest | 2024-09-12 07:34:52 +0000 UTC Normal Pod simplest-7478f7698d-6r8n2.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:35:17 | examples-simplest | 2024-09-12 07:34:52 +0000 UTC Normal Pod simplest-7478f7698d-6r8n2.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:35:17 | examples-simplest | 2024-09-12 07:34:52 +0000 UTC Normal Pod simplest-7478f7698d-6r8n2.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:35:17 | examples-simplest | 2024-09-12 07:34:52 +0000 UTC Normal Pod simplest-7478f7698d-6r8n2.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:35:17 | examples-simplest | 2024-09-12 07:34:57 +0000 UTC Normal Pod simplest-7478f7698d-6r8n2.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:35:17 | examples-simplest | 2024-09-12 07:34:57 +0000 UTC Normal Pod simplest-7478f7698d-6r8n2.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:35:17 | examples-simplest | 2024-09-12 07:34:57 +0000 UTC Normal ReplicaSet.apps simplest-7478f7698d SuccessfulDelete Deleted pod: simplest-7478f7698d-6r8n2 replicaset-controller logger.go:42: 07:35:17 | examples-simplest | 2024-09-12 07:34:57 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled down replica set simplest-7478f7698d to 0 from 1 deployment-controller logger.go:42: 07:35:17 | examples-simplest | 2024-09-12 07:34:57 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled up replica set simplest-78f459bbf5 to 1 deployment-controller logger.go:42: 07:35:17 | examples-simplest | 2024-09-12 07:34:58 +0000 UTC Normal Pod simplest-78f459bbf5-c29c8 Binding Scheduled Successfully assigned kuttl-test-unbiased-arachnid/simplest-78f459bbf5-c29c8 to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:35:17 | examples-simplest | 2024-09-12 07:34:58 +0000 UTC Normal Pod simplest-78f459bbf5-c29c8 AddedInterface Add eth0 [10.130.0.92/23] from ovn-kubernetes logger.go:42: 07:35:17 | examples-simplest | 2024-09-12 07:34:58 +0000 UTC Normal Pod simplest-78f459bbf5-c29c8.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" already present on machine kubelet logger.go:42: 07:35:17 | examples-simplest | 2024-09-12 07:34:58 +0000 UTC Normal Pod simplest-78f459bbf5-c29c8.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:35:17 | examples-simplest | 2024-09-12 07:34:58 +0000 UTC Normal Pod simplest-78f459bbf5-c29c8.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:35:17 | examples-simplest | 2024-09-12 07:34:58 +0000 UTC Normal Pod simplest-78f459bbf5-c29c8.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:35:17 | examples-simplest | 2024-09-12 07:34:58 +0000 UTC Normal Pod simplest-78f459bbf5-c29c8.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:35:17 | examples-simplest | 2024-09-12 07:34:58 +0000 UTC Normal Pod simplest-78f459bbf5-c29c8.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:35:17 | examples-simplest | 2024-09-12 07:34:58 +0000 UTC Normal ReplicaSet.apps simplest-78f459bbf5 SuccessfulCreate Created pod: simplest-78f459bbf5-c29c8 replicaset-controller logger.go:42: 07:35:17 | examples-simplest | 2024-09-12 07:35:04 +0000 UTC Normal Pod check-span-r7tmp Binding Scheduled Successfully assigned kuttl-test-unbiased-arachnid/check-span-r7tmp to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:35:17 | examples-simplest | 2024-09-12 07:35:04 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-r7tmp job-controller logger.go:42: 07:35:17 | examples-simplest | 2024-09-12 07:35:04 +0000 UTC Normal Pod report-span-lmcr4 Binding Scheduled Successfully assigned kuttl-test-unbiased-arachnid/report-span-lmcr4 to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:35:17 | examples-simplest | 2024-09-12 07:35:04 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-lmcr4 job-controller logger.go:42: 07:35:17 | examples-simplest | 2024-09-12 07:35:05 +0000 UTC Normal Pod check-span-r7tmp AddedInterface Add eth0 [10.130.0.94/23] from ovn-kubernetes logger.go:42: 07:35:17 | examples-simplest | 2024-09-12 07:35:05 +0000 UTC Normal Pod check-span-r7tmp.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:35:17 | examples-simplest | 2024-09-12 07:35:05 +0000 UTC Normal Pod report-span-lmcr4 AddedInterface Add eth0 [10.130.0.93/23] from ovn-kubernetes logger.go:42: 07:35:17 | examples-simplest | 2024-09-12 07:35:05 +0000 UTC Normal Pod report-span-lmcr4.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:35:17 | examples-simplest | 2024-09-12 07:35:06 +0000 UTC Normal Pod check-span-r7tmp.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 828ms (828ms including waiting) kubelet logger.go:42: 07:35:17 | examples-simplest | 2024-09-12 07:35:06 +0000 UTC Normal Pod check-span-r7tmp.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:35:17 | examples-simplest | 2024-09-12 07:35:06 +0000 UTC Normal Pod check-span-r7tmp.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:35:17 | examples-simplest | 2024-09-12 07:35:06 +0000 UTC Normal Pod report-span-lmcr4.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 855ms (855ms including waiting) kubelet logger.go:42: 07:35:17 | examples-simplest | 2024-09-12 07:35:06 +0000 UTC Normal Pod report-span-lmcr4.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:35:17 | examples-simplest | 2024-09-12 07:35:06 +0000 UTC Normal Pod report-span-lmcr4.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:35:17 | examples-simplest | 2024-09-12 07:35:17 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:35:17 | examples-simplest | Deleting namespace: kuttl-test-unbiased-arachnid === CONT kuttl/harness/examples-simple-prod-with-volumes logger.go:42: 07:35:30 | examples-simple-prod-with-volumes | Ignoring 03-check-volume.yaml.template as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:35:30 | examples-simple-prod-with-volumes | Creating namespace: kuttl-test-poetic-teal logger.go:42: 07:35:30 | examples-simple-prod-with-volumes/1-install | starting test step 1-install logger.go:42: 07:35:30 | examples-simple-prod-with-volumes/1-install | Jaeger:kuttl-test-poetic-teal/simple-prod created logger.go:42: 07:36:05 | examples-simple-prod-with-volumes/1-install | test step completed 1-install logger.go:42: 07:36:05 | examples-simple-prod-with-volumes/2-smoke-test | starting test step 2-smoke-test logger.go:42: 07:36:05 | examples-simple-prod-with-volumes/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE simple-prod /dev/null] logger.go:42: 07:36:06 | examples-simple-prod-with-volumes/2-smoke-test | Warning: resource jaegers/simple-prod is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:36:14 | examples-simple-prod-with-volumes/2-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:36:14 | examples-simple-prod-with-volumes/2-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:36:15 | examples-simple-prod-with-volumes/2-smoke-test | job.batch/report-span created logger.go:42: 07:36:15 | examples-simple-prod-with-volumes/2-smoke-test | job.batch/check-span created logger.go:42: 07:36:28 | examples-simple-prod-with-volumes/2-smoke-test | test step completed 2-smoke-test logger.go:42: 07:36:28 | examples-simple-prod-with-volumes/3-check-volume | starting test step 3-check-volume logger.go:42: 07:36:28 | examples-simple-prod-with-volumes/3-check-volume | running command: [sh -c kubectl exec $(kubectl get pods -n $NAMESPACE -l app=jaeger -l app.kubernetes.io/component=collector -o yaml | /tmp/jaeger-tests/bin/yq e '.items[0].metadata.name') -n $NAMESPACE -- ls /usr/share/elasticsearch/data] logger.go:42: 07:36:29 | examples-simple-prod-with-volumes/3-check-volume | test step completed 3-check-volume logger.go:42: 07:36:29 | examples-simple-prod-with-volumes | examples-simple-prod-with-volumes events from ns kuttl-test-poetic-teal: logger.go:42: 07:36:29 | examples-simple-prod-with-volumes | 2024-09-12 07:35:35 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestpoetictealsimpleprod-1-565c577cc5 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestpoetictealsimpleprod-1-565c577cwg7c7 replicaset-controller logger.go:42: 07:36:29 | examples-simple-prod-with-volumes | 2024-09-12 07:35:35 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpoetictealsimpleprod-1-565c577cwg7c7 Binding Scheduled Successfully assigned kuttl-test-poetic-teal/elasticsearch-cdm-kuttltestpoetictealsimpleprod-1-565c577cwg7c7 to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:36:29 | examples-simple-prod-with-volumes | 2024-09-12 07:35:35 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestpoetictealsimpleprod-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestpoetictealsimpleprod-1-565c577cc5 to 1 deployment-controller logger.go:42: 07:36:29 | examples-simple-prod-with-volumes | 2024-09-12 07:35:36 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpoetictealsimpleprod-1-565c577cwg7c7 AddedInterface Add eth0 [10.130.0.95/23] from ovn-kubernetes logger.go:42: 07:36:29 | examples-simple-prod-with-volumes | 2024-09-12 07:35:36 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpoetictealsimpleprod-1-565c577cwg7c7.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:bc97b8e13087050dfb1a75b02c5b780cbb8fb12a849a655169d072ba8bbf42b4" already present on machine kubelet logger.go:42: 07:36:29 | examples-simple-prod-with-volumes | 2024-09-12 07:35:36 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpoetictealsimpleprod-1-565c577cwg7c7.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:36:29 | examples-simple-prod-with-volumes | 2024-09-12 07:35:36 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpoetictealsimpleprod-1-565c577cwg7c7.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:36:29 | examples-simple-prod-with-volumes | 2024-09-12 07:35:36 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpoetictealsimpleprod-1-565c577cwg7c7.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:1d1d7d47b616995d18692f3a2d6232d7e1a0f41bc1503d87100f31939e080a78" already present on machine kubelet logger.go:42: 07:36:29 | examples-simple-prod-with-volumes | 2024-09-12 07:35:36 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpoetictealsimpleprod-1-565c577cwg7c7.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:36:29 | examples-simple-prod-with-volumes | 2024-09-12 07:35:36 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpoetictealsimpleprod-1-565c577cwg7c7.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:36:29 | examples-simple-prod-with-volumes | 2024-09-12 07:35:46 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestpoetictealsimpleprod-1-565c577cwg7c7.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 07:36:29 | examples-simple-prod-with-volumes | 2024-09-12 07:35:51 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestpoetictealsimpleprod-1-565c577cwg7c7.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:36:29 | examples-simple-prod-with-volumes | 2024-09-12 07:36:02 +0000 UTC Normal Pod simple-prod-collector-7c8b6c4bcf-9pvfk Binding Scheduled Successfully assigned kuttl-test-poetic-teal/simple-prod-collector-7c8b6c4bcf-9pvfk to ip-10-0-54-110.us-west-2.compute.internal default-scheduler logger.go:42: 07:36:29 | examples-simple-prod-with-volumes | 2024-09-12 07:36:02 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-7c8b6c4bcf SuccessfulCreate Created pod: simple-prod-collector-7c8b6c4bcf-9pvfk replicaset-controller logger.go:42: 07:36:29 | examples-simple-prod-with-volumes | 2024-09-12 07:36:02 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-7c8b6c4bcf to 1 deployment-controller logger.go:42: 07:36:29 | examples-simple-prod-with-volumes | 2024-09-12 07:36:02 +0000 UTC Normal Pod simple-prod-query-67fdf74d-d5ggp Binding Scheduled Successfully assigned kuttl-test-poetic-teal/simple-prod-query-67fdf74d-d5ggp to ip-10-0-52-42.us-west-2.compute.internal default-scheduler logger.go:42: 07:36:29 | examples-simple-prod-with-volumes | 2024-09-12 07:36:02 +0000 UTC Normal ReplicaSet.apps simple-prod-query-67fdf74d SuccessfulCreate Created pod: simple-prod-query-67fdf74d-d5ggp replicaset-controller logger.go:42: 07:36:29 | examples-simple-prod-with-volumes | 2024-09-12 07:36:02 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-67fdf74d to 1 deployment-controller logger.go:42: 07:36:29 | examples-simple-prod-with-volumes | 2024-09-12 07:36:03 +0000 UTC Normal Pod simple-prod-collector-7c8b6c4bcf-9pvfk AddedInterface Add eth0 [10.129.0.72/23] from ovn-kubernetes logger.go:42: 07:36:29 | examples-simple-prod-with-volumes | 2024-09-12 07:36:03 +0000 UTC Normal Pod simple-prod-collector-7c8b6c4bcf-9pvfk.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:4bab63290ff52e4f6328009f9c8b2c4496b230e9f8a98eac01736a66a291ff6c" already present on machine kubelet logger.go:42: 07:36:29 | examples-simple-prod-with-volumes | 2024-09-12 07:36:03 +0000 UTC Normal Pod simple-prod-collector-7c8b6c4bcf-9pvfk.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:36:29 | examples-simple-prod-with-volumes | 2024-09-12 07:36:03 +0000 UTC Normal Pod simple-prod-collector-7c8b6c4bcf-9pvfk.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:36:29 | examples-simple-prod-with-volumes | 2024-09-12 07:36:03 +0000 UTC Normal Pod simple-prod-query-67fdf74d-d5ggp AddedInterface Add eth0 [10.128.0.67/23] from ovn-kubernetes logger.go:42: 07:36:29 | examples-simple-prod-with-volumes | 2024-09-12 07:36:03 +0000 UTC Normal Pod simple-prod-query-67fdf74d-d5ggp.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a9186dcd910256c0f464b0a3928844a01de166a10c186c97ef4581bf288c23cb" already present on machine kubelet logger.go:42: 07:36:29 | examples-simple-prod-with-volumes | 2024-09-12 07:36:03 +0000 UTC Normal Pod simple-prod-query-67fdf74d-d5ggp.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:36:29 | examples-simple-prod-with-volumes | 2024-09-12 07:36:03 +0000 UTC Normal Pod simple-prod-query-67fdf74d-d5ggp.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:36:29 | examples-simple-prod-with-volumes | 2024-09-12 07:36:03 +0000 UTC Normal Pod simple-prod-query-67fdf74d-d5ggp.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:36:29 | examples-simple-prod-with-volumes | 2024-09-12 07:36:03 +0000 UTC Normal Pod simple-prod-query-67fdf74d-d5ggp.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:36:29 | examples-simple-prod-with-volumes | 2024-09-12 07:36:03 +0000 UTC Normal Pod simple-prod-query-67fdf74d-d5ggp.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:36:29 | examples-simple-prod-with-volumes | 2024-09-12 07:36:03 +0000 UTC Normal Pod simple-prod-query-67fdf74d-d5ggp.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" already present on machine kubelet logger.go:42: 07:36:29 | examples-simple-prod-with-volumes | 2024-09-12 07:36:03 +0000 UTC Normal Pod simple-prod-query-67fdf74d-d5ggp.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:36:29 | examples-simple-prod-with-volumes | 2024-09-12 07:36:04 +0000 UTC Normal Pod simple-prod-query-67fdf74d-d5ggp.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:36:29 | examples-simple-prod-with-volumes | 2024-09-12 07:36:07 +0000 UTC Normal ReplicaSet.apps simple-prod-query-67fdf74d SuccessfulDelete Deleted pod: simple-prod-query-67fdf74d-d5ggp replicaset-controller logger.go:42: 07:36:29 | examples-simple-prod-with-volumes | 2024-09-12 07:36:07 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled down replica set simple-prod-query-67fdf74d to 0 from 1 deployment-controller logger.go:42: 07:36:29 | examples-simple-prod-with-volumes | 2024-09-12 07:36:08 +0000 UTC Normal Pod simple-prod-query-67fdf74d-d5ggp.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:36:29 | examples-simple-prod-with-volumes | 2024-09-12 07:36:08 +0000 UTC Normal Pod simple-prod-query-67fdf74d-d5ggp.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:36:29 | examples-simple-prod-with-volumes | 2024-09-12 07:36:08 +0000 UTC Normal Pod simple-prod-query-67fdf74d-d5ggp.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:36:29 | examples-simple-prod-with-volumes | 2024-09-12 07:36:09 +0000 UTC Normal Pod simple-prod-query-695f7f65f5-rmlgk Binding Scheduled Successfully assigned kuttl-test-poetic-teal/simple-prod-query-695f7f65f5-rmlgk to ip-10-0-52-42.us-west-2.compute.internal default-scheduler logger.go:42: 07:36:29 | examples-simple-prod-with-volumes | 2024-09-12 07:36:09 +0000 UTC Normal Pod simple-prod-query-695f7f65f5-rmlgk AddedInterface Add eth0 [10.128.0.68/23] from ovn-kubernetes logger.go:42: 07:36:29 | examples-simple-prod-with-volumes | 2024-09-12 07:36:09 +0000 UTC Normal Pod simple-prod-query-695f7f65f5-rmlgk.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:a9186dcd910256c0f464b0a3928844a01de166a10c186c97ef4581bf288c23cb" already present on machine kubelet logger.go:42: 07:36:29 | examples-simple-prod-with-volumes | 2024-09-12 07:36:09 +0000 UTC Normal Pod simple-prod-query-695f7f65f5-rmlgk.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:36:29 | examples-simple-prod-with-volumes | 2024-09-12 07:36:09 +0000 UTC Normal Pod simple-prod-query-695f7f65f5-rmlgk.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:36:29 | examples-simple-prod-with-volumes | 2024-09-12 07:36:09 +0000 UTC Normal Pod simple-prod-query-695f7f65f5-rmlgk.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:36:29 | examples-simple-prod-with-volumes | 2024-09-12 07:36:09 +0000 UTC Normal Pod simple-prod-query-695f7f65f5-rmlgk.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:36:29 | examples-simple-prod-with-volumes | 2024-09-12 07:36:09 +0000 UTC Normal Pod simple-prod-query-695f7f65f5-rmlgk.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:36:29 | examples-simple-prod-with-volumes | 2024-09-12 07:36:09 +0000 UTC Normal Pod simple-prod-query-695f7f65f5-rmlgk.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" already present on machine kubelet logger.go:42: 07:36:29 | examples-simple-prod-with-volumes | 2024-09-12 07:36:09 +0000 UTC Normal Pod simple-prod-query-695f7f65f5-rmlgk.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:36:29 | examples-simple-prod-with-volumes | 2024-09-12 07:36:09 +0000 UTC Normal Pod simple-prod-query-695f7f65f5-rmlgk.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:36:29 | examples-simple-prod-with-volumes | 2024-09-12 07:36:09 +0000 UTC Normal ReplicaSet.apps simple-prod-query-695f7f65f5 SuccessfulCreate Created pod: simple-prod-query-695f7f65f5-rmlgk replicaset-controller logger.go:42: 07:36:29 | examples-simple-prod-with-volumes | 2024-09-12 07:36:09 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-695f7f65f5 to 1 deployment-controller logger.go:42: 07:36:29 | examples-simple-prod-with-volumes | 2024-09-12 07:36:15 +0000 UTC Normal Pod check-span-64klw Binding Scheduled Successfully assigned kuttl-test-poetic-teal/check-span-64klw to ip-10-0-54-110.us-west-2.compute.internal default-scheduler logger.go:42: 07:36:29 | examples-simple-prod-with-volumes | 2024-09-12 07:36:15 +0000 UTC Normal Pod check-span-64klw AddedInterface Add eth0 [10.129.0.73/23] from ovn-kubernetes logger.go:42: 07:36:29 | examples-simple-prod-with-volumes | 2024-09-12 07:36:15 +0000 UTC Normal Pod check-span-64klw.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:36:29 | examples-simple-prod-with-volumes | 2024-09-12 07:36:15 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-64klw job-controller logger.go:42: 07:36:29 | examples-simple-prod-with-volumes | 2024-09-12 07:36:15 +0000 UTC Normal Pod report-span-xjxl4 Binding Scheduled Successfully assigned kuttl-test-poetic-teal/report-span-xjxl4 to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:36:29 | examples-simple-prod-with-volumes | 2024-09-12 07:36:15 +0000 UTC Normal Pod report-span-xjxl4 AddedInterface Add eth0 [10.130.0.96/23] from ovn-kubernetes logger.go:42: 07:36:29 | examples-simple-prod-with-volumes | 2024-09-12 07:36:15 +0000 UTC Normal Pod report-span-xjxl4.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:36:29 | examples-simple-prod-with-volumes | 2024-09-12 07:36:15 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-xjxl4 job-controller logger.go:42: 07:36:29 | examples-simple-prod-with-volumes | 2024-09-12 07:36:16 +0000 UTC Normal Pod check-span-64klw.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 818ms (818ms including waiting) kubelet logger.go:42: 07:36:29 | examples-simple-prod-with-volumes | 2024-09-12 07:36:16 +0000 UTC Normal Pod check-span-64klw.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:36:29 | examples-simple-prod-with-volumes | 2024-09-12 07:36:16 +0000 UTC Normal Pod check-span-64klw.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:36:29 | examples-simple-prod-with-volumes | 2024-09-12 07:36:16 +0000 UTC Normal Pod report-span-xjxl4.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 842ms (842ms including waiting) kubelet logger.go:42: 07:36:29 | examples-simple-prod-with-volumes | 2024-09-12 07:36:16 +0000 UTC Normal Pod report-span-xjxl4.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:36:29 | examples-simple-prod-with-volumes | 2024-09-12 07:36:16 +0000 UTC Normal Pod report-span-xjxl4.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:36:29 | examples-simple-prod-with-volumes | 2024-09-12 07:36:20 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:36:29 | examples-simple-prod-with-volumes | 2024-09-12 07:36:20 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:36:29 | examples-simple-prod-with-volumes | 2024-09-12 07:36:20 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:36:29 | examples-simple-prod-with-volumes | 2024-09-12 07:36:27 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:36:29 | examples-simple-prod-with-volumes | Deleting namespace: kuttl-test-poetic-teal === CONT kuttl/harness/examples-business-application-injected-sidecar logger.go:42: 07:36:37 | examples-business-application-injected-sidecar | Creating namespace: kuttl-test-funny-mule logger.go:42: 07:36:37 | examples-business-application-injected-sidecar/0-install | starting test step 0-install logger.go:42: 07:36:37 | examples-business-application-injected-sidecar/0-install | Deployment:kuttl-test-funny-mule/myapp created logger.go:42: 07:36:37 | examples-business-application-injected-sidecar/0-install | test step completed 0-install logger.go:42: 07:36:37 | examples-business-application-injected-sidecar/1-install | starting test step 1-install logger.go:42: 07:36:37 | examples-business-application-injected-sidecar/1-install | Jaeger:kuttl-test-funny-mule/simplest created logger.go:42: 07:36:41 | examples-business-application-injected-sidecar/1-install | test step completed 1-install logger.go:42: 07:36:41 | examples-business-application-injected-sidecar/2-smoke-test | starting test step 2-smoke-test logger.go:42: 07:36:41 | examples-business-application-injected-sidecar/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE simplest /dev/null] logger.go:42: 07:36:42 | examples-business-application-injected-sidecar/2-smoke-test | Warning: resource jaegers/simplest is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:36:50 | examples-business-application-injected-sidecar/2-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://simplest-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:36:50 | examples-business-application-injected-sidecar/2-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:36:51 | examples-business-application-injected-sidecar/2-smoke-test | job.batch/report-span created logger.go:42: 07:36:51 | examples-business-application-injected-sidecar/2-smoke-test | job.batch/check-span created logger.go:42: 07:37:04 | examples-business-application-injected-sidecar/2-smoke-test | test step completed 2-smoke-test logger.go:42: 07:37:04 | examples-business-application-injected-sidecar | examples-business-application-injected-sidecar events from ns kuttl-test-funny-mule: logger.go:42: 07:37:04 | examples-business-application-injected-sidecar | 2024-09-12 07:36:37 +0000 UTC Normal Pod myapp-679f79d5f8-wdxr5 Binding Scheduled Successfully assigned kuttl-test-funny-mule/myapp-679f79d5f8-wdxr5 to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:37:04 | examples-business-application-injected-sidecar | 2024-09-12 07:36:37 +0000 UTC Normal Pod myapp-679f79d5f8-wdxr5 AddedInterface Add eth0 [10.130.0.97/23] from ovn-kubernetes logger.go:42: 07:37:04 | examples-business-application-injected-sidecar | 2024-09-12 07:36:37 +0000 UTC Normal Pod myapp-679f79d5f8-wdxr5.spec.containers{myapp} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:37:04 | examples-business-application-injected-sidecar | 2024-09-12 07:36:37 +0000 UTC Normal Pod myapp-679f79d5f8-wdxr5.spec.containers{myapp} Created Created container myapp kubelet logger.go:42: 07:37:04 | examples-business-application-injected-sidecar | 2024-09-12 07:36:37 +0000 UTC Normal Pod myapp-679f79d5f8-wdxr5.spec.containers{myapp} Started Started container myapp kubelet logger.go:42: 07:37:04 | examples-business-application-injected-sidecar | 2024-09-12 07:36:37 +0000 UTC Normal ReplicaSet.apps myapp-679f79d5f8 SuccessfulCreate Created pod: myapp-679f79d5f8-wdxr5 replicaset-controller logger.go:42: 07:37:04 | examples-business-application-injected-sidecar | 2024-09-12 07:36:37 +0000 UTC Normal Deployment.apps myapp ScalingReplicaSet Scaled up replica set myapp-679f79d5f8 to 1 deployment-controller logger.go:42: 07:37:04 | examples-business-application-injected-sidecar | 2024-09-12 07:36:37 +0000 UTC Normal Deployment.apps myapp ScalingReplicaSet Scaled up replica set myapp-564c69cc94 to 1 deployment-controller logger.go:42: 07:37:04 | examples-business-application-injected-sidecar | 2024-09-12 07:36:38 +0000 UTC Normal Pod myapp-564c69cc94-zppxr Binding Scheduled Successfully assigned kuttl-test-funny-mule/myapp-564c69cc94-zppxr to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:37:04 | examples-business-application-injected-sidecar | 2024-09-12 07:36:38 +0000 UTC Normal Pod myapp-564c69cc94-zppxr AddedInterface Add eth0 [10.130.0.98/23] from ovn-kubernetes logger.go:42: 07:37:04 | examples-business-application-injected-sidecar | 2024-09-12 07:36:38 +0000 UTC Normal Pod myapp-564c69cc94-zppxr.spec.containers{myapp} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:37:04 | examples-business-application-injected-sidecar | 2024-09-12 07:36:38 +0000 UTC Normal Pod myapp-564c69cc94-zppxr.spec.containers{myapp} Created Created container myapp kubelet logger.go:42: 07:37:04 | examples-business-application-injected-sidecar | 2024-09-12 07:36:38 +0000 UTC Normal Pod myapp-564c69cc94-zppxr.spec.containers{myapp} Started Started container myapp kubelet logger.go:42: 07:37:04 | examples-business-application-injected-sidecar | 2024-09-12 07:36:38 +0000 UTC Normal Pod myapp-564c69cc94-zppxr.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:69d728dd27fbd47fc667704adfa76746392f1f2331a927e5c436965d651ae147" already present on machine kubelet logger.go:42: 07:37:04 | examples-business-application-injected-sidecar | 2024-09-12 07:36:38 +0000 UTC Normal Pod myapp-564c69cc94-zppxr.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:37:04 | examples-business-application-injected-sidecar | 2024-09-12 07:36:38 +0000 UTC Normal Pod myapp-564c69cc94-zppxr.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:37:04 | examples-business-application-injected-sidecar | 2024-09-12 07:36:38 +0000 UTC Normal ReplicaSet.apps myapp-564c69cc94 SuccessfulCreate Created pod: myapp-564c69cc94-zppxr replicaset-controller logger.go:42: 07:37:04 | examples-business-application-injected-sidecar | 2024-09-12 07:36:38 +0000 UTC Normal Pod simplest-6588dd8b77-5f944 Binding Scheduled Successfully assigned kuttl-test-funny-mule/simplest-6588dd8b77-5f944 to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:37:04 | examples-business-application-injected-sidecar | 2024-09-12 07:36:38 +0000 UTC Normal Pod simplest-6588dd8b77-5f944 AddedInterface Add eth0 [10.130.0.99/23] from ovn-kubernetes logger.go:42: 07:37:04 | examples-business-application-injected-sidecar | 2024-09-12 07:36:38 +0000 UTC Normal Pod simplest-6588dd8b77-5f944.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" already present on machine kubelet logger.go:42: 07:37:04 | examples-business-application-injected-sidecar | 2024-09-12 07:36:38 +0000 UTC Normal ReplicaSet.apps simplest-6588dd8b77 SuccessfulCreate Created pod: simplest-6588dd8b77-5f944 replicaset-controller logger.go:42: 07:37:04 | examples-business-application-injected-sidecar | 2024-09-12 07:36:38 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled up replica set simplest-6588dd8b77 to 1 deployment-controller logger.go:42: 07:37:04 | examples-business-application-injected-sidecar | 2024-09-12 07:36:39 +0000 UTC Normal Pod simplest-6588dd8b77-5f944.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:37:04 | examples-business-application-injected-sidecar | 2024-09-12 07:36:39 +0000 UTC Normal Pod simplest-6588dd8b77-5f944.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:37:04 | examples-business-application-injected-sidecar | 2024-09-12 07:36:39 +0000 UTC Normal Pod simplest-6588dd8b77-5f944.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:37:04 | examples-business-application-injected-sidecar | 2024-09-12 07:36:39 +0000 UTC Normal Pod simplest-6588dd8b77-5f944.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:37:04 | examples-business-application-injected-sidecar | 2024-09-12 07:36:39 +0000 UTC Normal Pod simplest-6588dd8b77-5f944.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:37:04 | examples-business-application-injected-sidecar | 2024-09-12 07:36:40 +0000 UTC Normal Pod myapp-679f79d5f8-wdxr5.spec.containers{myapp} Killing Stopping container myapp kubelet logger.go:42: 07:37:04 | examples-business-application-injected-sidecar | 2024-09-12 07:36:40 +0000 UTC Normal ReplicaSet.apps myapp-679f79d5f8 SuccessfulDelete Deleted pod: myapp-679f79d5f8-wdxr5 replicaset-controller logger.go:42: 07:37:04 | examples-business-application-injected-sidecar | 2024-09-12 07:36:40 +0000 UTC Normal Deployment.apps myapp ScalingReplicaSet Scaled down replica set myapp-679f79d5f8 to 0 from 1 deployment-controller logger.go:42: 07:37:04 | examples-business-application-injected-sidecar | 2024-09-12 07:36:43 +0000 UTC Normal Pod simplest-6588dd8b77-5f944.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:37:04 | examples-business-application-injected-sidecar | 2024-09-12 07:36:43 +0000 UTC Normal Pod simplest-6588dd8b77-5f944.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:37:04 | examples-business-application-injected-sidecar | 2024-09-12 07:36:43 +0000 UTC Normal ReplicaSet.apps simplest-6588dd8b77 SuccessfulDelete Deleted pod: simplest-6588dd8b77-5f944 replicaset-controller logger.go:42: 07:37:04 | examples-business-application-injected-sidecar | 2024-09-12 07:36:43 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled down replica set simplest-6588dd8b77 to 0 from 1 deployment-controller logger.go:42: 07:37:04 | examples-business-application-injected-sidecar | 2024-09-12 07:36:44 +0000 UTC Normal Pod simplest-54c786656b-456cz Binding Scheduled Successfully assigned kuttl-test-funny-mule/simplest-54c786656b-456cz to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:37:04 | examples-business-application-injected-sidecar | 2024-09-12 07:36:44 +0000 UTC Normal Pod simplest-54c786656b-456cz AddedInterface Add eth0 [10.130.0.100/23] from ovn-kubernetes logger.go:42: 07:37:04 | examples-business-application-injected-sidecar | 2024-09-12 07:36:44 +0000 UTC Normal Pod simplest-54c786656b-456cz.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" already present on machine kubelet logger.go:42: 07:37:04 | examples-business-application-injected-sidecar | 2024-09-12 07:36:44 +0000 UTC Normal Pod simplest-54c786656b-456cz.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:37:04 | examples-business-application-injected-sidecar | 2024-09-12 07:36:44 +0000 UTC Normal Pod simplest-54c786656b-456cz.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:37:04 | examples-business-application-injected-sidecar | 2024-09-12 07:36:44 +0000 UTC Normal Pod simplest-54c786656b-456cz.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:37:04 | examples-business-application-injected-sidecar | 2024-09-12 07:36:44 +0000 UTC Normal Pod simplest-54c786656b-456cz.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:37:04 | examples-business-application-injected-sidecar | 2024-09-12 07:36:44 +0000 UTC Normal Pod simplest-54c786656b-456cz.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:37:04 | examples-business-application-injected-sidecar | 2024-09-12 07:36:44 +0000 UTC Normal ReplicaSet.apps simplest-54c786656b SuccessfulCreate Created pod: simplest-54c786656b-456cz replicaset-controller logger.go:42: 07:37:04 | examples-business-application-injected-sidecar | 2024-09-12 07:36:44 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled up replica set simplest-54c786656b to 1 deployment-controller logger.go:42: 07:37:04 | examples-business-application-injected-sidecar | 2024-09-12 07:36:49 +0000 UTC Warning Pod myapp-564c69cc94-zppxr.spec.containers{myapp} Unhealthy Liveness probe failed: Get "http://10.130.0.98:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:37:04 | examples-business-application-injected-sidecar | 2024-09-12 07:36:51 +0000 UTC Normal Pod check-span-t7t4r Binding Scheduled Successfully assigned kuttl-test-funny-mule/check-span-t7t4r to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:37:04 | examples-business-application-injected-sidecar | 2024-09-12 07:36:51 +0000 UTC Normal Pod check-span-t7t4r AddedInterface Add eth0 [10.130.0.102/23] from ovn-kubernetes logger.go:42: 07:37:04 | examples-business-application-injected-sidecar | 2024-09-12 07:36:51 +0000 UTC Normal Pod check-span-t7t4r.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:37:04 | examples-business-application-injected-sidecar | 2024-09-12 07:36:51 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-t7t4r job-controller logger.go:42: 07:37:04 | examples-business-application-injected-sidecar | 2024-09-12 07:36:51 +0000 UTC Normal Pod report-span-lq7v2 Binding Scheduled Successfully assigned kuttl-test-funny-mule/report-span-lq7v2 to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:37:04 | examples-business-application-injected-sidecar | 2024-09-12 07:36:51 +0000 UTC Normal Pod report-span-lq7v2 AddedInterface Add eth0 [10.130.0.101/23] from ovn-kubernetes logger.go:42: 07:37:04 | examples-business-application-injected-sidecar | 2024-09-12 07:36:51 +0000 UTC Normal Pod report-span-lq7v2.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:37:04 | examples-business-application-injected-sidecar | 2024-09-12 07:36:51 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-lq7v2 job-controller logger.go:42: 07:37:04 | examples-business-application-injected-sidecar | 2024-09-12 07:36:52 +0000 UTC Normal Pod check-span-t7t4r.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 890ms (890ms including waiting) kubelet logger.go:42: 07:37:04 | examples-business-application-injected-sidecar | 2024-09-12 07:36:52 +0000 UTC Normal Pod check-span-t7t4r.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:37:04 | examples-business-application-injected-sidecar | 2024-09-12 07:36:52 +0000 UTC Normal Pod check-span-t7t4r.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:37:04 | examples-business-application-injected-sidecar | 2024-09-12 07:36:52 +0000 UTC Normal Pod report-span-lq7v2.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 828ms (828ms including waiting) kubelet logger.go:42: 07:37:04 | examples-business-application-injected-sidecar | 2024-09-12 07:36:52 +0000 UTC Normal Pod report-span-lq7v2.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:37:04 | examples-business-application-injected-sidecar | 2024-09-12 07:36:52 +0000 UTC Normal Pod report-span-lq7v2.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:37:04 | examples-business-application-injected-sidecar | 2024-09-12 07:37:03 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:37:04 | examples-business-application-injected-sidecar | Deleting namespace: kuttl-test-funny-mule === CONT kuttl/harness/examples-service-types logger.go:42: 07:37:16 | examples-service-types | Creating namespace: kuttl-test-workable-pigeon logger.go:42: 07:37:16 | examples-service-types/0-install | starting test step 0-install logger.go:42: 07:37:17 | examples-service-types/0-install | Jaeger:kuttl-test-workable-pigeon/service-types created logger.go:42: 07:37:20 | examples-service-types/0-install | test step completed 0-install logger.go:42: 07:37:20 | examples-service-types/1-smoke-test | starting test step 1-smoke-test logger.go:42: 07:37:20 | examples-service-types/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE service-types /dev/null] logger.go:42: 07:37:22 | examples-service-types/1-smoke-test | Warning: resource jaegers/service-types is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:37:29 | examples-service-types/1-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://service-types-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://service-types-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:37:30 | examples-service-types/1-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:37:30 | examples-service-types/1-smoke-test | job.batch/report-span created logger.go:42: 07:37:30 | examples-service-types/1-smoke-test | job.batch/check-span created logger.go:42: 07:37:42 | examples-service-types/1-smoke-test | test step completed 1-smoke-test logger.go:42: 07:37:42 | examples-service-types/2- | starting test step 2- logger.go:42: 07:37:43 | examples-service-types/2- | test step completed 2- logger.go:42: 07:37:43 | examples-service-types | examples-service-types events from ns kuttl-test-workable-pigeon: logger.go:42: 07:37:43 | examples-service-types | 2024-09-12 07:37:17 +0000 UTC Normal Pod service-types-5d998675cf-g7q5t Binding Scheduled Successfully assigned kuttl-test-workable-pigeon/service-types-5d998675cf-g7q5t to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:37:43 | examples-service-types | 2024-09-12 07:37:17 +0000 UTC Normal ReplicaSet.apps service-types-5d998675cf SuccessfulCreate Created pod: service-types-5d998675cf-g7q5t replicaset-controller logger.go:42: 07:37:43 | examples-service-types | 2024-09-12 07:37:17 +0000 UTC Normal Service service-types-query EnsuringLoadBalancer Ensuring load balancer service-controller logger.go:42: 07:37:43 | examples-service-types | 2024-09-12 07:37:17 +0000 UTC Normal Deployment.apps service-types ScalingReplicaSet Scaled up replica set service-types-5d998675cf to 1 deployment-controller logger.go:42: 07:37:43 | examples-service-types | 2024-09-12 07:37:18 +0000 UTC Normal Pod service-types-5d998675cf-g7q5t AddedInterface Add eth0 [10.130.0.103/23] from ovn-kubernetes logger.go:42: 07:37:43 | examples-service-types | 2024-09-12 07:37:18 +0000 UTC Normal Pod service-types-5d998675cf-g7q5t.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" already present on machine kubelet logger.go:42: 07:37:43 | examples-service-types | 2024-09-12 07:37:18 +0000 UTC Normal Pod service-types-5d998675cf-g7q5t.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:37:43 | examples-service-types | 2024-09-12 07:37:18 +0000 UTC Normal Pod service-types-5d998675cf-g7q5t.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:37:43 | examples-service-types | 2024-09-12 07:37:18 +0000 UTC Normal Pod service-types-5d998675cf-g7q5t.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:37:43 | examples-service-types | 2024-09-12 07:37:18 +0000 UTC Normal Pod service-types-5d998675cf-g7q5t.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:37:43 | examples-service-types | 2024-09-12 07:37:18 +0000 UTC Normal Pod service-types-5d998675cf-g7q5t.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:37:43 | examples-service-types | 2024-09-12 07:37:20 +0000 UTC Normal Service service-types-collector EnsuringLoadBalancer Ensuring load balancer service-controller logger.go:42: 07:37:43 | examples-service-types | 2024-09-12 07:37:20 +0000 UTC Normal Service service-types-query EnsuredLoadBalancer Ensured load balancer service-controller logger.go:42: 07:37:43 | examples-service-types | 2024-09-12 07:37:22 +0000 UTC Normal ReplicaSet.apps service-types-5d998675cf SuccessfulDelete Deleted pod: service-types-5d998675cf-g7q5t replicaset-controller logger.go:42: 07:37:43 | examples-service-types | 2024-09-12 07:37:22 +0000 UTC Normal Deployment.apps service-types ScalingReplicaSet Scaled down replica set service-types-5d998675cf to 0 from 1 deployment-controller logger.go:42: 07:37:43 | examples-service-types | 2024-09-12 07:37:23 +0000 UTC Normal Pod service-types-5d998675cf-g7q5t.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:37:43 | examples-service-types | 2024-09-12 07:37:23 +0000 UTC Normal Pod service-types-5d998675cf-g7q5t.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:37:43 | examples-service-types | 2024-09-12 07:37:23 +0000 UTC Normal Service service-types-collector EnsuredLoadBalancer Ensured load balancer service-controller logger.go:42: 07:37:43 | examples-service-types | 2024-09-12 07:37:24 +0000 UTC Normal Pod service-types-86b4f6b87d-rrmdn Binding Scheduled Successfully assigned kuttl-test-workable-pigeon/service-types-86b4f6b87d-rrmdn to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:37:43 | examples-service-types | 2024-09-12 07:37:24 +0000 UTC Normal ReplicaSet.apps service-types-86b4f6b87d SuccessfulCreate Created pod: service-types-86b4f6b87d-rrmdn replicaset-controller logger.go:42: 07:37:43 | examples-service-types | 2024-09-12 07:37:24 +0000 UTC Normal Deployment.apps service-types ScalingReplicaSet Scaled up replica set service-types-86b4f6b87d to 1 deployment-controller logger.go:42: 07:37:43 | examples-service-types | 2024-09-12 07:37:25 +0000 UTC Normal Pod service-types-86b4f6b87d-rrmdn AddedInterface Add eth0 [10.130.0.104/23] from ovn-kubernetes logger.go:42: 07:37:43 | examples-service-types | 2024-09-12 07:37:25 +0000 UTC Normal Pod service-types-86b4f6b87d-rrmdn.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" already present on machine kubelet logger.go:42: 07:37:43 | examples-service-types | 2024-09-12 07:37:25 +0000 UTC Normal Pod service-types-86b4f6b87d-rrmdn.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:37:43 | examples-service-types | 2024-09-12 07:37:25 +0000 UTC Normal Pod service-types-86b4f6b87d-rrmdn.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:37:43 | examples-service-types | 2024-09-12 07:37:25 +0000 UTC Normal Pod service-types-86b4f6b87d-rrmdn.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:37:43 | examples-service-types | 2024-09-12 07:37:25 +0000 UTC Normal Pod service-types-86b4f6b87d-rrmdn.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:37:43 | examples-service-types | 2024-09-12 07:37:25 +0000 UTC Normal Pod service-types-86b4f6b87d-rrmdn.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:37:43 | examples-service-types | 2024-09-12 07:37:30 +0000 UTC Normal Pod check-span-jckd9 Binding Scheduled Successfully assigned kuttl-test-workable-pigeon/check-span-jckd9 to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:37:43 | examples-service-types | 2024-09-12 07:37:30 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-jckd9 job-controller logger.go:42: 07:37:43 | examples-service-types | 2024-09-12 07:37:30 +0000 UTC Normal Pod report-span-wk4dd Binding Scheduled Successfully assigned kuttl-test-workable-pigeon/report-span-wk4dd to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:37:43 | examples-service-types | 2024-09-12 07:37:30 +0000 UTC Normal Pod report-span-wk4dd AddedInterface Add eth0 [10.130.0.105/23] from ovn-kubernetes logger.go:42: 07:37:43 | examples-service-types | 2024-09-12 07:37:30 +0000 UTC Normal Pod report-span-wk4dd.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:37:43 | examples-service-types | 2024-09-12 07:37:30 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-wk4dd job-controller logger.go:42: 07:37:43 | examples-service-types | 2024-09-12 07:37:31 +0000 UTC Normal Pod check-span-jckd9 AddedInterface Add eth0 [10.130.0.106/23] from ovn-kubernetes logger.go:42: 07:37:43 | examples-service-types | 2024-09-12 07:37:31 +0000 UTC Normal Pod check-span-jckd9.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:37:43 | examples-service-types | 2024-09-12 07:37:31 +0000 UTC Normal Pod check-span-jckd9.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 818ms (818ms including waiting) kubelet logger.go:42: 07:37:43 | examples-service-types | 2024-09-12 07:37:31 +0000 UTC Normal Pod report-span-wk4dd.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 816ms (816ms including waiting) kubelet logger.go:42: 07:37:43 | examples-service-types | 2024-09-12 07:37:31 +0000 UTC Normal Pod report-span-wk4dd.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:37:43 | examples-service-types | 2024-09-12 07:37:31 +0000 UTC Normal Pod report-span-wk4dd.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:37:43 | examples-service-types | 2024-09-12 07:37:32 +0000 UTC Normal Pod check-span-jckd9.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:37:43 | examples-service-types | 2024-09-12 07:37:32 +0000 UTC Normal Pod check-span-jckd9.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:37:43 | examples-service-types | 2024-09-12 07:37:42 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:37:43 | examples-service-types | Deleting namespace: kuttl-test-workable-pigeon === CONT kuttl/harness/examples-openshift-with-htpasswd logger.go:42: 07:38:10 | examples-openshift-with-htpasswd | Ignoring 00-install.yaml.template as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:38:10 | examples-openshift-with-htpasswd | Ignoring ensure-ingress-host.sh as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:38:10 | examples-openshift-with-htpasswd | Creating namespace: kuttl-test-beloved-lacewing logger.go:42: 07:38:10 | examples-openshift-with-htpasswd/0-install | starting test step 0-install logger.go:42: 07:38:10 | examples-openshift-with-htpasswd/0-install | Secret:kuttl-test-beloved-lacewing/htpasswd created logger.go:42: 07:38:10 | examples-openshift-with-htpasswd/0-install | test step completed 0-install logger.go:42: 07:38:10 | examples-openshift-with-htpasswd/1-install | starting test step 1-install logger.go:42: 07:38:11 | examples-openshift-with-htpasswd/1-install | Jaeger:kuttl-test-beloved-lacewing/with-htpasswd created logger.go:42: 07:38:14 | examples-openshift-with-htpasswd/1-install | test step completed 1-install logger.go:42: 07:38:14 | examples-openshift-with-htpasswd/2-check-unsecured | starting test step 2-check-unsecured logger.go:42: 07:38:14 | examples-openshift-with-htpasswd/2-check-unsecured | running command: [./ensure-ingress-host.sh] logger.go:42: 07:38:14 | examples-openshift-with-htpasswd/2-check-unsecured | Checking the Ingress host value was populated logger.go:42: 07:38:14 | examples-openshift-with-htpasswd/2-check-unsecured | Try number 0 logger.go:42: 07:38:14 | examples-openshift-with-htpasswd/2-check-unsecured | Hostname is with-htpasswd-kuttl-test-beloved-lacewing.apps.rosa.ci-rosa-h-u0tm.f61o.s3.devshift.org logger.go:42: 07:38:14 | examples-openshift-with-htpasswd/2-check-unsecured | running command: [sh -c INSECURE=true ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 403 true $NAMESPACE with-htpasswd] logger.go:42: 07:38:14 | examples-openshift-with-htpasswd/2-check-unsecured | Checking an expected HTTP response logger.go:42: 07:38:14 | examples-openshift-with-htpasswd/2-check-unsecured | Running in OpenShift logger.go:42: 07:38:14 | examples-openshift-with-htpasswd/2-check-unsecured | Not using any secret logger.go:42: 07:38:14 | examples-openshift-with-htpasswd/2-check-unsecured | Try number 1/30 the https://with-htpasswd-kuttl-test-beloved-lacewing.apps.rosa.ci-rosa-h-u0tm.f61o.s3.devshift.org/search logger.go:42: 07:38:15 | examples-openshift-with-htpasswd/2-check-unsecured | HTTP response is 503. 403 expected. Waiting 10 s logger.go:42: 07:38:25 | examples-openshift-with-htpasswd/2-check-unsecured | Try number 2/30 the https://with-htpasswd-kuttl-test-beloved-lacewing.apps.rosa.ci-rosa-h-u0tm.f61o.s3.devshift.org/search logger.go:42: 07:38:25 | examples-openshift-with-htpasswd/2-check-unsecured | curl response asserted properly logger.go:42: 07:38:25 | examples-openshift-with-htpasswd/2-check-unsecured | test step completed 2-check-unsecured logger.go:42: 07:38:25 | examples-openshift-with-htpasswd/3-check-unauthorized | starting test step 3-check-unauthorized logger.go:42: 07:38:25 | examples-openshift-with-htpasswd/3-check-unauthorized | running command: [./ensure-ingress-host.sh] logger.go:42: 07:38:25 | examples-openshift-with-htpasswd/3-check-unauthorized | Checking the Ingress host value was populated logger.go:42: 07:38:25 | examples-openshift-with-htpasswd/3-check-unauthorized | Try number 0 logger.go:42: 07:38:25 | examples-openshift-with-htpasswd/3-check-unauthorized | Hostname is with-htpasswd-kuttl-test-beloved-lacewing.apps.rosa.ci-rosa-h-u0tm.f61o.s3.devshift.org logger.go:42: 07:38:25 | examples-openshift-with-htpasswd/3-check-unauthorized | running command: [sh -c JAEGER_USERNAME=wronguser JAEGER_PASSWORD=wrongpassword ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 403 true $NAMESPACE with-htpasswd] logger.go:42: 07:38:26 | examples-openshift-with-htpasswd/3-check-unauthorized | Checking an expected HTTP response logger.go:42: 07:38:26 | examples-openshift-with-htpasswd/3-check-unauthorized | Running in OpenShift logger.go:42: 07:38:26 | examples-openshift-with-htpasswd/3-check-unauthorized | Using Jaeger basic authentication logger.go:42: 07:38:26 | examples-openshift-with-htpasswd/3-check-unauthorized | Try number 1/30 the https://with-htpasswd-kuttl-test-beloved-lacewing.apps.rosa.ci-rosa-h-u0tm.f61o.s3.devshift.org/search logger.go:42: 07:38:26 | examples-openshift-with-htpasswd/3-check-unauthorized | curl response asserted properly logger.go:42: 07:38:26 | examples-openshift-with-htpasswd/3-check-unauthorized | test step completed 3-check-unauthorized logger.go:42: 07:38:26 | examples-openshift-with-htpasswd/4-check-authorized | starting test step 4-check-authorized logger.go:42: 07:38:26 | examples-openshift-with-htpasswd/4-check-authorized | running command: [./ensure-ingress-host.sh] logger.go:42: 07:38:26 | examples-openshift-with-htpasswd/4-check-authorized | Checking the Ingress host value was populated logger.go:42: 07:38:26 | examples-openshift-with-htpasswd/4-check-authorized | Try number 0 logger.go:42: 07:38:26 | examples-openshift-with-htpasswd/4-check-authorized | Hostname is with-htpasswd-kuttl-test-beloved-lacewing.apps.rosa.ci-rosa-h-u0tm.f61o.s3.devshift.org logger.go:42: 07:38:26 | examples-openshift-with-htpasswd/4-check-authorized | running command: [sh -c JAEGER_USERNAME=awesomeuser JAEGER_PASSWORD=awesomepassword ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 200 true $NAMESPACE with-htpasswd] logger.go:42: 07:38:27 | examples-openshift-with-htpasswd/4-check-authorized | Checking an expected HTTP response logger.go:42: 07:38:27 | examples-openshift-with-htpasswd/4-check-authorized | Running in OpenShift logger.go:42: 07:38:27 | examples-openshift-with-htpasswd/4-check-authorized | Using Jaeger basic authentication logger.go:42: 07:38:27 | examples-openshift-with-htpasswd/4-check-authorized | Try number 1/30 the https://with-htpasswd-kuttl-test-beloved-lacewing.apps.rosa.ci-rosa-h-u0tm.f61o.s3.devshift.org/search logger.go:42: 07:38:27 | examples-openshift-with-htpasswd/4-check-authorized | curl response asserted properly logger.go:42: 07:38:27 | examples-openshift-with-htpasswd/4-check-authorized | test step completed 4-check-authorized logger.go:42: 07:38:27 | examples-openshift-with-htpasswd | examples-openshift-with-htpasswd events from ns kuttl-test-beloved-lacewing: logger.go:42: 07:38:27 | examples-openshift-with-htpasswd | 2024-09-12 07:38:11 +0000 UTC Normal Pod with-htpasswd-7b89cd4bbd-h9d5s Binding Scheduled Successfully assigned kuttl-test-beloved-lacewing/with-htpasswd-7b89cd4bbd-h9d5s to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:38:27 | examples-openshift-with-htpasswd | 2024-09-12 07:38:11 +0000 UTC Normal Pod with-htpasswd-7b89cd4bbd-h9d5s AddedInterface Add eth0 [10.130.0.107/23] from ovn-kubernetes logger.go:42: 07:38:27 | examples-openshift-with-htpasswd | 2024-09-12 07:38:11 +0000 UTC Normal ReplicaSet.apps with-htpasswd-7b89cd4bbd SuccessfulCreate Created pod: with-htpasswd-7b89cd4bbd-h9d5s replicaset-controller logger.go:42: 07:38:27 | examples-openshift-with-htpasswd | 2024-09-12 07:38:11 +0000 UTC Normal Deployment.apps with-htpasswd ScalingReplicaSet Scaled up replica set with-htpasswd-7b89cd4bbd to 1 deployment-controller logger.go:42: 07:38:27 | examples-openshift-with-htpasswd | 2024-09-12 07:38:12 +0000 UTC Normal Pod with-htpasswd-7b89cd4bbd-h9d5s.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" already present on machine kubelet logger.go:42: 07:38:27 | examples-openshift-with-htpasswd | 2024-09-12 07:38:12 +0000 UTC Normal Pod with-htpasswd-7b89cd4bbd-h9d5s.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:38:27 | examples-openshift-with-htpasswd | 2024-09-12 07:38:12 +0000 UTC Normal Pod with-htpasswd-7b89cd4bbd-h9d5s.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:38:27 | examples-openshift-with-htpasswd | 2024-09-12 07:38:12 +0000 UTC Normal Pod with-htpasswd-7b89cd4bbd-h9d5s.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:38:27 | examples-openshift-with-htpasswd | 2024-09-12 07:38:12 +0000 UTC Normal Pod with-htpasswd-7b89cd4bbd-h9d5s.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:38:27 | examples-openshift-with-htpasswd | 2024-09-12 07:38:12 +0000 UTC Normal Pod with-htpasswd-7b89cd4bbd-h9d5s.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:38:27 | examples-openshift-with-htpasswd | Deleting namespace: kuttl-test-beloved-lacewing === CONT kuttl/harness/examples-all-in-one-with-options logger.go:42: 07:38:34 | examples-all-in-one-with-options | Creating namespace: kuttl-test-current-badger logger.go:42: 07:38:34 | examples-all-in-one-with-options/0-install | starting test step 0-install logger.go:42: 07:38:34 | examples-all-in-one-with-options/0-install | Jaeger:kuttl-test-current-badger/my-jaeger created logger.go:42: 07:38:36 | examples-all-in-one-with-options/0-install | test step completed 0-install logger.go:42: 07:38:36 | examples-all-in-one-with-options/1-smoke-test | starting test step 1-smoke-test logger.go:42: 07:38:36 | examples-all-in-one-with-options/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:38:38 | examples-all-in-one-with-options/1-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:38:46 | examples-all-in-one-with-options/1-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443/jaeger MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:38:47 | examples-all-in-one-with-options/1-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:38:47 | examples-all-in-one-with-options/1-smoke-test | job.batch/report-span created logger.go:42: 07:38:47 | examples-all-in-one-with-options/1-smoke-test | job.batch/check-span created logger.go:42: 07:38:59 | examples-all-in-one-with-options/1-smoke-test | test step completed 1-smoke-test logger.go:42: 07:38:59 | examples-all-in-one-with-options | examples-all-in-one-with-options events from ns kuttl-test-current-badger: logger.go:42: 07:38:59 | examples-all-in-one-with-options | 2024-09-12 07:38:35 +0000 UTC Normal Pod my-jaeger-67db5c78d9-dwbhg Binding Scheduled Successfully assigned kuttl-test-current-badger/my-jaeger-67db5c78d9-dwbhg to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:38:59 | examples-all-in-one-with-options | 2024-09-12 07:38:35 +0000 UTC Normal Pod my-jaeger-67db5c78d9-dwbhg AddedInterface Add eth0 [10.130.0.108/23] from ovn-kubernetes logger.go:42: 07:38:59 | examples-all-in-one-with-options | 2024-09-12 07:38:35 +0000 UTC Normal Pod my-jaeger-67db5c78d9-dwbhg.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" already present on machine kubelet logger.go:42: 07:38:59 | examples-all-in-one-with-options | 2024-09-12 07:38:35 +0000 UTC Normal Pod my-jaeger-67db5c78d9-dwbhg.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:38:59 | examples-all-in-one-with-options | 2024-09-12 07:38:35 +0000 UTC Normal Pod my-jaeger-67db5c78d9-dwbhg.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:38:59 | examples-all-in-one-with-options | 2024-09-12 07:38:35 +0000 UTC Normal Pod my-jaeger-67db5c78d9-dwbhg.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:38:59 | examples-all-in-one-with-options | 2024-09-12 07:38:35 +0000 UTC Normal Pod my-jaeger-67db5c78d9-dwbhg.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:38:59 | examples-all-in-one-with-options | 2024-09-12 07:38:35 +0000 UTC Normal Pod my-jaeger-67db5c78d9-dwbhg.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:38:59 | examples-all-in-one-with-options | 2024-09-12 07:38:35 +0000 UTC Normal ReplicaSet.apps my-jaeger-67db5c78d9 SuccessfulCreate Created pod: my-jaeger-67db5c78d9-dwbhg replicaset-controller logger.go:42: 07:38:59 | examples-all-in-one-with-options | 2024-09-12 07:38:35 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-67db5c78d9 to 1 deployment-controller logger.go:42: 07:38:59 | examples-all-in-one-with-options | 2024-09-12 07:38:39 +0000 UTC Normal Pod my-jaeger-67db5c78d9-dwbhg.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:38:59 | examples-all-in-one-with-options | 2024-09-12 07:38:39 +0000 UTC Normal Pod my-jaeger-67db5c78d9-dwbhg.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:38:59 | examples-all-in-one-with-options | 2024-09-12 07:38:39 +0000 UTC Normal ReplicaSet.apps my-jaeger-67db5c78d9 SuccessfulDelete Deleted pod: my-jaeger-67db5c78d9-dwbhg replicaset-controller logger.go:42: 07:38:59 | examples-all-in-one-with-options | 2024-09-12 07:38:39 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled down replica set my-jaeger-67db5c78d9 to 0 from 1 deployment-controller logger.go:42: 07:38:59 | examples-all-in-one-with-options | 2024-09-12 07:38:40 +0000 UTC Normal Pod my-jaeger-666986f879-9bh4h Binding Scheduled Successfully assigned kuttl-test-current-badger/my-jaeger-666986f879-9bh4h to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:38:59 | examples-all-in-one-with-options | 2024-09-12 07:38:40 +0000 UTC Normal ReplicaSet.apps my-jaeger-666986f879 SuccessfulCreate Created pod: my-jaeger-666986f879-9bh4h replicaset-controller logger.go:42: 07:38:59 | examples-all-in-one-with-options | 2024-09-12 07:38:40 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-666986f879 to 1 deployment-controller logger.go:42: 07:38:59 | examples-all-in-one-with-options | 2024-09-12 07:38:41 +0000 UTC Normal Pod my-jaeger-666986f879-9bh4h AddedInterface Add eth0 [10.130.0.109/23] from ovn-kubernetes logger.go:42: 07:38:59 | examples-all-in-one-with-options | 2024-09-12 07:38:41 +0000 UTC Normal Pod my-jaeger-666986f879-9bh4h.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1d8eef711323bbd14830846b3267011dd20cb1b15b84f16ce514e19c65531d34" already present on machine kubelet logger.go:42: 07:38:59 | examples-all-in-one-with-options | 2024-09-12 07:38:41 +0000 UTC Normal Pod my-jaeger-666986f879-9bh4h.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:38:59 | examples-all-in-one-with-options | 2024-09-12 07:38:41 +0000 UTC Normal Pod my-jaeger-666986f879-9bh4h.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:38:59 | examples-all-in-one-with-options | 2024-09-12 07:38:41 +0000 UTC Normal Pod my-jaeger-666986f879-9bh4h.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:38:59 | examples-all-in-one-with-options | 2024-09-12 07:38:41 +0000 UTC Normal Pod my-jaeger-666986f879-9bh4h.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:38:59 | examples-all-in-one-with-options | 2024-09-12 07:38:41 +0000 UTC Normal Pod my-jaeger-666986f879-9bh4h.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:38:59 | examples-all-in-one-with-options | 2024-09-12 07:38:47 +0000 UTC Normal Pod check-span-vww27 Binding Scheduled Successfully assigned kuttl-test-current-badger/check-span-vww27 to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:38:59 | examples-all-in-one-with-options | 2024-09-12 07:38:47 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-vww27 job-controller logger.go:42: 07:38:59 | examples-all-in-one-with-options | 2024-09-12 07:38:47 +0000 UTC Normal Pod report-span-7pspj Binding Scheduled Successfully assigned kuttl-test-current-badger/report-span-7pspj to ip-10-0-49-197.us-west-2.compute.internal default-scheduler logger.go:42: 07:38:59 | examples-all-in-one-with-options | 2024-09-12 07:38:47 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-7pspj job-controller logger.go:42: 07:38:59 | examples-all-in-one-with-options | 2024-09-12 07:38:48 +0000 UTC Normal Pod check-span-vww27 AddedInterface Add eth0 [10.130.0.111/23] from ovn-kubernetes logger.go:42: 07:38:59 | examples-all-in-one-with-options | 2024-09-12 07:38:48 +0000 UTC Normal Pod check-span-vww27.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:38:59 | examples-all-in-one-with-options | 2024-09-12 07:38:48 +0000 UTC Normal Pod report-span-7pspj AddedInterface Add eth0 [10.130.0.110/23] from ovn-kubernetes logger.go:42: 07:38:59 | examples-all-in-one-with-options | 2024-09-12 07:38:48 +0000 UTC Normal Pod report-span-7pspj.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:38:59 | examples-all-in-one-with-options | 2024-09-12 07:38:48 +0000 UTC Normal Pod report-span-7pspj.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 869ms (869ms including waiting) kubelet logger.go:42: 07:38:59 | examples-all-in-one-with-options | 2024-09-12 07:38:49 +0000 UTC Normal Pod check-span-vww27.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 965ms (965ms including waiting) kubelet logger.go:42: 07:38:59 | examples-all-in-one-with-options | 2024-09-12 07:38:49 +0000 UTC Normal Pod check-span-vww27.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:38:59 | examples-all-in-one-with-options | 2024-09-12 07:38:49 +0000 UTC Normal Pod check-span-vww27.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:38:59 | examples-all-in-one-with-options | 2024-09-12 07:38:49 +0000 UTC Normal Pod report-span-7pspj.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:38:59 | examples-all-in-one-with-options | 2024-09-12 07:38:49 +0000 UTC Normal Pod report-span-7pspj.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:38:59 | examples-all-in-one-with-options | 2024-09-12 07:38:59 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:38:59 | examples-all-in-one-with-options | Deleting namespace: kuttl-test-current-badger === CONT kuttl/harness/examples-agent-with-priority-class logger.go:42: 07:39:12 | examples-agent-with-priority-class | Creating namespace: kuttl-test-massive-chamois logger.go:42: 07:39:12 | examples-agent-with-priority-class/0-install | starting test step 0-install logger.go:42: 07:39:13 | examples-agent-with-priority-class/0-install | SecurityContextConstraints:/daemonset-with-hostport created logger.go:42: 07:39:13 | examples-agent-with-priority-class/0-install | ServiceAccount:kuttl-test-massive-chamois/jaeger-agent-daemonset created logger.go:42: 07:39:13 | examples-agent-with-priority-class/0-install | test step completed 0-install logger.go:42: 07:39:13 | examples-agent-with-priority-class/1-add-policy | starting test step 1-add-policy logger.go:42: 07:39:13 | examples-agent-with-priority-class/1-add-policy | running command: [sh -c oc adm policy --namespace $NAMESPACE add-scc-to-user daemonset-with-hostport -z jaeger-agent-daemonset] logger.go:42: 07:39:13 | examples-agent-with-priority-class/1-add-policy | clusterrole.rbac.authorization.k8s.io/system:openshift:scc:daemonset-with-hostport added: "jaeger-agent-daemonset" logger.go:42: 07:39:13 | examples-agent-with-priority-class/1-add-policy | running command: [sh -c sleep 5] logger.go:42: 07:39:19 | examples-agent-with-priority-class/1-add-policy | test step completed 1-add-policy logger.go:42: 07:39:19 | examples-agent-with-priority-class/2-install | starting test step 2-install logger.go:42: 07:39:19 | examples-agent-with-priority-class/2-install | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE agent-as-daemonset /dev/null] logger.go:42: 07:39:27 | examples-agent-with-priority-class/2-install | Error from server (NotFound): deployments.apps "agent-as-daemonset" not found logger.go:42: 07:39:27 | examples-agent-with-priority-class/2-install | command failure, skipping 2 additional commands logger.go:42: 07:39:28 | examples-agent-with-priority-class/2-install | PriorityClass:/high-priority created logger.go:42: 07:39:28 | examples-agent-with-priority-class/2-install | Jaeger:kuttl-test-massive-chamois/agent-as-daemonset updated case.go:364: failed in step 2-install case.go:366: exit status 1 logger.go:42: 07:39:28 | examples-agent-with-priority-class | examples-agent-with-priority-class events from ns kuttl-test-massive-chamois: logger.go:42: 07:39:28 | examples-agent-with-priority-class | Deleting namespace: kuttl-test-massive-chamois === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- FAIL: kuttl (614.88s) --- FAIL: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (6.10s) --- PASS: kuttl/harness/examples-simple-prod (66.71s) --- PASS: kuttl/harness/examples-with-sampling (161.26s) --- PASS: kuttl/harness/examples-with-cassandra (57.46s) --- PASS: kuttl/harness/examples-with-badger (37.31s) --- PASS: kuttl/harness/examples-simplest (39.41s) --- PASS: kuttl/harness/examples-simple-prod-with-volumes (66.91s) --- PASS: kuttl/harness/examples-business-application-injected-sidecar (39.81s) --- PASS: kuttl/harness/examples-service-types (53.49s) --- PASS: kuttl/harness/examples-openshift-with-htpasswd (23.74s) --- PASS: kuttl/harness/examples-all-in-one-with-options (38.71s) --- FAIL: kuttl/harness/examples-agent-with-priority-class (23.26s) FAIL + exit_code=1 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name examples --report --output /logs/artifacts/examples.xml ./artifacts/kuttl-report.xml time="2024-09-12T07:39:37Z" level=debug msg="Setting a new name for the test suites" time="2024-09-12T07:39:37Z" level=debug msg="Removing 'artifacts' TestCase" time="2024-09-12T07:39:37Z" level=debug msg="normalizing test case names" time="2024-09-12T07:39:37Z" level=debug msg="examples/artifacts -> examples_artifacts" time="2024-09-12T07:39:37Z" level=debug msg="examples/examples-simple-prod -> examples_examples_simple_prod" time="2024-09-12T07:39:37Z" level=debug msg="examples/examples-with-sampling -> examples_examples_with_sampling" time="2024-09-12T07:39:37Z" level=debug msg="examples/examples-with-cassandra -> examples_examples_with_cassandra" time="2024-09-12T07:39:37Z" level=debug msg="examples/examples-with-badger -> examples_examples_with_badger" time="2024-09-12T07:39:37Z" level=debug msg="examples/examples-simplest -> examples_examples_simplest" time="2024-09-12T07:39:37Z" level=debug msg="examples/examples-simple-prod-with-volumes -> examples_examples_simple_prod_with_volumes" time="2024-09-12T07:39:37Z" level=debug msg="examples/examples-business-application-injected-sidecar -> examples_examples_business_application_injected_sidecar" time="2024-09-12T07:39:37Z" level=debug msg="examples/examples-service-types -> examples_examples_service_types" time="2024-09-12T07:39:37Z" level=debug msg="examples/examples-openshift-with-htpasswd -> examples_examples_openshift_with_htpasswd" time="2024-09-12T07:39:37Z" level=debug msg="examples/examples-all-in-one-with-options -> examples_examples_all_in_one_with_options" time="2024-09-12T07:39:37Z" level=debug msg="examples/examples-agent-with-priority-class -> examples_examples_agent_with_priority_class" +---------------------------------------------------------+--------+ | NAME | RESULT | +---------------------------------------------------------+--------+ | examples_artifacts | passed | | examples_examples_simple_prod | passed | | examples_examples_with_sampling | passed | | examples_examples_with_cassandra | passed | | examples_examples_with_badger | passed | | examples_examples_simplest | passed | | examples_examples_simple_prod_with_volumes | passed | | examples_examples_business_application_injected_sidecar | passed | | examples_examples_service_types | passed | | examples_examples_openshift_with_htpasswd | passed | | examples_examples_all_in_one_with_options | passed | | examples_examples_agent_with_priority_class | failed | +---------------------------------------------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh generate false true + '[' 3 -ne 3 ']' + test_suite_name=generate + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/generate.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-generate make[2]: Entering directory '/tmp/jaeger-tests' test -s /tmp/jaeger-tests/bin/operator-sdk || curl -sLo /tmp/jaeger-tests/bin/operator-sdk https://github.com/operator-framework/operator-sdk/releases/download/v1.32.0/operator-sdk_`go env GOOS`_`go env GOARCH` ./hack/install/install-golangci-lint.sh Installing golangci-lint golangci-lint 1.55.2 is installed already ./hack/install/install-goimports.sh Installing goimports Try 0... go install golang.org/x/tools/cmd/goimports@v0.1.12 >>>> Formatting code... ./.ci/format.sh >>>> Building... ./hack/install/install-dependencies.sh Installing go dependencies Try 0... go mod download GOOS= GOARCH= CGO_ENABLED=0 GO111MODULE=on go build -ldflags "-X "github.com/jaegertracing/jaeger-operator/pkg/version".version="1.60.1" -X "github.com/jaegertracing/jaeger-operator/pkg/version".buildDate=2024-09-12T07:39:38Z -X "github.com/jaegertracing/jaeger-operator/pkg/version".defaultJaeger="1.60.0"" -o "bin/jaeger-operator" main.go JAEGER_VERSION="1.60.0" ./tests/e2e/generate/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.29 True False 51m Cluster version is 4.15.29' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.15.29 True False 51m Cluster version is 4.15.29' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/generate/render.sh ++ export SUITE_DIR=./tests/e2e/generate ++ SUITE_DIR=./tests/e2e/generate ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/generate ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + '[' true = true ']' + skip_test generate 'This test was skipped until https://github.com/jaegertracing/jaeger-operator/issues/2145 is fixed' + '[' 2 -ne 2 ']' + test_name=generate + message='This test was skipped until https://github.com/jaegertracing/jaeger-operator/issues/2145 is fixed' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/generate/_build + '[' _build '!=' _build ']' + rm -rf generate + warning 'generate: This test was skipped until https://github.com/jaegertracing/jaeger-operator/issues/2145 is fixed' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: generate: This test was skipped until https://github.com/jaegertracing/jaeger-operator/issues/2145 is fixed\e[0m' WAR: generate: This test was skipped until https://github.com/jaegertracing/jaeger-operator/issues/2145 is fixed make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running generate E2E tests' Running generate E2E tests + cd tests/e2e/generate/_build + set +e + KUBECONFIG=/tmp/kubeconfig-2486720052 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-rosa-h-u0tm.f61o.s3.devshift.org:443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 1 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === CONT kuttl/harness/artifacts logger.go:42: 07:39:54 | artifacts | Creating namespace: kuttl-test-leading-kid logger.go:42: 07:39:54 | artifacts | artifacts events from ns kuttl-test-leading-kid: logger.go:42: 07:39:54 | artifacts | Deleting namespace: kuttl-test-leading-kid === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (7.93s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (7.33s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name generate --report --output /logs/artifacts/generate.xml ./artifacts/kuttl-report.xml time="2024-09-12T07:40:02Z" level=debug msg="Setting a new name for the test suites" time="2024-09-12T07:40:02Z" level=debug msg="Removing 'artifacts' TestCase" time="2024-09-12T07:40:02Z" level=debug msg="normalizing test case names" time="2024-09-12T07:40:02Z" level=debug msg="generate/artifacts -> generate_artifacts" +--------------------+--------+ | NAME | RESULT | +--------------------+--------+ | generate_artifacts | passed | +--------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests'