% Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 3831 100 3831 0 0 23687 0 --:--:-- --:--:-- --:--:-- 23795 % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 9315 100 9315 0 0 48742 0 --:--:-- --:--:-- --:--:-- 48769 % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 2607 100 2607 0 0 15959 0 --:--:-- --:--:-- --:--:-- 15896 100 2607 100 2607 0 0 15953 0 --:--:-- --:--:-- --:--:-- 15896 % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 1953 100 1953 0 0 13043 0 --:--:-- --:--:-- --:--:-- 13107 % Total % Received % Xferd Average Speed Time Time Time Current Dload Upload Total Spent Left Speed 0 0 0 0 0 0 0 0 --:--:-- --:--:-- --:--:-- 0 100 350 100 350 0 0 2327 0 --:--:-- --:--:-- --:--:-- 2333 Installing kuttl Try 0... curl -sLo /tmp/jaeger-tests/hack/install/../../bin/kubectl-kuttl https://github.com/kudobuilder/kuttl/releases/download/v0.15.0/kubectl-kuttl_0.15.0_linux_x86_64 KUBECONFIG file is: /tmp/kubeconfig-591533165 for suite in sidecar streaming ui miscellaneous elasticsearch examples generate upgrade; do \ make run-e2e-tests-$suite ; \ done make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh sidecar false true + '[' 3 -ne 3 ']' + test_suite_name=sidecar + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/sidecar.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-sidecar make[2]: Entering directory '/tmp/jaeger-tests' ./tests/e2e/sidecar/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.18.0-0.nightly-2024-12-05-191355 True False 7m20s Cluster version is 4.18.0-0.nightly-2024-12-05-191355' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.18.0-0.nightly-2024-12-05-191355 True False 7m20s Cluster version is 4.18.0-0.nightly-2024-12-05-191355' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/sidecar/render.sh ++ export SUITE_DIR=./tests/e2e/sidecar ++ SUITE_DIR=./tests/e2e/sidecar ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/sidecar ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + jaeger_service_name=order + start_test sidecar-deployment + '[' 1 -ne 1 ']' + test_name=sidecar-deployment + echo =========================================================================== =========================================================================== + info 'Rendering files for test sidecar-deployment' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test sidecar-deployment\e[0m' Rendering files for test sidecar-deployment + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/sidecar/_build + '[' _build '!=' _build ']' + mkdir -p sidecar-deployment + cd sidecar-deployment + render_install_vertx 01 + '[' 1 -ne 1 ']' + test_step=01 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-assert.yaml.template -o ./01-assert.yaml + render_find_service agent-as-sidecar allInOne order 00 03 + '[' 5 -ne 5 ']' + jaeger=agent-as-sidecar + deployment_strategy=allInOne + service_name=order + job_number=00 + test_step=03 + export JAEGER_NAME=agent-as-sidecar + JAEGER_NAME=agent-as-sidecar + export JOB_NUMBER=00 + JOB_NUMBER=00 + export SERVICE_NAME=order + SERVICE_NAME=order + export JAEGER_QUERY_ENDPOINT + '[' true = true ']' + '[' allInOne '!=' allInOne ']' + template=/tmp/jaeger-tests/tests/templates/find-service.yaml.template + JAEGER_QUERY_ENDPOINT=http://agent-as-sidecar-query:16686 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/find-service.yaml.template -o ./03-find-service.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-find-service.yaml.template -o ./03-assert.yaml + unset JAEGER_NAME + unset SERVICE_NAME + unset JOB_NUMBER + unset JAEGER_COLLECTOR_ENDPOINT + render_find_service agent-as-sidecar2 allInOne order 01 06 + '[' 5 -ne 5 ']' + jaeger=agent-as-sidecar2 + deployment_strategy=allInOne + service_name=order + job_number=01 + test_step=06 + export JAEGER_NAME=agent-as-sidecar2 + JAEGER_NAME=agent-as-sidecar2 + export JOB_NUMBER=01 + JOB_NUMBER=01 + export SERVICE_NAME=order + SERVICE_NAME=order + export JAEGER_QUERY_ENDPOINT + '[' true = true ']' + '[' allInOne '!=' allInOne ']' + template=/tmp/jaeger-tests/tests/templates/find-service.yaml.template + JAEGER_QUERY_ENDPOINT=http://agent-as-sidecar2-query:16686 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/find-service.yaml.template -o ./06-find-service.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-find-service.yaml.template -o ./06-assert.yaml + unset JAEGER_NAME + unset SERVICE_NAME + unset JOB_NUMBER + unset JAEGER_COLLECTOR_ENDPOINT + start_test sidecar-namespace + '[' 1 -ne 1 ']' + test_name=sidecar-namespace + echo =========================================================================== =========================================================================== + info 'Rendering files for test sidecar-namespace' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test sidecar-namespace\e[0m' Rendering files for test sidecar-namespace + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/sidecar/_build/sidecar-deployment + '[' sidecar-deployment '!=' _build ']' + cd .. + mkdir -p sidecar-namespace + cd sidecar-namespace + render_install_vertx 01 + '[' 1 -ne 1 ']' + test_step=01 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-assert.yaml.template -o ./01-assert.yaml + render_find_service agent-as-sidecar allInOne order 00 03 + '[' 5 -ne 5 ']' + jaeger=agent-as-sidecar + deployment_strategy=allInOne + service_name=order + job_number=00 + test_step=03 + export JAEGER_NAME=agent-as-sidecar + JAEGER_NAME=agent-as-sidecar + export JOB_NUMBER=00 + JOB_NUMBER=00 + export SERVICE_NAME=order + SERVICE_NAME=order + export JAEGER_QUERY_ENDPOINT + '[' true = true ']' + '[' allInOne '!=' allInOne ']' + template=/tmp/jaeger-tests/tests/templates/find-service.yaml.template + JAEGER_QUERY_ENDPOINT=http://agent-as-sidecar-query:16686 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/find-service.yaml.template -o ./03-find-service.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-find-service.yaml.template -o ./03-assert.yaml + unset JAEGER_NAME + unset SERVICE_NAME + unset JOB_NUMBER + unset JAEGER_COLLECTOR_ENDPOINT + render_find_service agent-as-sidecar2 allInOne order 01 06 + '[' 5 -ne 5 ']' + jaeger=agent-as-sidecar2 + deployment_strategy=allInOne + service_name=order + job_number=01 + test_step=06 + export JAEGER_NAME=agent-as-sidecar2 + JAEGER_NAME=agent-as-sidecar2 + export JOB_NUMBER=01 + JOB_NUMBER=01 + export SERVICE_NAME=order + SERVICE_NAME=order + export JAEGER_QUERY_ENDPOINT + '[' true = true ']' + '[' allInOne '!=' allInOne ']' + template=/tmp/jaeger-tests/tests/templates/find-service.yaml.template + JAEGER_QUERY_ENDPOINT=http://agent-as-sidecar2-query:16686 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/find-service.yaml.template -o ./06-find-service.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-find-service.yaml.template -o ./06-assert.yaml + unset JAEGER_NAME + unset SERVICE_NAME + unset JOB_NUMBER + unset JAEGER_COLLECTOR_ENDPOINT + start_test sidecar-skip-webhook + '[' 1 -ne 1 ']' + test_name=sidecar-skip-webhook + echo =========================================================================== =========================================================================== + info 'Rendering files for test sidecar-skip-webhook' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test sidecar-skip-webhook\e[0m' Rendering files for test sidecar-skip-webhook + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/sidecar/_build/sidecar-namespace + '[' sidecar-namespace '!=' _build ']' + cd .. + mkdir -p sidecar-skip-webhook + cd sidecar-skip-webhook + render_install_vertx 01 + '[' 1 -ne 1 ']' + test_step=01 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/vertex-assert.yaml.template -o ./01-assert.yaml make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running sidecar E2E tests' Running sidecar E2E tests + cd tests/e2e/sidecar/_build + set +e + KUBECONFIG=/tmp/kubeconfig-591533165 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-54q9kfc5-9210b.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 4 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/sidecar-deployment === PAUSE kuttl/harness/sidecar-deployment === RUN kuttl/harness/sidecar-namespace === PAUSE kuttl/harness/sidecar-namespace === RUN kuttl/harness/sidecar-skip-webhook === PAUSE kuttl/harness/sidecar-skip-webhook === CONT kuttl/harness/artifacts logger.go:42: 07:12:57 | artifacts | Creating namespace: kuttl-test-holy-pig logger.go:42: 07:12:57 | artifacts | artifacts events from ns kuttl-test-holy-pig: logger.go:42: 07:12:57 | artifacts | Deleting namespace: kuttl-test-holy-pig === CONT kuttl/harness/sidecar-namespace logger.go:42: 07:13:03 | sidecar-namespace | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:13:03 | sidecar-namespace | Creating namespace: kuttl-test-viable-quagga logger.go:42: 07:13:03 | sidecar-namespace/0-install | starting test step 0-install logger.go:42: 07:13:03 | sidecar-namespace/0-install | Jaeger:kuttl-test-viable-quagga/agent-as-sidecar created logger.go:42: 07:13:14 | sidecar-namespace/0-install | test step completed 0-install logger.go:42: 07:13:14 | sidecar-namespace/1-install | starting test step 1-install logger.go:42: 07:13:14 | sidecar-namespace/1-install | Deployment:kuttl-test-viable-quagga/vertx-create-span-sidecar created logger.go:42: 07:13:19 | sidecar-namespace/1-install | test step completed 1-install logger.go:42: 07:13:19 | sidecar-namespace/2-enable-injection | starting test step 2-enable-injection logger.go:42: 07:13:19 | sidecar-namespace/2-enable-injection | running command: [sh -c kubectl annotate --overwrite namespaces $NAMESPACE "sidecar.jaegertracing.io/inject"="true"] logger.go:42: 07:13:19 | sidecar-namespace/2-enable-injection | namespace/kuttl-test-viable-quagga annotated logger.go:42: 07:13:28 | sidecar-namespace/2-enable-injection | test step completed 2-enable-injection logger.go:42: 07:13:28 | sidecar-namespace/3-find-service | starting test step 3-find-service logger.go:42: 07:13:29 | sidecar-namespace/3-find-service | Job:kuttl-test-viable-quagga/00-find-service created logger.go:42: 07:13:36 | sidecar-namespace/3-find-service | test step completed 3-find-service logger.go:42: 07:13:36 | sidecar-namespace/4-other-instance | starting test step 4-other-instance logger.go:42: 07:13:36 | sidecar-namespace/4-other-instance | Jaeger:kuttl-test-viable-quagga/agent-as-sidecar2 created logger.go:42: 07:13:49 | sidecar-namespace/4-other-instance | test step completed 4-other-instance logger.go:42: 07:13:49 | sidecar-namespace/5-delete-first-instance | starting test step 5-delete-first-instance logger.go:42: 07:13:49 | sidecar-namespace/5-delete-first-instance | test step completed 5-delete-first-instance logger.go:42: 07:13:49 | sidecar-namespace/6-find-service | starting test step 6-find-service logger.go:42: 07:13:49 | sidecar-namespace/6-find-service | Job:kuttl-test-viable-quagga/01-find-service created logger.go:42: 07:14:02 | sidecar-namespace/6-find-service | test step completed 6-find-service logger.go:42: 07:14:02 | sidecar-namespace/7-disable-injection | starting test step 7-disable-injection logger.go:42: 07:14:02 | sidecar-namespace/7-disable-injection | running command: [sh -c kubectl annotate --overwrite namespaces $NAMESPACE "sidecar.jaegertracing.io/inject"="false"] logger.go:42: 07:14:03 | sidecar-namespace/7-disable-injection | namespace/kuttl-test-viable-quagga annotated logger.go:42: 07:14:05 | sidecar-namespace/7-disable-injection | test step completed 7-disable-injection logger.go:42: 07:14:05 | sidecar-namespace | sidecar-namespace events from ns kuttl-test-viable-quagga: logger.go:42: 07:14:05 | sidecar-namespace | 2024-12-09 07:13:07 +0000 UTC Normal Pod agent-as-sidecar-54bb7ccbc4-5s4g8 Binding Scheduled Successfully assigned kuttl-test-viable-quagga/agent-as-sidecar-54bb7ccbc4-5s4g8 to ip-10-0-95-250.ec2.internal default-scheduler logger.go:42: 07:14:05 | sidecar-namespace | 2024-12-09 07:13:07 +0000 UTC Normal ReplicaSet.apps agent-as-sidecar-54bb7ccbc4 SuccessfulCreate Created pod: agent-as-sidecar-54bb7ccbc4-5s4g8 replicaset-controller logger.go:42: 07:14:05 | sidecar-namespace | 2024-12-09 07:13:07 +0000 UTC Normal Deployment.apps agent-as-sidecar ScalingReplicaSet Scaled up replica set agent-as-sidecar-54bb7ccbc4 to 1 deployment-controller logger.go:42: 07:14:05 | sidecar-namespace | 2024-12-09 07:13:08 +0000 UTC Normal Pod agent-as-sidecar-54bb7ccbc4-5s4g8 AddedInterface Add eth0 [10.131.0.16/23] from ovn-kubernetes multus logger.go:42: 07:14:05 | sidecar-namespace | 2024-12-09 07:13:08 +0000 UTC Normal Pod agent-as-sidecar-54bb7ccbc4-5s4g8.spec.containers{jaeger} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1fa5ed13cc8a023f1e987f6cafe86adc195c373cc2b774539df6d0fd02b780a7" kubelet logger.go:42: 07:14:05 | sidecar-namespace | 2024-12-09 07:13:11 +0000 UTC Normal Pod agent-as-sidecar-54bb7ccbc4-5s4g8.spec.containers{jaeger} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1fa5ed13cc8a023f1e987f6cafe86adc195c373cc2b774539df6d0fd02b780a7" in 3.53s (3.53s including waiting). Image size: 147411949 bytes. kubelet logger.go:42: 07:14:05 | sidecar-namespace | 2024-12-09 07:13:12 +0000 UTC Normal Pod agent-as-sidecar-54bb7ccbc4-5s4g8.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:14:05 | sidecar-namespace | 2024-12-09 07:13:12 +0000 UTC Normal Pod agent-as-sidecar-54bb7ccbc4-5s4g8.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:14:05 | sidecar-namespace | 2024-12-09 07:13:14 +0000 UTC Normal Pod vertx-create-span-sidecar-797645c8fc-8tfpw Binding Scheduled Successfully assigned kuttl-test-viable-quagga/vertx-create-span-sidecar-797645c8fc-8tfpw to ip-10-0-79-227.ec2.internal default-scheduler logger.go:42: 07:14:05 | sidecar-namespace | 2024-12-09 07:13:14 +0000 UTC Normal Pod vertx-create-span-sidecar-797645c8fc-8tfpw AddedInterface Add eth0 [10.129.2.23/23] from ovn-kubernetes multus logger.go:42: 07:14:05 | sidecar-namespace | 2024-12-09 07:13:14 +0000 UTC Normal Pod vertx-create-span-sidecar-797645c8fc-8tfpw.spec.containers{vertx-create-span-sidecar} Pulling Pulling image "jaegertracing/vertx-create-span:operator-e2e-tests" kubelet logger.go:42: 07:14:05 | sidecar-namespace | 2024-12-09 07:13:14 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-797645c8fc SuccessfulCreate Created pod: vertx-create-span-sidecar-797645c8fc-8tfpw replicaset-controller logger.go:42: 07:14:05 | sidecar-namespace | 2024-12-09 07:13:14 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-797645c8fc to 1 deployment-controller logger.go:42: 07:14:05 | sidecar-namespace | 2024-12-09 07:13:18 +0000 UTC Normal Pod vertx-create-span-sidecar-797645c8fc-8tfpw.spec.containers{vertx-create-span-sidecar} Pulled Successfully pulled image "jaegertracing/vertx-create-span:operator-e2e-tests" in 4.083s (4.083s including waiting). Image size: 282912835 bytes. kubelet logger.go:42: 07:14:05 | sidecar-namespace | 2024-12-09 07:13:19 +0000 UTC Normal Pod vertx-create-span-sidecar-6bdf95c49c-hwrlp Binding Scheduled Successfully assigned kuttl-test-viable-quagga/vertx-create-span-sidecar-6bdf95c49c-hwrlp to ip-10-0-95-250.ec2.internal default-scheduler logger.go:42: 07:14:05 | sidecar-namespace | 2024-12-09 07:13:19 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-6bdf95c49c SuccessfulCreate Created pod: vertx-create-span-sidecar-6bdf95c49c-hwrlp replicaset-controller logger.go:42: 07:14:05 | sidecar-namespace | 2024-12-09 07:13:19 +0000 UTC Normal Pod vertx-create-span-sidecar-797645c8fc-8tfpw.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:14:05 | sidecar-namespace | 2024-12-09 07:13:19 +0000 UTC Normal Pod vertx-create-span-sidecar-797645c8fc-8tfpw.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:14:05 | sidecar-namespace | 2024-12-09 07:13:19 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-6bdf95c49c to 1 deployment-controller logger.go:42: 07:14:05 | sidecar-namespace | 2024-12-09 07:13:20 +0000 UTC Normal Pod vertx-create-span-sidecar-6bdf95c49c-hwrlp AddedInterface Add eth0 [10.131.0.17/23] from ovn-kubernetes multus logger.go:42: 07:14:05 | sidecar-namespace | 2024-12-09 07:13:20 +0000 UTC Normal Pod vertx-create-span-sidecar-6bdf95c49c-hwrlp.spec.containers{vertx-create-span-sidecar} Pulling Pulling image "jaegertracing/vertx-create-span:operator-e2e-tests" kubelet logger.go:42: 07:14:05 | sidecar-namespace | 2024-12-09 07:13:23 +0000 UTC Normal Pod vertx-create-span-sidecar-6bdf95c49c-hwrlp.spec.containers{vertx-create-span-sidecar} Pulled Successfully pulled image "jaegertracing/vertx-create-span:operator-e2e-tests" in 2.859s (2.859s including waiting). Image size: 282912835 bytes. kubelet logger.go:42: 07:14:05 | sidecar-namespace | 2024-12-09 07:13:23 +0000 UTC Normal Pod vertx-create-span-sidecar-6bdf95c49c-hwrlp.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:14:05 | sidecar-namespace | 2024-12-09 07:13:23 +0000 UTC Normal Pod vertx-create-span-sidecar-6bdf95c49c-hwrlp.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:14:05 | sidecar-namespace | 2024-12-09 07:13:23 +0000 UTC Normal Pod vertx-create-span-sidecar-6bdf95c49c-hwrlp.spec.containers{jaeger-agent} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" kubelet logger.go:42: 07:14:05 | sidecar-namespace | 2024-12-09 07:13:27 +0000 UTC Warning Pod vertx-create-span-sidecar-797645c8fc-8tfpw.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.129.2.23:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:14:05 | sidecar-namespace | 2024-12-09 07:13:27 +0000 UTC Warning Pod vertx-create-span-sidecar-797645c8fc-8tfpw.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.23:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:14:05 | sidecar-namespace | 2024-12-09 07:13:28 +0000 UTC Normal Pod vertx-create-span-sidecar-6bdf95c49c-hwrlp.spec.containers{jaeger-agent} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" in 5.036s (5.036s including waiting). Image size: 115697219 bytes. kubelet logger.go:42: 07:14:05 | sidecar-namespace | 2024-12-09 07:13:28 +0000 UTC Normal Pod vertx-create-span-sidecar-6bdf95c49c-hwrlp.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:14:05 | sidecar-namespace | 2024-12-09 07:13:28 +0000 UTC Normal Pod vertx-create-span-sidecar-6bdf95c49c-hwrlp.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:14:05 | sidecar-namespace | 2024-12-09 07:13:29 +0000 UTC Normal Pod 00-find-service-22rpd Binding Scheduled Successfully assigned kuttl-test-viable-quagga/00-find-service-22rpd to ip-10-0-79-227.ec2.internal default-scheduler logger.go:42: 07:14:05 | sidecar-namespace | 2024-12-09 07:13:29 +0000 UTC Normal Pod 00-find-service-22rpd AddedInterface Add eth0 [10.129.2.24/23] from ovn-kubernetes multus logger.go:42: 07:14:05 | sidecar-namespace | 2024-12-09 07:13:29 +0000 UTC Normal Pod 00-find-service-22rpd.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:14:05 | sidecar-namespace | 2024-12-09 07:13:29 +0000 UTC Normal Job.batch 00-find-service SuccessfulCreate Created pod: 00-find-service-22rpd job-controller logger.go:42: 07:14:05 | sidecar-namespace | 2024-12-09 07:13:29 +0000 UTC Normal Pod vertx-create-span-sidecar-797645c8fc-8tfpw.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 07:14:05 | sidecar-namespace | 2024-12-09 07:13:30 +0000 UTC Warning Pod vertx-create-span-sidecar-797645c8fc-8tfpw.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.23:8080/": read tcp 10.129.2.2:45720->10.129.2.23:8080: read: connection reset by peer kubelet logger.go:42: 07:14:05 | sidecar-namespace | 2024-12-09 07:13:30 +0000 UTC Warning Pod vertx-create-span-sidecar-797645c8fc-8tfpw.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.23:8080/": dial tcp 10.129.2.23:8080: connect: connection refused kubelet logger.go:42: 07:14:05 | sidecar-namespace | 2024-12-09 07:13:30 +0000 UTC Normal Pod vertx-create-span-sidecar-797645c8fc-8tfpw.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:14:05 | sidecar-namespace | 2024-12-09 07:13:31 +0000 UTC Warning Pod vertx-create-span-sidecar-6bdf95c49c-hwrlp.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.131.0.17:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:14:05 | sidecar-namespace | 2024-12-09 07:13:31 +0000 UTC Warning Pod vertx-create-span-sidecar-6bdf95c49c-hwrlp.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.131.0.17:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:14:05 | sidecar-namespace | 2024-12-09 07:13:32 +0000 UTC Normal Pod 00-find-service-22rpd.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 2.698s (2.698s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:14:05 | sidecar-namespace | 2024-12-09 07:13:32 +0000 UTC Normal Pod 00-find-service-22rpd.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:14:05 | sidecar-namespace | 2024-12-09 07:13:32 +0000 UTC Normal Pod 00-find-service-22rpd.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:14:05 | sidecar-namespace | 2024-12-09 07:13:33 +0000 UTC Normal Pod vertx-create-span-sidecar-6bdf95c49c-hwrlp.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 07:14:05 | sidecar-namespace | 2024-12-09 07:13:33 +0000 UTC Warning Pod vertx-create-span-sidecar-6bdf95c49c-hwrlp.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.131.0.17:8080/": read tcp 10.131.0.2:54748->10.131.0.17:8080: read: connection reset by peer kubelet logger.go:42: 07:14:05 | sidecar-namespace | 2024-12-09 07:13:33 +0000 UTC Warning Pod vertx-create-span-sidecar-6bdf95c49c-hwrlp.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.131.0.17:8080/": dial tcp 10.131.0.17:8080: connect: connection refused kubelet logger.go:42: 07:14:05 | sidecar-namespace | 2024-12-09 07:13:33 +0000 UTC Normal Pod vertx-create-span-sidecar-6bdf95c49c-hwrlp.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:14:05 | sidecar-namespace | 2024-12-09 07:13:35 +0000 UTC Normal Job.batch 00-find-service Completed Job completed job-controller logger.go:42: 07:14:05 | sidecar-namespace | 2024-12-09 07:13:41 +0000 UTC Warning Pod vertx-create-span-sidecar-797645c8fc-8tfpw.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.23:8080/": read tcp 10.129.2.2:52906->10.129.2.23:8080: read: connection reset by peer kubelet logger.go:42: 07:14:05 | sidecar-namespace | 2024-12-09 07:13:42 +0000 UTC Normal Pod agent-as-sidecar2-85549847fb-c6k78 Binding Scheduled Successfully assigned kuttl-test-viable-quagga/agent-as-sidecar2-85549847fb-c6k78 to ip-10-0-39-223.ec2.internal default-scheduler logger.go:42: 07:14:05 | sidecar-namespace | 2024-12-09 07:13:42 +0000 UTC Normal ReplicaSet.apps agent-as-sidecar2-85549847fb SuccessfulCreate Created pod: agent-as-sidecar2-85549847fb-c6k78 replicaset-controller logger.go:42: 07:14:05 | sidecar-namespace | 2024-12-09 07:13:42 +0000 UTC Normal Deployment.apps agent-as-sidecar2 ScalingReplicaSet Scaled up replica set agent-as-sidecar2-85549847fb to 1 deployment-controller logger.go:42: 07:14:05 | sidecar-namespace | 2024-12-09 07:13:43 +0000 UTC Normal Pod agent-as-sidecar2-85549847fb-c6k78 AddedInterface Add eth0 [10.128.2.20/23] from ovn-kubernetes multus logger.go:42: 07:14:05 | sidecar-namespace | 2024-12-09 07:13:43 +0000 UTC Normal Pod agent-as-sidecar2-85549847fb-c6k78.spec.containers{jaeger} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1fa5ed13cc8a023f1e987f6cafe86adc195c373cc2b774539df6d0fd02b780a7" kubelet logger.go:42: 07:14:05 | sidecar-namespace | 2024-12-09 07:13:47 +0000 UTC Normal Pod agent-as-sidecar2-85549847fb-c6k78.spec.containers{jaeger} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1fa5ed13cc8a023f1e987f6cafe86adc195c373cc2b774539df6d0fd02b780a7" in 3.439s (3.439s including waiting). Image size: 147411949 bytes. kubelet logger.go:42: 07:14:05 | sidecar-namespace | 2024-12-09 07:13:47 +0000 UTC Normal Pod agent-as-sidecar2-85549847fb-c6k78.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:14:05 | sidecar-namespace | 2024-12-09 07:13:47 +0000 UTC Normal Pod agent-as-sidecar2-85549847fb-c6k78.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:14:05 | sidecar-namespace | 2024-12-09 07:13:49 +0000 UTC Normal Pod 01-find-service-ddsw7 Binding Scheduled Successfully assigned kuttl-test-viable-quagga/01-find-service-ddsw7 to ip-10-0-79-227.ec2.internal default-scheduler logger.go:42: 07:14:05 | sidecar-namespace | 2024-12-09 07:13:49 +0000 UTC Normal Job.batch 01-find-service SuccessfulCreate Created pod: 01-find-service-ddsw7 job-controller logger.go:42: 07:14:05 | sidecar-namespace | 2024-12-09 07:13:49 +0000 UTC Normal Pod agent-as-sidecar-54bb7ccbc4-5s4g8.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:14:05 | sidecar-namespace | 2024-12-09 07:13:50 +0000 UTC Normal Pod 01-find-service-ddsw7 AddedInterface Add eth0 [10.129.2.25/23] from ovn-kubernetes multus logger.go:42: 07:14:05 | sidecar-namespace | 2024-12-09 07:13:50 +0000 UTC Normal Pod 01-find-service-ddsw7.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:14:05 | sidecar-namespace | 2024-12-09 07:13:51 +0000 UTC Normal Pod 01-find-service-ddsw7.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 1.052s (1.052s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:14:05 | sidecar-namespace | 2024-12-09 07:13:51 +0000 UTC Normal Pod 01-find-service-ddsw7.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:14:05 | sidecar-namespace | 2024-12-09 07:13:51 +0000 UTC Normal Pod 01-find-service-ddsw7.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:14:05 | sidecar-namespace | 2024-12-09 07:13:51 +0000 UTC Normal Pod vertx-create-span-sidecar-6c59dc9cf5-dgbdn Binding Scheduled Successfully assigned kuttl-test-viable-quagga/vertx-create-span-sidecar-6c59dc9cf5-dgbdn to ip-10-0-79-227.ec2.internal default-scheduler logger.go:42: 07:14:05 | sidecar-namespace | 2024-12-09 07:13:51 +0000 UTC Normal Pod vertx-create-span-sidecar-6c59dc9cf5-dgbdn AddedInterface Add eth0 [10.129.2.26/23] from ovn-kubernetes multus logger.go:42: 07:14:05 | sidecar-namespace | 2024-12-09 07:13:51 +0000 UTC Normal Pod vertx-create-span-sidecar-6c59dc9cf5-dgbdn.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:14:05 | sidecar-namespace | 2024-12-09 07:13:51 +0000 UTC Normal Pod vertx-create-span-sidecar-6c59dc9cf5-dgbdn.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:14:05 | sidecar-namespace | 2024-12-09 07:13:51 +0000 UTC Normal Pod vertx-create-span-sidecar-6c59dc9cf5-dgbdn.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:14:05 | sidecar-namespace | 2024-12-09 07:13:51 +0000 UTC Normal Pod vertx-create-span-sidecar-6c59dc9cf5-dgbdn.spec.containers{jaeger-agent} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" kubelet logger.go:42: 07:14:05 | sidecar-namespace | 2024-12-09 07:13:51 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-6c59dc9cf5 SuccessfulCreate Created pod: vertx-create-span-sidecar-6c59dc9cf5-dgbdn replicaset-controller logger.go:42: 07:14:05 | sidecar-namespace | 2024-12-09 07:13:51 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-797645c8fc SuccessfulDelete Deleted pod: vertx-create-span-sidecar-797645c8fc-8tfpw replicaset-controller logger.go:42: 07:14:05 | sidecar-namespace | 2024-12-09 07:13:51 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled down replica set vertx-create-span-sidecar-797645c8fc to 0 from 1 deployment-controller logger.go:42: 07:14:05 | sidecar-namespace | 2024-12-09 07:13:51 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-6c59dc9cf5 to 1 from 0 deployment-controller logger.go:42: 07:14:05 | sidecar-namespace | 2024-12-09 07:13:53 +0000 UTC Normal Pod vertx-create-span-sidecar-6c59dc9cf5-dgbdn.spec.containers{jaeger-agent} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" in 1.885s (1.885s including waiting). Image size: 115697219 bytes. kubelet logger.go:42: 07:14:05 | sidecar-namespace | 2024-12-09 07:13:53 +0000 UTC Normal Pod vertx-create-span-sidecar-6c59dc9cf5-dgbdn.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:14:05 | sidecar-namespace | 2024-12-09 07:13:53 +0000 UTC Normal Pod vertx-create-span-sidecar-6c59dc9cf5-dgbdn.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:14:05 | sidecar-namespace | 2024-12-09 07:13:59 +0000 UTC Warning Pod vertx-create-span-sidecar-6c59dc9cf5-dgbdn.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.129.2.26:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:14:05 | sidecar-namespace | 2024-12-09 07:13:59 +0000 UTC Warning Pod vertx-create-span-sidecar-6c59dc9cf5-dgbdn.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.26:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:14:05 | sidecar-namespace | 2024-12-09 07:14:01 +0000 UTC Normal Pod vertx-create-span-sidecar-6c59dc9cf5-dgbdn.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 07:14:05 | sidecar-namespace | 2024-12-09 07:14:02 +0000 UTC Normal Job.batch 01-find-service Completed Job completed job-controller logger.go:42: 07:14:05 | sidecar-namespace | 2024-12-09 07:14:02 +0000 UTC Warning Pod vertx-create-span-sidecar-6c59dc9cf5-dgbdn.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.26:8080/": read tcp 10.129.2.2:53694->10.129.2.26:8080: read: connection reset by peer kubelet logger.go:42: 07:14:05 | sidecar-namespace | 2024-12-09 07:14:02 +0000 UTC Warning Pod vertx-create-span-sidecar-6c59dc9cf5-dgbdn.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.26:8080/": dial tcp 10.129.2.26:8080: connect: connection refused kubelet logger.go:42: 07:14:05 | sidecar-namespace | 2024-12-09 07:14:03 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-6bdf95c49c SuccessfulDelete Deleted pod: vertx-create-span-sidecar-6bdf95c49c-hwrlp replicaset-controller logger.go:42: 07:14:05 | sidecar-namespace | 2024-12-09 07:14:03 +0000 UTC Normal Pod vertx-create-span-sidecar-7d4fcc9fb-dpl27 Binding Scheduled Successfully assigned kuttl-test-viable-quagga/vertx-create-span-sidecar-7d4fcc9fb-dpl27 to ip-10-0-79-227.ec2.internal default-scheduler logger.go:42: 07:14:05 | sidecar-namespace | 2024-12-09 07:14:03 +0000 UTC Normal Pod vertx-create-span-sidecar-7d4fcc9fb-dpl27 AddedInterface Add eth0 [10.129.2.27/23] from ovn-kubernetes multus logger.go:42: 07:14:05 | sidecar-namespace | 2024-12-09 07:14:03 +0000 UTC Normal Pod vertx-create-span-sidecar-7d4fcc9fb-dpl27.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:14:05 | sidecar-namespace | 2024-12-09 07:14:03 +0000 UTC Normal Pod vertx-create-span-sidecar-7d4fcc9fb-dpl27.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:14:05 | sidecar-namespace | 2024-12-09 07:14:03 +0000 UTC Normal Pod vertx-create-span-sidecar-7d4fcc9fb-dpl27.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:14:05 | sidecar-namespace | 2024-12-09 07:14:03 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-7d4fcc9fb SuccessfulCreate Created pod: vertx-create-span-sidecar-7d4fcc9fb-dpl27 replicaset-controller logger.go:42: 07:14:05 | sidecar-namespace | 2024-12-09 07:14:03 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled down replica set vertx-create-span-sidecar-6bdf95c49c to 0 from 1 deployment-controller logger.go:42: 07:14:05 | sidecar-namespace | 2024-12-09 07:14:03 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-7d4fcc9fb to 1 from 0 deployment-controller logger.go:42: 07:14:05 | sidecar-namespace | Deleting namespace: kuttl-test-viable-quagga === CONT kuttl/harness/sidecar-skip-webhook logger.go:42: 07:14:11 | sidecar-skip-webhook | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:14:11 | sidecar-skip-webhook | Creating namespace: kuttl-test-ruling-buzzard logger.go:42: 07:14:11 | sidecar-skip-webhook/0-install | starting test step 0-install logger.go:42: 07:14:12 | sidecar-skip-webhook/0-install | Jaeger:kuttl-test-ruling-buzzard/agent-as-sidecar created logger.go:42: 07:14:18 | sidecar-skip-webhook/0-install | test step completed 0-install logger.go:42: 07:14:18 | sidecar-skip-webhook/1-install | starting test step 1-install logger.go:42: 07:14:18 | sidecar-skip-webhook/1-install | Deployment:kuttl-test-ruling-buzzard/vertx-create-span-sidecar created logger.go:42: 07:14:20 | sidecar-skip-webhook/1-install | test step completed 1-install logger.go:42: 07:14:20 | sidecar-skip-webhook/2-add-anotation-and-label | starting test step 2-add-anotation-and-label logger.go:42: 07:14:20 | sidecar-skip-webhook/2-add-anotation-and-label | running command: [kubectl label deployment vertx-create-span-sidecar app.kubernetes.io/name=jaeger-operator --namespace kuttl-test-ruling-buzzard] logger.go:42: 07:14:20 | sidecar-skip-webhook/2-add-anotation-and-label | deployment.apps/vertx-create-span-sidecar labeled logger.go:42: 07:14:20 | sidecar-skip-webhook/2-add-anotation-and-label | running command: [kubectl annotate --overwrite deployment vertx-create-span-sidecar sidecar.jaegertracing.io/inject=true --namespace kuttl-test-ruling-buzzard] logger.go:42: 07:14:20 | sidecar-skip-webhook/2-add-anotation-and-label | deployment.apps/vertx-create-span-sidecar annotated logger.go:42: 07:14:20 | sidecar-skip-webhook/2-add-anotation-and-label | test step completed 2-add-anotation-and-label logger.go:42: 07:14:20 | sidecar-skip-webhook/3-remove-label | starting test step 3-remove-label logger.go:42: 07:14:20 | sidecar-skip-webhook/3-remove-label | running command: [kubectl label deployment vertx-create-span-sidecar app.kubernetes.io/name- --namespace kuttl-test-ruling-buzzard] logger.go:42: 07:14:20 | sidecar-skip-webhook/3-remove-label | deployment.apps/vertx-create-span-sidecar unlabeled logger.go:42: 07:14:23 | sidecar-skip-webhook/3-remove-label | test step completed 3-remove-label logger.go:42: 07:14:23 | sidecar-skip-webhook | sidecar-skip-webhook events from ns kuttl-test-ruling-buzzard: logger.go:42: 07:14:23 | sidecar-skip-webhook | 2024-12-09 07:14:16 +0000 UTC Normal Pod agent-as-sidecar-65fff6759b-2qmwn Binding Scheduled Successfully assigned kuttl-test-ruling-buzzard/agent-as-sidecar-65fff6759b-2qmwn to ip-10-0-95-250.ec2.internal default-scheduler logger.go:42: 07:14:23 | sidecar-skip-webhook | 2024-12-09 07:14:16 +0000 UTC Normal Pod agent-as-sidecar-65fff6759b-2qmwn AddedInterface Add eth0 [10.131.0.18/23] from ovn-kubernetes multus logger.go:42: 07:14:23 | sidecar-skip-webhook | 2024-12-09 07:14:16 +0000 UTC Normal Pod agent-as-sidecar-65fff6759b-2qmwn.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1fa5ed13cc8a023f1e987f6cafe86adc195c373cc2b774539df6d0fd02b780a7" already present on machine kubelet logger.go:42: 07:14:23 | sidecar-skip-webhook | 2024-12-09 07:14:16 +0000 UTC Normal Pod agent-as-sidecar-65fff6759b-2qmwn.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:14:23 | sidecar-skip-webhook | 2024-12-09 07:14:16 +0000 UTC Normal Pod agent-as-sidecar-65fff6759b-2qmwn.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:14:23 | sidecar-skip-webhook | 2024-12-09 07:14:16 +0000 UTC Normal ReplicaSet.apps agent-as-sidecar-65fff6759b SuccessfulCreate Created pod: agent-as-sidecar-65fff6759b-2qmwn replicaset-controller logger.go:42: 07:14:23 | sidecar-skip-webhook | 2024-12-09 07:14:16 +0000 UTC Normal Deployment.apps agent-as-sidecar ScalingReplicaSet Scaled up replica set agent-as-sidecar-65fff6759b to 1 deployment-controller logger.go:42: 07:14:23 | sidecar-skip-webhook | 2024-12-09 07:14:18 +0000 UTC Normal Pod vertx-create-span-sidecar-797645c8fc-td4pz Binding Scheduled Successfully assigned kuttl-test-ruling-buzzard/vertx-create-span-sidecar-797645c8fc-td4pz to ip-10-0-79-227.ec2.internal default-scheduler logger.go:42: 07:14:23 | sidecar-skip-webhook | 2024-12-09 07:14:18 +0000 UTC Normal Pod vertx-create-span-sidecar-797645c8fc-td4pz AddedInterface Add eth0 [10.129.2.28/23] from ovn-kubernetes multus logger.go:42: 07:14:23 | sidecar-skip-webhook | 2024-12-09 07:14:18 +0000 UTC Normal Pod vertx-create-span-sidecar-797645c8fc-td4pz.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:14:23 | sidecar-skip-webhook | 2024-12-09 07:14:18 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-797645c8fc SuccessfulCreate Created pod: vertx-create-span-sidecar-797645c8fc-td4pz replicaset-controller logger.go:42: 07:14:23 | sidecar-skip-webhook | 2024-12-09 07:14:18 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-797645c8fc to 1 deployment-controller logger.go:42: 07:14:23 | sidecar-skip-webhook | 2024-12-09 07:14:19 +0000 UTC Normal Pod vertx-create-span-sidecar-797645c8fc-td4pz.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:14:23 | sidecar-skip-webhook | 2024-12-09 07:14:19 +0000 UTC Normal Pod vertx-create-span-sidecar-797645c8fc-td4pz.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:14:23 | sidecar-skip-webhook | 2024-12-09 07:14:20 +0000 UTC Normal Pod vertx-create-span-sidecar-75d85d649f-9jwjk Binding Scheduled Successfully assigned kuttl-test-ruling-buzzard/vertx-create-span-sidecar-75d85d649f-9jwjk to ip-10-0-79-227.ec2.internal default-scheduler logger.go:42: 07:14:23 | sidecar-skip-webhook | 2024-12-09 07:14:20 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-75d85d649f SuccessfulCreate Created pod: vertx-create-span-sidecar-75d85d649f-9jwjk replicaset-controller logger.go:42: 07:14:23 | sidecar-skip-webhook | 2024-12-09 07:14:20 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-75d85d649f to 1 deployment-controller logger.go:42: 07:14:23 | sidecar-skip-webhook | 2024-12-09 07:14:21 +0000 UTC Normal Pod vertx-create-span-sidecar-75d85d649f-9jwjk AddedInterface Add eth0 [10.129.2.29/23] from ovn-kubernetes multus logger.go:42: 07:14:23 | sidecar-skip-webhook | 2024-12-09 07:14:21 +0000 UTC Normal Pod vertx-create-span-sidecar-75d85d649f-9jwjk.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:14:23 | sidecar-skip-webhook | 2024-12-09 07:14:21 +0000 UTC Normal Pod vertx-create-span-sidecar-75d85d649f-9jwjk.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:14:23 | sidecar-skip-webhook | 2024-12-09 07:14:21 +0000 UTC Normal Pod vertx-create-span-sidecar-75d85d649f-9jwjk.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:14:23 | sidecar-skip-webhook | 2024-12-09 07:14:21 +0000 UTC Normal Pod vertx-create-span-sidecar-75d85d649f-9jwjk.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" already present on machine kubelet logger.go:42: 07:14:23 | sidecar-skip-webhook | 2024-12-09 07:14:21 +0000 UTC Normal Pod vertx-create-span-sidecar-75d85d649f-9jwjk.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:14:23 | sidecar-skip-webhook | 2024-12-09 07:14:21 +0000 UTC Normal Pod vertx-create-span-sidecar-75d85d649f-9jwjk.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:14:23 | sidecar-skip-webhook | Deleting namespace: kuttl-test-ruling-buzzard === CONT kuttl/harness/sidecar-deployment logger.go:42: 07:14:29 | sidecar-deployment | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:14:29 | sidecar-deployment | Creating namespace: kuttl-test-lucky-tortoise logger.go:42: 07:14:29 | sidecar-deployment/0-install | starting test step 0-install logger.go:42: 07:14:29 | sidecar-deployment/0-install | Jaeger:kuttl-test-lucky-tortoise/agent-as-sidecar created logger.go:42: 07:14:36 | sidecar-deployment/0-install | test step completed 0-install logger.go:42: 07:14:36 | sidecar-deployment/1-install | starting test step 1-install logger.go:42: 07:14:36 | sidecar-deployment/1-install | Deployment:kuttl-test-lucky-tortoise/vertx-create-span-sidecar created logger.go:42: 07:14:39 | sidecar-deployment/1-install | test step completed 1-install logger.go:42: 07:14:39 | sidecar-deployment/2-enable-injection | starting test step 2-enable-injection logger.go:42: 07:14:39 | sidecar-deployment/2-enable-injection | running command: [kubectl annotate --overwrite deployment vertx-create-span-sidecar sidecar.jaegertracing.io/inject=true --namespace kuttl-test-lucky-tortoise] logger.go:42: 07:14:39 | sidecar-deployment/2-enable-injection | deployment.apps/vertx-create-span-sidecar annotated logger.go:42: 07:14:41 | sidecar-deployment/2-enable-injection | test step completed 2-enable-injection logger.go:42: 07:14:41 | sidecar-deployment/3-find-service | starting test step 3-find-service logger.go:42: 07:14:41 | sidecar-deployment/3-find-service | Job:kuttl-test-lucky-tortoise/00-find-service created logger.go:42: 07:14:55 | sidecar-deployment/3-find-service | test step completed 3-find-service logger.go:42: 07:14:55 | sidecar-deployment/4-other-instance | starting test step 4-other-instance logger.go:42: 07:14:55 | sidecar-deployment/4-other-instance | Jaeger:kuttl-test-lucky-tortoise/agent-as-sidecar2 created logger.go:42: 07:15:02 | sidecar-deployment/4-other-instance | test step completed 4-other-instance logger.go:42: 07:15:02 | sidecar-deployment/5-delete-first-instance | starting test step 5-delete-first-instance logger.go:42: 07:15:03 | sidecar-deployment/5-delete-first-instance | test step completed 5-delete-first-instance logger.go:42: 07:15:03 | sidecar-deployment/6-find-service | starting test step 6-find-service logger.go:42: 07:15:03 | sidecar-deployment/6-find-service | Job:kuttl-test-lucky-tortoise/01-find-service created logger.go:42: 07:15:26 | sidecar-deployment/6-find-service | test step completed 6-find-service logger.go:42: 07:15:26 | sidecar-deployment/7-disable-injection | starting test step 7-disable-injection logger.go:42: 07:15:26 | sidecar-deployment/7-disable-injection | running command: [kubectl annotate --overwrite deployment vertx-create-span-sidecar sidecar.jaegertracing.io/inject=false --namespace kuttl-test-lucky-tortoise] logger.go:42: 07:15:26 | sidecar-deployment/7-disable-injection | deployment.apps/vertx-create-span-sidecar annotated logger.go:42: 07:15:27 | sidecar-deployment/7-disable-injection | test step completed 7-disable-injection logger.go:42: 07:15:27 | sidecar-deployment | sidecar-deployment events from ns kuttl-test-lucky-tortoise: logger.go:42: 07:15:27 | sidecar-deployment | 2024-12-09 07:14:33 +0000 UTC Normal Pod agent-as-sidecar-7bfcd64f74-d24g4 Binding Scheduled Successfully assigned kuttl-test-lucky-tortoise/agent-as-sidecar-7bfcd64f74-d24g4 to ip-10-0-95-250.ec2.internal default-scheduler logger.go:42: 07:15:27 | sidecar-deployment | 2024-12-09 07:14:33 +0000 UTC Normal ReplicaSet.apps agent-as-sidecar-7bfcd64f74 SuccessfulCreate Created pod: agent-as-sidecar-7bfcd64f74-d24g4 replicaset-controller logger.go:42: 07:15:27 | sidecar-deployment | 2024-12-09 07:14:33 +0000 UTC Normal Deployment.apps agent-as-sidecar ScalingReplicaSet Scaled up replica set agent-as-sidecar-7bfcd64f74 to 1 deployment-controller logger.go:42: 07:15:27 | sidecar-deployment | 2024-12-09 07:14:34 +0000 UTC Normal Pod agent-as-sidecar-7bfcd64f74-d24g4 AddedInterface Add eth0 [10.131.0.19/23] from ovn-kubernetes multus logger.go:42: 07:15:27 | sidecar-deployment | 2024-12-09 07:14:34 +0000 UTC Normal Pod agent-as-sidecar-7bfcd64f74-d24g4.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1fa5ed13cc8a023f1e987f6cafe86adc195c373cc2b774539df6d0fd02b780a7" already present on machine kubelet logger.go:42: 07:15:27 | sidecar-deployment | 2024-12-09 07:14:34 +0000 UTC Normal Pod agent-as-sidecar-7bfcd64f74-d24g4.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:15:27 | sidecar-deployment | 2024-12-09 07:14:34 +0000 UTC Normal Pod agent-as-sidecar-7bfcd64f74-d24g4.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:15:27 | sidecar-deployment | 2024-12-09 07:14:36 +0000 UTC Normal Pod vertx-create-span-sidecar-797645c8fc-bfnv7 Binding Scheduled Successfully assigned kuttl-test-lucky-tortoise/vertx-create-span-sidecar-797645c8fc-bfnv7 to ip-10-0-79-227.ec2.internal default-scheduler logger.go:42: 07:15:27 | sidecar-deployment | 2024-12-09 07:14:36 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-797645c8fc SuccessfulCreate Created pod: vertx-create-span-sidecar-797645c8fc-bfnv7 replicaset-controller logger.go:42: 07:15:27 | sidecar-deployment | 2024-12-09 07:14:36 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-797645c8fc to 1 deployment-controller logger.go:42: 07:15:27 | sidecar-deployment | 2024-12-09 07:14:37 +0000 UTC Normal Pod vertx-create-span-sidecar-797645c8fc-bfnv7 AddedInterface Add eth0 [10.129.2.30/23] from ovn-kubernetes multus logger.go:42: 07:15:27 | sidecar-deployment | 2024-12-09 07:14:37 +0000 UTC Normal Pod vertx-create-span-sidecar-797645c8fc-bfnv7.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:15:27 | sidecar-deployment | 2024-12-09 07:14:37 +0000 UTC Normal Pod vertx-create-span-sidecar-797645c8fc-bfnv7.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:15:27 | sidecar-deployment | 2024-12-09 07:14:37 +0000 UTC Normal Pod vertx-create-span-sidecar-797645c8fc-bfnv7.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:15:27 | sidecar-deployment | 2024-12-09 07:14:39 +0000 UTC Normal Pod vertx-create-span-sidecar-57cccc5b89-qnfgv Binding Scheduled Successfully assigned kuttl-test-lucky-tortoise/vertx-create-span-sidecar-57cccc5b89-qnfgv to ip-10-0-79-227.ec2.internal default-scheduler logger.go:42: 07:15:27 | sidecar-deployment | 2024-12-09 07:14:39 +0000 UTC Normal Pod vertx-create-span-sidecar-57cccc5b89-qnfgv AddedInterface Add eth0 [10.129.2.31/23] from ovn-kubernetes multus logger.go:42: 07:15:27 | sidecar-deployment | 2024-12-09 07:14:39 +0000 UTC Normal Pod vertx-create-span-sidecar-57cccc5b89-qnfgv.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:15:27 | sidecar-deployment | 2024-12-09 07:14:39 +0000 UTC Normal Pod vertx-create-span-sidecar-57cccc5b89-qnfgv.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:15:27 | sidecar-deployment | 2024-12-09 07:14:39 +0000 UTC Normal Pod vertx-create-span-sidecar-57cccc5b89-qnfgv.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:15:27 | sidecar-deployment | 2024-12-09 07:14:39 +0000 UTC Normal Pod vertx-create-span-sidecar-57cccc5b89-qnfgv.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" already present on machine kubelet logger.go:42: 07:15:27 | sidecar-deployment | 2024-12-09 07:14:39 +0000 UTC Normal Pod vertx-create-span-sidecar-57cccc5b89-qnfgv.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:15:27 | sidecar-deployment | 2024-12-09 07:14:39 +0000 UTC Normal Pod vertx-create-span-sidecar-57cccc5b89-qnfgv.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:15:27 | sidecar-deployment | 2024-12-09 07:14:39 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-57cccc5b89 SuccessfulCreate Created pod: vertx-create-span-sidecar-57cccc5b89-qnfgv replicaset-controller logger.go:42: 07:15:27 | sidecar-deployment | 2024-12-09 07:14:39 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-57cccc5b89 to 1 deployment-controller logger.go:42: 07:15:27 | sidecar-deployment | 2024-12-09 07:14:41 +0000 UTC Normal Pod 00-find-service-bn6ms Binding Scheduled Successfully assigned kuttl-test-lucky-tortoise/00-find-service-bn6ms to ip-10-0-39-223.ec2.internal default-scheduler logger.go:42: 07:15:27 | sidecar-deployment | 2024-12-09 07:14:41 +0000 UTC Normal Job.batch 00-find-service SuccessfulCreate Created pod: 00-find-service-bn6ms job-controller logger.go:42: 07:15:27 | sidecar-deployment | 2024-12-09 07:14:42 +0000 UTC Normal Pod 00-find-service-bn6ms AddedInterface Add eth0 [10.128.2.21/23] from ovn-kubernetes multus logger.go:42: 07:15:27 | sidecar-deployment | 2024-12-09 07:14:42 +0000 UTC Normal Pod 00-find-service-bn6ms.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:15:27 | sidecar-deployment | 2024-12-09 07:14:44 +0000 UTC Normal Pod 00-find-service-bn6ms.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 1.998s (1.998s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:15:27 | sidecar-deployment | 2024-12-09 07:14:44 +0000 UTC Normal Pod 00-find-service-bn6ms.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:15:27 | sidecar-deployment | 2024-12-09 07:14:44 +0000 UTC Normal Pod 00-find-service-bn6ms.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:15:27 | sidecar-deployment | 2024-12-09 07:14:45 +0000 UTC Warning Pod vertx-create-span-sidecar-797645c8fc-bfnv7.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.129.2.30:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:15:27 | sidecar-deployment | 2024-12-09 07:14:45 +0000 UTC Warning Pod vertx-create-span-sidecar-797645c8fc-bfnv7.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.30:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:15:27 | sidecar-deployment | 2024-12-09 07:14:47 +0000 UTC Warning Pod vertx-create-span-sidecar-57cccc5b89-qnfgv.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.129.2.31:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:15:27 | sidecar-deployment | 2024-12-09 07:14:47 +0000 UTC Warning Pod vertx-create-span-sidecar-57cccc5b89-qnfgv.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.31:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:15:27 | sidecar-deployment | 2024-12-09 07:14:47 +0000 UTC Normal Pod vertx-create-span-sidecar-797645c8fc-bfnv7.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 07:15:27 | sidecar-deployment | 2024-12-09 07:14:47 +0000 UTC Warning Pod vertx-create-span-sidecar-797645c8fc-bfnv7.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.30:8080/": read tcp 10.129.2.2:33372->10.129.2.30:8080: read: connection reset by peer kubelet logger.go:42: 07:15:27 | sidecar-deployment | 2024-12-09 07:14:47 +0000 UTC Warning Pod vertx-create-span-sidecar-797645c8fc-bfnv7.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.30:8080/": dial tcp 10.129.2.30:8080: connect: connection refused kubelet logger.go:42: 07:15:27 | sidecar-deployment | 2024-12-09 07:14:49 +0000 UTC Normal Pod vertx-create-span-sidecar-57cccc5b89-qnfgv.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 07:15:27 | sidecar-deployment | 2024-12-09 07:14:50 +0000 UTC Warning Pod vertx-create-span-sidecar-57cccc5b89-qnfgv.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.31:8080/": read tcp 10.129.2.2:54096->10.129.2.31:8080: read: connection reset by peer kubelet logger.go:42: 07:15:27 | sidecar-deployment | 2024-12-09 07:14:50 +0000 UTC Warning Pod vertx-create-span-sidecar-57cccc5b89-qnfgv.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.31:8080/": dial tcp 10.129.2.31:8080: connect: connection refused kubelet logger.go:42: 07:15:27 | sidecar-deployment | 2024-12-09 07:14:54 +0000 UTC Normal Job.batch 00-find-service Completed Job completed job-controller logger.go:42: 07:15:27 | sidecar-deployment | 2024-12-09 07:14:58 +0000 UTC Warning Pod vertx-create-span-sidecar-797645c8fc-bfnv7.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.129.2.30:8080/": read tcp 10.129.2.2:51716->10.129.2.30:8080: read: connection reset by peer kubelet logger.go:42: 07:15:27 | sidecar-deployment | 2024-12-09 07:14:59 +0000 UTC Normal Pod agent-as-sidecar2-7b7d85464d-rxc2s Binding Scheduled Successfully assigned kuttl-test-lucky-tortoise/agent-as-sidecar2-7b7d85464d-rxc2s to ip-10-0-39-223.ec2.internal default-scheduler logger.go:42: 07:15:27 | sidecar-deployment | 2024-12-09 07:14:59 +0000 UTC Normal ReplicaSet.apps agent-as-sidecar2-7b7d85464d SuccessfulCreate Created pod: agent-as-sidecar2-7b7d85464d-rxc2s replicaset-controller logger.go:42: 07:15:27 | sidecar-deployment | 2024-12-09 07:14:59 +0000 UTC Normal Deployment.apps agent-as-sidecar2 ScalingReplicaSet Scaled up replica set agent-as-sidecar2-7b7d85464d to 1 deployment-controller logger.go:42: 07:15:27 | sidecar-deployment | 2024-12-09 07:15:00 +0000 UTC Normal Pod agent-as-sidecar2-7b7d85464d-rxc2s AddedInterface Add eth0 [10.128.2.22/23] from ovn-kubernetes multus logger.go:42: 07:15:27 | sidecar-deployment | 2024-12-09 07:15:00 +0000 UTC Normal Pod agent-as-sidecar2-7b7d85464d-rxc2s.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1fa5ed13cc8a023f1e987f6cafe86adc195c373cc2b774539df6d0fd02b780a7" already present on machine kubelet logger.go:42: 07:15:27 | sidecar-deployment | 2024-12-09 07:15:00 +0000 UTC Normal Pod agent-as-sidecar2-7b7d85464d-rxc2s.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:15:27 | sidecar-deployment | 2024-12-09 07:15:00 +0000 UTC Normal Pod agent-as-sidecar2-7b7d85464d-rxc2s.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:15:27 | sidecar-deployment | 2024-12-09 07:15:03 +0000 UTC Normal Pod 01-find-service-svfpv Binding Scheduled Successfully assigned kuttl-test-lucky-tortoise/01-find-service-svfpv to ip-10-0-95-250.ec2.internal default-scheduler logger.go:42: 07:15:27 | sidecar-deployment | 2024-12-09 07:15:03 +0000 UTC Normal Pod 01-find-service-svfpv AddedInterface Add eth0 [10.131.0.20/23] from ovn-kubernetes multus logger.go:42: 07:15:27 | sidecar-deployment | 2024-12-09 07:15:03 +0000 UTC Normal Pod 01-find-service-svfpv.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:15:27 | sidecar-deployment | 2024-12-09 07:15:03 +0000 UTC Normal Job.batch 01-find-service SuccessfulCreate Created pod: 01-find-service-svfpv job-controller logger.go:42: 07:15:27 | sidecar-deployment | 2024-12-09 07:15:03 +0000 UTC Normal Pod agent-as-sidecar-7bfcd64f74-d24g4.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:15:27 | sidecar-deployment | 2024-12-09 07:15:05 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled down replica set vertx-create-span-sidecar-797645c8fc to 0 from 1 deployment-controller logger.go:42: 07:15:27 | sidecar-deployment | 2024-12-09 07:15:06 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-797645c8fc SuccessfulDelete Deleted pod: vertx-create-span-sidecar-797645c8fc-bfnv7 replicaset-controller logger.go:42: 07:15:27 | sidecar-deployment | 2024-12-09 07:15:06 +0000 UTC Normal Pod vertx-create-span-sidecar-7cf5c59df9-cf6pp Binding Scheduled Successfully assigned kuttl-test-lucky-tortoise/vertx-create-span-sidecar-7cf5c59df9-cf6pp to ip-10-0-95-250.ec2.internal default-scheduler logger.go:42: 07:15:27 | sidecar-deployment | 2024-12-09 07:15:06 +0000 UTC Normal Pod vertx-create-span-sidecar-7cf5c59df9-cf6pp AddedInterface Add eth0 [10.131.0.21/23] from ovn-kubernetes multus logger.go:42: 07:15:27 | sidecar-deployment | 2024-12-09 07:15:06 +0000 UTC Normal Pod vertx-create-span-sidecar-7cf5c59df9-cf6pp.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:15:27 | sidecar-deployment | 2024-12-09 07:15:06 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-7cf5c59df9 SuccessfulCreate Created pod: vertx-create-span-sidecar-7cf5c59df9-cf6pp replicaset-controller logger.go:42: 07:15:27 | sidecar-deployment | 2024-12-09 07:15:06 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-7cf5c59df9 to 1 from 0 deployment-controller logger.go:42: 07:15:27 | sidecar-deployment | 2024-12-09 07:15:07 +0000 UTC Normal Pod 01-find-service-svfpv.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 3.193s (3.193s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:15:27 | sidecar-deployment | 2024-12-09 07:15:07 +0000 UTC Normal Pod 01-find-service-svfpv.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:15:27 | sidecar-deployment | 2024-12-09 07:15:07 +0000 UTC Normal Pod 01-find-service-svfpv.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:15:27 | sidecar-deployment | 2024-12-09 07:15:07 +0000 UTC Normal Pod vertx-create-span-sidecar-7cf5c59df9-cf6pp.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:15:27 | sidecar-deployment | 2024-12-09 07:15:07 +0000 UTC Normal Pod vertx-create-span-sidecar-7cf5c59df9-cf6pp.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:15:27 | sidecar-deployment | 2024-12-09 07:15:07 +0000 UTC Normal Pod vertx-create-span-sidecar-7cf5c59df9-cf6pp.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" already present on machine kubelet logger.go:42: 07:15:27 | sidecar-deployment | 2024-12-09 07:15:07 +0000 UTC Normal Pod vertx-create-span-sidecar-7cf5c59df9-cf6pp.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:15:27 | sidecar-deployment | 2024-12-09 07:15:07 +0000 UTC Normal Pod vertx-create-span-sidecar-7cf5c59df9-cf6pp.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:15:27 | sidecar-deployment | 2024-12-09 07:15:15 +0000 UTC Warning Pod vertx-create-span-sidecar-7cf5c59df9-cf6pp.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.131.0.21:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:15:27 | sidecar-deployment | 2024-12-09 07:15:15 +0000 UTC Warning Pod vertx-create-span-sidecar-7cf5c59df9-cf6pp.spec.containers{vertx-create-span-sidecar} Unhealthy Liveness probe failed: Get "http://10.131.0.21:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:15:27 | sidecar-deployment | 2024-12-09 07:15:17 +0000 UTC Normal Pod vertx-create-span-sidecar-7cf5c59df9-cf6pp.spec.containers{vertx-create-span-sidecar} Killing Container vertx-create-span-sidecar failed liveness probe, will be restarted kubelet logger.go:42: 07:15:27 | sidecar-deployment | 2024-12-09 07:15:17 +0000 UTC Warning Pod vertx-create-span-sidecar-7cf5c59df9-cf6pp.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.131.0.21:8080/": read tcp 10.131.0.2:40914->10.131.0.21:8080: read: connection reset by peer kubelet logger.go:42: 07:15:27 | sidecar-deployment | 2024-12-09 07:15:17 +0000 UTC Warning Pod vertx-create-span-sidecar-7cf5c59df9-cf6pp.spec.containers{vertx-create-span-sidecar} Unhealthy Readiness probe failed: Get "http://10.131.0.21:8080/": dial tcp 10.131.0.21:8080: connect: connection refused kubelet logger.go:42: 07:15:27 | sidecar-deployment | 2024-12-09 07:15:26 +0000 UTC Normal Job.batch 01-find-service Completed Job completed job-controller logger.go:42: 07:15:27 | sidecar-deployment | 2024-12-09 07:15:26 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-57cccc5b89 SuccessfulDelete Deleted pod: vertx-create-span-sidecar-57cccc5b89-qnfgv replicaset-controller logger.go:42: 07:15:27 | sidecar-deployment | 2024-12-09 07:15:26 +0000 UTC Normal Pod vertx-create-span-sidecar-c786c97dd-q7lxd Binding Scheduled Successfully assigned kuttl-test-lucky-tortoise/vertx-create-span-sidecar-c786c97dd-q7lxd to ip-10-0-79-227.ec2.internal default-scheduler logger.go:42: 07:15:27 | sidecar-deployment | 2024-12-09 07:15:26 +0000 UTC Normal ReplicaSet.apps vertx-create-span-sidecar-c786c97dd SuccessfulCreate Created pod: vertx-create-span-sidecar-c786c97dd-q7lxd replicaset-controller logger.go:42: 07:15:27 | sidecar-deployment | 2024-12-09 07:15:26 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled down replica set vertx-create-span-sidecar-57cccc5b89 to 0 from 1 deployment-controller logger.go:42: 07:15:27 | sidecar-deployment | 2024-12-09 07:15:26 +0000 UTC Normal Deployment.apps vertx-create-span-sidecar ScalingReplicaSet Scaled up replica set vertx-create-span-sidecar-c786c97dd to 1 from 0 deployment-controller logger.go:42: 07:15:27 | sidecar-deployment | 2024-12-09 07:15:27 +0000 UTC Normal Pod vertx-create-span-sidecar-c786c97dd-q7lxd AddedInterface Add eth0 [10.129.2.33/23] from ovn-kubernetes multus logger.go:42: 07:15:27 | sidecar-deployment | 2024-12-09 07:15:27 +0000 UTC Normal Pod vertx-create-span-sidecar-c786c97dd-q7lxd.spec.containers{vertx-create-span-sidecar} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:15:27 | sidecar-deployment | 2024-12-09 07:15:27 +0000 UTC Normal Pod vertx-create-span-sidecar-c786c97dd-q7lxd.spec.containers{vertx-create-span-sidecar} Created Created container vertx-create-span-sidecar kubelet logger.go:42: 07:15:27 | sidecar-deployment | 2024-12-09 07:15:27 +0000 UTC Normal Pod vertx-create-span-sidecar-c786c97dd-q7lxd.spec.containers{vertx-create-span-sidecar} Started Started container vertx-create-span-sidecar kubelet logger.go:42: 07:15:28 | sidecar-deployment | Deleting namespace: kuttl-test-lucky-tortoise === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (157.49s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (6.33s) --- PASS: kuttl/harness/sidecar-namespace (68.11s) --- PASS: kuttl/harness/sidecar-skip-webhook (17.70s) --- PASS: kuttl/harness/sidecar-deployment (65.17s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name sidecar --report --output /logs/artifacts/sidecar.xml ./artifacts/kuttl-report.xml time="2024-12-09T07:15:35Z" level=debug msg="Setting a new name for the test suites" time="2024-12-09T07:15:35Z" level=debug msg="Removing 'artifacts' TestCase" time="2024-12-09T07:15:35Z" level=debug msg="normalizing test case names" time="2024-12-09T07:15:35Z" level=debug msg="sidecar/artifacts -> sidecar_artifacts" time="2024-12-09T07:15:35Z" level=debug msg="sidecar/sidecar-namespace -> sidecar_sidecar_namespace" time="2024-12-09T07:15:35Z" level=debug msg="sidecar/sidecar-skip-webhook -> sidecar_sidecar_skip_webhook" time="2024-12-09T07:15:35Z" level=debug msg="sidecar/sidecar-deployment -> sidecar_sidecar_deployment" +------------------------------+--------+ | NAME | RESULT | +------------------------------+--------+ | sidecar_artifacts | passed | | sidecar_sidecar_namespace | passed | | sidecar_sidecar_skip_webhook | passed | | sidecar_sidecar_deployment | passed | +------------------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh streaming false true + '[' 3 -ne 3 ']' + test_suite_name=streaming + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/streaming.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-streaming make[2]: Entering directory '/tmp/jaeger-tests' >>>> Elasticsearch image not loaded because SKIP_ES_EXTERNAL is true KAFKA_VERSION=3.6.0 \ SKIP_KAFKA=false \ SKIP_ES_EXTERNAL=true \ ./tests/e2e/streaming/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.18.0-0.nightly-2024-12-05-191355 True False 10m Cluster version is 4.18.0-0.nightly-2024-12-05-191355' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.18.0-0.nightly-2024-12-05-191355 True False 10m Cluster version is 4.18.0-0.nightly-2024-12-05-191355' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/streaming/render.sh ++ export SUITE_DIR=./tests/e2e/streaming ++ SUITE_DIR=./tests/e2e/streaming ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/streaming ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + is_secured=false + '[' true = true ']' + is_secured=true + '[' false = true ']' + start_test streaming-simple + '[' 1 -ne 1 ']' + test_name=streaming-simple + echo =========================================================================== =========================================================================== + info 'Rendering files for test streaming-simple' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test streaming-simple\e[0m' Rendering files for test streaming-simple + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/streaming/_build + '[' _build '!=' _build ']' + mkdir -p streaming-simple + cd streaming-simple + render_install_kafka my-cluster 00 + '[' 2 -ne 2 ']' + cluster_name=my-cluster + test_step=00 + CLUSTER_NAME=my-cluster + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/kafka-install.yaml.template -o ./00-install.yaml + render_assert_kafka false my-cluster 00 + '[' 3 -ne 3 ']' + autoprovisioned=false + cluster_name=my-cluster + test_step=00 + '[' false = true ']' + '[' false = true ']' + '[' false = false ']' + replicas=1 + CLUSTER_NAME=my-cluster + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-zookeeper-cluster.yaml.template -o ./00-assert.yaml ++ expr 00 + 1 + CLUSTER_NAME=my-cluster + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-kafka-cluster.yaml.template -o ./01-assert.yaml ++ expr 00 + 2 + CLUSTER_NAME=my-cluster + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-entity-operator.yaml.template -o ./02-assert.yaml + render_install_elasticsearch upstream 03 + '[' 2 -ne 2 ']' + deploy_mode=upstream + test_step=03 + '[' upstream = upstream ']' + '[' true = true ']' + template=/tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template + /tmp/jaeger-tests/bin/yq eval -s '"elasticsearch_" + $index' /tmp/jaeger-tests/tests/elasticsearch.yml + /tmp/jaeger-tests/bin/yq eval -i '.spec.template.spec.serviceAccountName="deploy-elasticsearch"' ./elasticsearch_0.yml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template -o ./03-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/elasticsearch-assert.yaml.template -o ./03-assert.yaml + JAEGER_NAME=simple-streaming + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/streaming-jaeger-assert.yaml.template -o ./04-assert.yaml + render_smoke_test simple-streaming true 05 + '[' 3 -ne 3 ']' + jaeger=simple-streaming + is_secured=true + test_step=05 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simple-streaming-query:443 + JAEGER_QUERY_ENDPOINT=https://simple-streaming-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simple-streaming-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simple-streaming-collector-headless:14268 + export JAEGER_NAME=simple-streaming + JAEGER_NAME=simple-streaming + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./05-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./05-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + '[' false = true ']' + start_test streaming-with-tls + '[' 1 -ne 1 ']' + test_name=streaming-with-tls + echo =========================================================================== =========================================================================== + info 'Rendering files for test streaming-with-tls' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test streaming-with-tls\e[0m' Rendering files for test streaming-with-tls + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/streaming/_build/streaming-simple + '[' streaming-simple '!=' _build ']' + cd .. + mkdir -p streaming-with-tls + cd streaming-with-tls + render_install_kafka my-cluster 00 + '[' 2 -ne 2 ']' + cluster_name=my-cluster + test_step=00 + CLUSTER_NAME=my-cluster + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/kafka-install.yaml.template -o ./00-install.yaml + render_assert_kafka false my-cluster 00 + '[' 3 -ne 3 ']' + autoprovisioned=false + cluster_name=my-cluster + test_step=00 + '[' false = true ']' + '[' false = true ']' + '[' false = false ']' + replicas=1 + CLUSTER_NAME=my-cluster + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-zookeeper-cluster.yaml.template -o ./00-assert.yaml ++ expr 00 + 1 + CLUSTER_NAME=my-cluster + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-kafka-cluster.yaml.template -o ./01-assert.yaml ++ expr 00 + 2 + CLUSTER_NAME=my-cluster + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-entity-operator.yaml.template -o ./02-assert.yaml + render_install_elasticsearch upstream 03 + '[' 2 -ne 2 ']' + deploy_mode=upstream + test_step=03 + '[' upstream = upstream ']' + '[' true = true ']' + template=/tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template + /tmp/jaeger-tests/bin/yq eval -s '"elasticsearch_" + $index' /tmp/jaeger-tests/tests/elasticsearch.yml + /tmp/jaeger-tests/bin/yq eval -i '.spec.template.spec.serviceAccountName="deploy-elasticsearch"' ./elasticsearch_0.yml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template -o ./03-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/elasticsearch-assert.yaml.template -o ./03-assert.yaml + render_smoke_test tls-streaming true 05 + '[' 3 -ne 3 ']' + jaeger=tls-streaming + is_secured=true + test_step=05 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://tls-streaming-query:443 + JAEGER_QUERY_ENDPOINT=https://tls-streaming-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://tls-streaming-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://tls-streaming-collector-headless:14268 + export JAEGER_NAME=tls-streaming + JAEGER_NAME=tls-streaming + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./05-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./05-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + '[' false = true ']' + start_test streaming-with-autoprovisioning-autoscale + '[' 1 -ne 1 ']' + test_name=streaming-with-autoprovisioning-autoscale + echo =========================================================================== =========================================================================== + info 'Rendering files for test streaming-with-autoprovisioning-autoscale' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test streaming-with-autoprovisioning-autoscale\e[0m' Rendering files for test streaming-with-autoprovisioning-autoscale + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/streaming/_build/streaming-with-tls + '[' streaming-with-tls '!=' _build ']' + cd .. + mkdir -p streaming-with-autoprovisioning-autoscale + cd streaming-with-autoprovisioning-autoscale + '[' true = true ']' + rm ./00-install.yaml ./00-assert.yaml + render_install_elasticsearch upstream 01 + '[' 2 -ne 2 ']' + deploy_mode=upstream + test_step=01 + '[' upstream = upstream ']' + '[' true = true ']' + template=/tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template + /tmp/jaeger-tests/bin/yq eval -s '"elasticsearch_" + $index' /tmp/jaeger-tests/tests/elasticsearch.yml + /tmp/jaeger-tests/bin/yq eval -i '.spec.template.spec.serviceAccountName="deploy-elasticsearch"' ./elasticsearch_0.yml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/elasticsearch-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/elasticsearch-assert.yaml.template -o ./01-assert.yaml + jaeger_name=auto-provisioned + /tmp/jaeger-tests/bin/yq e -i '.spec.ingester.resources.requests.memory="20Mi"' ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.ingester.resources.requests.memory="500m"' ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.ingester.autoscale=true ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.ingester.minReplicas=1 ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.ingester.maxReplicas=2 ./02-install.yaml + render_assert_kafka true auto-provisioned 03 + '[' 3 -ne 3 ']' + autoprovisioned=true + cluster_name=auto-provisioned + test_step=03 + '[' true = true ']' + is_kafka_minimal_enabled + namespaces=(observability openshift-operators openshift-distributed-tracing) + for i in "${namespaces[@]}" ++ kubectl get pods -n observability -l name=jaeger-operator -o yaml ++ /tmp/jaeger-tests/bin/yq e '.items[0].spec.containers[0].env[] | select(.name=="KAFKA-PROVISIONING-MINIMAL").value' + enabled= + '[' '' == true ']' + for i in "${namespaces[@]}" ++ kubectl get pods -n openshift-operators -l name=jaeger-operator -o yaml ++ /tmp/jaeger-tests/bin/yq e '.items[0].spec.containers[0].env[] | select(.name=="KAFKA-PROVISIONING-MINIMAL").value' + enabled= + '[' '' == true ']' + for i in "${namespaces[@]}" ++ kubectl get pods -n openshift-distributed-tracing -l name=jaeger-operator -o yaml ++ /tmp/jaeger-tests/bin/yq e '.items[0].spec.containers[0].env[] | select(.name=="KAFKA-PROVISIONING-MINIMAL").value' + enabled=true + '[' true == true ']' + return 0 + replicas=1 + CLUSTER_NAME=auto-provisioned + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-zookeeper-cluster.yaml.template -o ./03-assert.yaml ++ expr 03 + 1 + CLUSTER_NAME=auto-provisioned + REPLICAS=1 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-kafka-cluster.yaml.template -o ./04-assert.yaml ++ expr 03 + 2 + CLUSTER_NAME=auto-provisioned + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-entity-operator.yaml.template -o ./05-assert.yaml + version_lt 1.30 1.23 ++ echo 1.30 1.23 ++ tr ' ' '\n' ++ sort -rV ++ head -n 1 + test 1.30 '!=' 1.30 + rm ./08-assert.yaml + skip_test streaming-with-tls 'This test is flaky in Prow CI' + '[' 2 -ne 2 ']' + test_name=streaming-with-tls + message='This test is flaky in Prow CI' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/streaming/_build/streaming-with-autoprovisioning-autoscale + '[' streaming-with-autoprovisioning-autoscale '!=' _build ']' + cd .. + rm -rf streaming-with-tls + warning 'streaming-with-tls: This test is flaky in Prow CI' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: streaming-with-tls: This test is flaky in Prow CI\e[0m' WAR: streaming-with-tls: This test is flaky in Prow CI + skip_test streaming-simple 'This test is flaky in Prow CI' + '[' 2 -ne 2 ']' + test_name=streaming-simple + message='This test is flaky in Prow CI' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/streaming/_build + '[' _build '!=' _build ']' + rm -rf streaming-simple + warning 'streaming-simple: This test is flaky in Prow CI' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: streaming-simple: This test is flaky in Prow CI\e[0m' WAR: streaming-simple: This test is flaky in Prow CI make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running streaming E2E tests' Running streaming E2E tests + cd tests/e2e/streaming/_build + set +e + KUBECONFIG=/tmp/kubeconfig-591533165 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-54q9kfc5-9210b.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 2 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/streaming-with-autoprovisioning-autoscale === PAUSE kuttl/harness/streaming-with-autoprovisioning-autoscale === CONT kuttl/harness/artifacts logger.go:42: 07:15:48 | artifacts | Creating namespace: kuttl-test-climbing-viper logger.go:42: 07:15:48 | artifacts | artifacts events from ns kuttl-test-climbing-viper: logger.go:42: 07:15:48 | artifacts | Deleting namespace: kuttl-test-climbing-viper === CONT kuttl/harness/streaming-with-autoprovisioning-autoscale logger.go:42: 07:15:54 | streaming-with-autoprovisioning-autoscale | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:15:54 | streaming-with-autoprovisioning-autoscale | Ignoring elasticsearch_0.yml as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:15:54 | streaming-with-autoprovisioning-autoscale | Ignoring elasticsearch_1.yml as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:15:54 | streaming-with-autoprovisioning-autoscale | Creating namespace: kuttl-test-singular-heron logger.go:42: 07:15:54 | streaming-with-autoprovisioning-autoscale/1-install | starting test step 1-install logger.go:42: 07:15:54 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c oc create sa deploy-elasticsearch -n $NAMESPACE 2>&1 | grep -v "already exists" || true] logger.go:42: 07:15:54 | streaming-with-autoprovisioning-autoscale/1-install | serviceaccount/deploy-elasticsearch created logger.go:42: 07:15:54 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c oc adm policy add-scc-to-user privileged -z deploy-elasticsearch -n $NAMESPACE 2>&1 | grep -v "already exists" || true] logger.go:42: 07:15:54 | streaming-with-autoprovisioning-autoscale/1-install | clusterrole.rbac.authorization.k8s.io/system:openshift:scc:privileged added: "deploy-elasticsearch" logger.go:42: 07:15:54 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c sleep 6] logger.go:42: 07:16:00 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c kubectl apply -f elasticsearch_0.yml -n $NAMESPACE] logger.go:42: 07:16:00 | streaming-with-autoprovisioning-autoscale/1-install | statefulset.apps/elasticsearch created logger.go:42: 07:16:00 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c sleep 3] logger.go:42: 07:16:03 | streaming-with-autoprovisioning-autoscale/1-install | running command: [sh -c kubectl apply -f elasticsearch_1.yml -n $NAMESPACE] logger.go:42: 07:16:03 | streaming-with-autoprovisioning-autoscale/1-install | service/elasticsearch created logger.go:42: 07:16:21 | streaming-with-autoprovisioning-autoscale/1-install | test step completed 1-install logger.go:42: 07:16:21 | streaming-with-autoprovisioning-autoscale/2-install | starting test step 2-install logger.go:42: 07:16:21 | streaming-with-autoprovisioning-autoscale/2-install | Jaeger:kuttl-test-singular-heron/auto-provisioned created logger.go:42: 07:16:21 | streaming-with-autoprovisioning-autoscale/2-install | test step completed 2-install logger.go:42: 07:16:21 | streaming-with-autoprovisioning-autoscale/3- | starting test step 3- logger.go:42: 07:17:09 | streaming-with-autoprovisioning-autoscale/3- | test step completed 3- logger.go:42: 07:17:09 | streaming-with-autoprovisioning-autoscale/4- | starting test step 4- logger.go:42: 07:17:39 | streaming-with-autoprovisioning-autoscale/4- | test step completed 4- logger.go:42: 07:17:39 | streaming-with-autoprovisioning-autoscale/5- | starting test step 5- logger.go:42: 07:18:02 | streaming-with-autoprovisioning-autoscale/5- | test step completed 5- logger.go:42: 07:18:02 | streaming-with-autoprovisioning-autoscale/6- | starting test step 6- logger.go:42: 07:18:25 | streaming-with-autoprovisioning-autoscale/6- | test step completed 6- logger.go:42: 07:18:25 | streaming-with-autoprovisioning-autoscale/7- | starting test step 7- logger.go:42: 07:18:26 | streaming-with-autoprovisioning-autoscale/7- | test step completed 7- logger.go:42: 07:18:26 | streaming-with-autoprovisioning-autoscale | streaming-with-autoprovisioning-autoscale events from ns kuttl-test-singular-heron: logger.go:42: 07:18:26 | streaming-with-autoprovisioning-autoscale | 2024-12-09 07:16:00 +0000 UTC Normal Pod elasticsearch-0 Binding Scheduled Successfully assigned kuttl-test-singular-heron/elasticsearch-0 to ip-10-0-79-227.ec2.internal default-scheduler logger.go:42: 07:18:26 | streaming-with-autoprovisioning-autoscale | 2024-12-09 07:16:00 +0000 UTC Normal StatefulSet.apps elasticsearch SuccessfulCreate create Pod elasticsearch-0 in StatefulSet elasticsearch successful statefulset-controller logger.go:42: 07:18:26 | streaming-with-autoprovisioning-autoscale | 2024-12-09 07:16:01 +0000 UTC Normal Pod elasticsearch-0 AddedInterface Add eth0 [10.129.2.34/23] from ovn-kubernetes multus logger.go:42: 07:18:26 | streaming-with-autoprovisioning-autoscale | 2024-12-09 07:16:01 +0000 UTC Normal Pod elasticsearch-0.spec.containers{elasticsearch} Pulling Pulling image "docker.elastic.co/elasticsearch/elasticsearch-oss:6.8.6" kubelet logger.go:42: 07:18:26 | streaming-with-autoprovisioning-autoscale | 2024-12-09 07:16:12 +0000 UTC Normal Pod elasticsearch-0.spec.containers{elasticsearch} Pulled Successfully pulled image "docker.elastic.co/elasticsearch/elasticsearch-oss:6.8.6" in 10.757s (10.757s including waiting). Image size: 758467647 bytes. kubelet logger.go:42: 07:18:26 | streaming-with-autoprovisioning-autoscale | 2024-12-09 07:16:12 +0000 UTC Normal Pod elasticsearch-0.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:18:26 | streaming-with-autoprovisioning-autoscale | 2024-12-09 07:16:12 +0000 UTC Normal Pod elasticsearch-0.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:18:26 | streaming-with-autoprovisioning-autoscale | 2024-12-09 07:16:32 +0000 UTC Normal PodDisruptionBudget.policy auto-provisioned-zookeeper NoPods No matching pods found controllermanager logger.go:42: 07:18:26 | streaming-with-autoprovisioning-autoscale | 2024-12-09 07:16:32 +0000 UTC Normal PersistentVolumeClaim data-auto-provisioned-zookeeper-0 WaitForFirstConsumer waiting for first consumer to be created before binding persistentvolume-controller logger.go:42: 07:18:26 | streaming-with-autoprovisioning-autoscale | 2024-12-09 07:16:33 +0000 UTC Normal PersistentVolumeClaim data-auto-provisioned-zookeeper-0 Provisioning External provisioner is provisioning volume for claim "kuttl-test-singular-heron/data-auto-provisioned-zookeeper-0" ebs.csi.aws.com_aws-ebs-csi-driver-controller-587676f5d9-xlc98_a4aab91d-8dc4-420f-b14b-b02c13ee89f7 logger.go:42: 07:18:26 | streaming-with-autoprovisioning-autoscale | 2024-12-09 07:16:33 +0000 UTC Normal PersistentVolumeClaim data-auto-provisioned-zookeeper-0 ExternalProvisioning Waiting for a volume to be created either by the external provisioner 'ebs.csi.aws.com' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered. persistentvolume-controller logger.go:42: 07:18:26 | streaming-with-autoprovisioning-autoscale | 2024-12-09 07:16:35 +0000 UTC Normal PersistentVolumeClaim data-auto-provisioned-zookeeper-0 ProvisioningSucceeded Successfully provisioned volume pvc-7f2c0b94-2cd6-4638-bd8e-dbe6a03361aa ebs.csi.aws.com_aws-ebs-csi-driver-controller-587676f5d9-xlc98_a4aab91d-8dc4-420f-b14b-b02c13ee89f7 logger.go:42: 07:18:26 | streaming-with-autoprovisioning-autoscale | 2024-12-09 07:16:36 +0000 UTC Normal Pod auto-provisioned-zookeeper-0 Binding Scheduled Successfully assigned kuttl-test-singular-heron/auto-provisioned-zookeeper-0 to ip-10-0-95-250.ec2.internal default-scheduler logger.go:42: 07:18:26 | streaming-with-autoprovisioning-autoscale | 2024-12-09 07:16:38 +0000 UTC Normal Pod auto-provisioned-zookeeper-0 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-7f2c0b94-2cd6-4638-bd8e-dbe6a03361aa" attachdetach-controller logger.go:42: 07:18:26 | streaming-with-autoprovisioning-autoscale | 2024-12-09 07:16:39 +0000 UTC Normal Pod auto-provisioned-zookeeper-0 AddedInterface Add eth0 [10.131.0.22/23] from ovn-kubernetes multus logger.go:42: 07:18:26 | streaming-with-autoprovisioning-autoscale | 2024-12-09 07:16:39 +0000 UTC Normal Pod auto-provisioned-zookeeper-0.spec.containers{zookeeper} Pulling Pulling image "registry.redhat.io/amq-streams/kafka-38-rhel9@sha256:c9a60d03827466ee37dd4aff6803eda2c1d65cb2508cb57d13e1d47f982e5a20" kubelet logger.go:42: 07:18:26 | streaming-with-autoprovisioning-autoscale | 2024-12-09 07:16:46 +0000 UTC Normal Pod auto-provisioned-zookeeper-0.spec.containers{zookeeper} Pulled Successfully pulled image "registry.redhat.io/amq-streams/kafka-38-rhel9@sha256:c9a60d03827466ee37dd4aff6803eda2c1d65cb2508cb57d13e1d47f982e5a20" in 7.005s (7.005s including waiting). Image size: 616797249 bytes. kubelet logger.go:42: 07:18:26 | streaming-with-autoprovisioning-autoscale | 2024-12-09 07:16:46 +0000 UTC Normal Pod auto-provisioned-zookeeper-0.spec.containers{zookeeper} Created Created container zookeeper kubelet logger.go:42: 07:18:26 | streaming-with-autoprovisioning-autoscale | 2024-12-09 07:16:46 +0000 UTC Normal Pod auto-provisioned-zookeeper-0.spec.containers{zookeeper} Started Started container zookeeper kubelet logger.go:42: 07:18:26 | streaming-with-autoprovisioning-autoscale | 2024-12-09 07:17:10 +0000 UTC Normal PodDisruptionBudget.policy auto-provisioned-kafka NoPods No matching pods found controllermanager logger.go:42: 07:18:26 | streaming-with-autoprovisioning-autoscale | 2024-12-09 07:17:10 +0000 UTC Normal PersistentVolumeClaim data-0-auto-provisioned-kafka-0 WaitForFirstConsumer waiting for first consumer to be created before binding persistentvolume-controller logger.go:42: 07:18:26 | streaming-with-autoprovisioning-autoscale | 2024-12-09 07:17:11 +0000 UTC Normal PersistentVolumeClaim data-0-auto-provisioned-kafka-0 Provisioning External provisioner is provisioning volume for claim "kuttl-test-singular-heron/data-0-auto-provisioned-kafka-0" ebs.csi.aws.com_aws-ebs-csi-driver-controller-587676f5d9-xlc98_a4aab91d-8dc4-420f-b14b-b02c13ee89f7 logger.go:42: 07:18:26 | streaming-with-autoprovisioning-autoscale | 2024-12-09 07:17:11 +0000 UTC Normal PersistentVolumeClaim data-0-auto-provisioned-kafka-0 ExternalProvisioning Waiting for a volume to be created either by the external provisioner 'ebs.csi.aws.com' or manually by the system administrator. If volume creation is delayed, please verify that the provisioner is running and correctly registered. persistentvolume-controller logger.go:42: 07:18:26 | streaming-with-autoprovisioning-autoscale | 2024-12-09 07:17:13 +0000 UTC Normal PersistentVolumeClaim data-0-auto-provisioned-kafka-0 ProvisioningSucceeded Successfully provisioned volume pvc-55152b9b-d772-47b3-98f7-cdd97f30c717 ebs.csi.aws.com_aws-ebs-csi-driver-controller-587676f5d9-xlc98_a4aab91d-8dc4-420f-b14b-b02c13ee89f7 logger.go:42: 07:18:26 | streaming-with-autoprovisioning-autoscale | 2024-12-09 07:17:14 +0000 UTC Normal Pod auto-provisioned-kafka-0 Binding Scheduled Successfully assigned kuttl-test-singular-heron/auto-provisioned-kafka-0 to ip-10-0-95-250.ec2.internal default-scheduler logger.go:42: 07:18:26 | streaming-with-autoprovisioning-autoscale | 2024-12-09 07:17:17 +0000 UTC Normal Pod auto-provisioned-kafka-0 SuccessfulAttachVolume AttachVolume.Attach succeeded for volume "pvc-55152b9b-d772-47b3-98f7-cdd97f30c717" attachdetach-controller logger.go:42: 07:18:26 | streaming-with-autoprovisioning-autoscale | 2024-12-09 07:17:19 +0000 UTC Normal Pod auto-provisioned-kafka-0 AddedInterface Add eth0 [10.131.0.23/23] from ovn-kubernetes multus logger.go:42: 07:18:26 | streaming-with-autoprovisioning-autoscale | 2024-12-09 07:17:19 +0000 UTC Normal Pod auto-provisioned-kafka-0.spec.containers{kafka} Pulled Container image "registry.redhat.io/amq-streams/kafka-38-rhel9@sha256:c9a60d03827466ee37dd4aff6803eda2c1d65cb2508cb57d13e1d47f982e5a20" already present on machine kubelet logger.go:42: 07:18:26 | streaming-with-autoprovisioning-autoscale | 2024-12-09 07:17:19 +0000 UTC Normal Pod auto-provisioned-kafka-0.spec.containers{kafka} Created Created container kafka kubelet logger.go:42: 07:18:26 | streaming-with-autoprovisioning-autoscale | 2024-12-09 07:17:19 +0000 UTC Normal Pod auto-provisioned-kafka-0.spec.containers{kafka} Started Started container kafka kubelet logger.go:42: 07:18:26 | streaming-with-autoprovisioning-autoscale | 2024-12-09 07:17:41 +0000 UTC Normal Pod auto-provisioned-entity-operator-64c66f8cc4-lv2tc Binding Scheduled Successfully assigned kuttl-test-singular-heron/auto-provisioned-entity-operator-64c66f8cc4-lv2tc to ip-10-0-95-250.ec2.internal default-scheduler logger.go:42: 07:18:26 | streaming-with-autoprovisioning-autoscale | 2024-12-09 07:17:41 +0000 UTC Normal ReplicaSet.apps auto-provisioned-entity-operator-64c66f8cc4 SuccessfulCreate Created pod: auto-provisioned-entity-operator-64c66f8cc4-lv2tc replicaset-controller logger.go:42: 07:18:26 | streaming-with-autoprovisioning-autoscale | 2024-12-09 07:17:41 +0000 UTC Normal Deployment.apps auto-provisioned-entity-operator ScalingReplicaSet Scaled up replica set auto-provisioned-entity-operator-64c66f8cc4 to 1 deployment-controller logger.go:42: 07:18:26 | streaming-with-autoprovisioning-autoscale | 2024-12-09 07:17:42 +0000 UTC Normal Pod auto-provisioned-entity-operator-64c66f8cc4-lv2tc AddedInterface Add eth0 [10.131.0.24/23] from ovn-kubernetes multus logger.go:42: 07:18:26 | streaming-with-autoprovisioning-autoscale | 2024-12-09 07:17:42 +0000 UTC Normal Pod auto-provisioned-entity-operator-64c66f8cc4-lv2tc.spec.containers{topic-operator} Pulled Container image "registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:07a9540a8b906c6d52e6a9684cfb838aadf9849cc24b1d80218ea3ad5545cb5a" already present on machine kubelet logger.go:42: 07:18:26 | streaming-with-autoprovisioning-autoscale | 2024-12-09 07:17:42 +0000 UTC Normal Pod auto-provisioned-entity-operator-64c66f8cc4-lv2tc.spec.containers{topic-operator} Created Created container topic-operator kubelet logger.go:42: 07:18:26 | streaming-with-autoprovisioning-autoscale | 2024-12-09 07:17:42 +0000 UTC Normal Pod auto-provisioned-entity-operator-64c66f8cc4-lv2tc.spec.containers{topic-operator} Started Started container topic-operator kubelet logger.go:42: 07:18:26 | streaming-with-autoprovisioning-autoscale | 2024-12-09 07:17:42 +0000 UTC Normal Pod auto-provisioned-entity-operator-64c66f8cc4-lv2tc.spec.containers{user-operator} Pulled Container image "registry.redhat.io/amq-streams/strimzi-rhel9-operator@sha256:07a9540a8b906c6d52e6a9684cfb838aadf9849cc24b1d80218ea3ad5545cb5a" already present on machine kubelet logger.go:42: 07:18:26 | streaming-with-autoprovisioning-autoscale | 2024-12-09 07:17:42 +0000 UTC Normal Pod auto-provisioned-entity-operator-64c66f8cc4-lv2tc.spec.containers{user-operator} Created Created container user-operator kubelet logger.go:42: 07:18:26 | streaming-with-autoprovisioning-autoscale | 2024-12-09 07:17:42 +0000 UTC Normal Pod auto-provisioned-entity-operator-64c66f8cc4-lv2tc.spec.containers{user-operator} Started Started container user-operator kubelet logger.go:42: 07:18:26 | streaming-with-autoprovisioning-autoscale | 2024-12-09 07:18:04 +0000 UTC Normal Pod auto-provisioned-collector-5bccbdbcd9-b975f Binding Scheduled Successfully assigned kuttl-test-singular-heron/auto-provisioned-collector-5bccbdbcd9-b975f to ip-10-0-79-227.ec2.internal default-scheduler logger.go:42: 07:18:26 | streaming-with-autoprovisioning-autoscale | 2024-12-09 07:18:04 +0000 UTC Normal ReplicaSet.apps auto-provisioned-collector-5bccbdbcd9 SuccessfulCreate Created pod: auto-provisioned-collector-5bccbdbcd9-b975f replicaset-controller logger.go:42: 07:18:26 | streaming-with-autoprovisioning-autoscale | 2024-12-09 07:18:04 +0000 UTC Normal Deployment.apps auto-provisioned-collector ScalingReplicaSet Scaled up replica set auto-provisioned-collector-5bccbdbcd9 to 1 deployment-controller logger.go:42: 07:18:26 | streaming-with-autoprovisioning-autoscale | 2024-12-09 07:18:04 +0000 UTC Normal Pod auto-provisioned-ingester-7f676c6fb4-kxsdv Binding Scheduled Successfully assigned kuttl-test-singular-heron/auto-provisioned-ingester-7f676c6fb4-kxsdv to ip-10-0-79-227.ec2.internal default-scheduler logger.go:42: 07:18:26 | streaming-with-autoprovisioning-autoscale | 2024-12-09 07:18:04 +0000 UTC Normal ReplicaSet.apps auto-provisioned-ingester-7f676c6fb4 SuccessfulCreate Created pod: auto-provisioned-ingester-7f676c6fb4-kxsdv replicaset-controller logger.go:42: 07:18:26 | streaming-with-autoprovisioning-autoscale | 2024-12-09 07:18:04 +0000 UTC Normal Deployment.apps auto-provisioned-ingester ScalingReplicaSet Scaled up replica set auto-provisioned-ingester-7f676c6fb4 to 1 deployment-controller logger.go:42: 07:18:26 | streaming-with-autoprovisioning-autoscale | 2024-12-09 07:18:04 +0000 UTC Normal Pod auto-provisioned-query-57fc868cd4-n2gfz Binding Scheduled Successfully assigned kuttl-test-singular-heron/auto-provisioned-query-57fc868cd4-n2gfz to ip-10-0-39-223.ec2.internal default-scheduler logger.go:42: 07:18:26 | streaming-with-autoprovisioning-autoscale | 2024-12-09 07:18:04 +0000 UTC Normal ReplicaSet.apps auto-provisioned-query-57fc868cd4 SuccessfulCreate Created pod: auto-provisioned-query-57fc868cd4-n2gfz replicaset-controller logger.go:42: 07:18:26 | streaming-with-autoprovisioning-autoscale | 2024-12-09 07:18:04 +0000 UTC Normal Deployment.apps auto-provisioned-query ScalingReplicaSet Scaled up replica set auto-provisioned-query-57fc868cd4 to 1 deployment-controller logger.go:42: 07:18:26 | streaming-with-autoprovisioning-autoscale | 2024-12-09 07:18:05 +0000 UTC Normal Pod auto-provisioned-collector-5bccbdbcd9-b975f AddedInterface Add eth0 [10.129.2.35/23] from ovn-kubernetes multus logger.go:42: 07:18:26 | streaming-with-autoprovisioning-autoscale | 2024-12-09 07:18:05 +0000 UTC Normal Pod auto-provisioned-collector-5bccbdbcd9-b975f.spec.containers{jaeger-collector} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:44686d560aa27cae8ff8693f88c4cb6e2edf1737010ec1f80709cb42250b729d" kubelet logger.go:42: 07:18:26 | streaming-with-autoprovisioning-autoscale | 2024-12-09 07:18:05 +0000 UTC Normal Pod auto-provisioned-ingester-7f676c6fb4-kxsdv AddedInterface Add eth0 [10.129.2.36/23] from ovn-kubernetes multus logger.go:42: 07:18:26 | streaming-with-autoprovisioning-autoscale | 2024-12-09 07:18:05 +0000 UTC Normal Pod auto-provisioned-ingester-7f676c6fb4-kxsdv.spec.containers{jaeger-ingester} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-ingester-rhel8@sha256:e01a99dddffe65385d77dd6692558aa9a47df36f8b2a8d141b6ad561139981b2" kubelet logger.go:42: 07:18:26 | streaming-with-autoprovisioning-autoscale | 2024-12-09 07:18:05 +0000 UTC Normal Pod auto-provisioned-query-57fc868cd4-n2gfz AddedInterface Add eth0 [10.128.2.23/23] from ovn-kubernetes multus logger.go:42: 07:18:26 | streaming-with-autoprovisioning-autoscale | 2024-12-09 07:18:05 +0000 UTC Normal Pod auto-provisioned-query-57fc868cd4-n2gfz.spec.containers{jaeger-query} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:24a83160aa930e1b72c2a2442a33b28af2b06c24b058e09afcd0d495a8066d6d" kubelet logger.go:42: 07:18:26 | streaming-with-autoprovisioning-autoscale | 2024-12-09 07:18:07 +0000 UTC Normal Pod auto-provisioned-collector-5bccbdbcd9-b975f.spec.containers{jaeger-collector} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:44686d560aa27cae8ff8693f88c4cb6e2edf1737010ec1f80709cb42250b729d" in 2.58s (2.58s including waiting). Image size: 142489237 bytes. kubelet logger.go:42: 07:18:26 | streaming-with-autoprovisioning-autoscale | 2024-12-09 07:18:07 +0000 UTC Normal Pod auto-provisioned-collector-5bccbdbcd9-b975f.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:18:26 | streaming-with-autoprovisioning-autoscale | 2024-12-09 07:18:07 +0000 UTC Normal Pod auto-provisioned-collector-5bccbdbcd9-b975f.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:18:26 | streaming-with-autoprovisioning-autoscale | 2024-12-09 07:18:10 +0000 UTC Normal Pod auto-provisioned-query-57fc868cd4-n2gfz.spec.containers{jaeger-query} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:24a83160aa930e1b72c2a2442a33b28af2b06c24b058e09afcd0d495a8066d6d" in 5.129s (5.129s including waiting). Image size: 192936692 bytes. kubelet logger.go:42: 07:18:26 | streaming-with-autoprovisioning-autoscale | 2024-12-09 07:18:10 +0000 UTC Normal Pod auto-provisioned-query-57fc868cd4-n2gfz.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:18:26 | streaming-with-autoprovisioning-autoscale | 2024-12-09 07:18:10 +0000 UTC Normal Pod auto-provisioned-query-57fc868cd4-n2gfz.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:18:26 | streaming-with-autoprovisioning-autoscale | 2024-12-09 07:18:10 +0000 UTC Normal Pod auto-provisioned-query-57fc868cd4-n2gfz.spec.containers{oauth-proxy} Pulling Pulling image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" kubelet logger.go:42: 07:18:26 | streaming-with-autoprovisioning-autoscale | 2024-12-09 07:18:11 +0000 UTC Normal Pod auto-provisioned-ingester-7f676c6fb4-kxsdv.spec.containers{jaeger-ingester} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-ingester-rhel8@sha256:e01a99dddffe65385d77dd6692558aa9a47df36f8b2a8d141b6ad561139981b2" in 6.521s (6.521s including waiting). Image size: 139765359 bytes. kubelet logger.go:42: 07:18:26 | streaming-with-autoprovisioning-autoscale | 2024-12-09 07:18:11 +0000 UTC Normal Pod auto-provisioned-ingester-7f676c6fb4-kxsdv.spec.containers{jaeger-ingester} Created Created container jaeger-ingester kubelet logger.go:42: 07:18:26 | streaming-with-autoprovisioning-autoscale | 2024-12-09 07:18:11 +0000 UTC Normal Pod auto-provisioned-ingester-7f676c6fb4-kxsdv.spec.containers{jaeger-ingester} Started Started container jaeger-ingester kubelet logger.go:42: 07:18:26 | streaming-with-autoprovisioning-autoscale | 2024-12-09 07:18:12 +0000 UTC Warning Pod auto-provisioned-ingester-7f676c6fb4-kxsdv.spec.containers{jaeger-ingester} Unhealthy Readiness probe failed: HTTP probe failed with statuscode: 503 kubelet logger.go:42: 07:18:26 | streaming-with-autoprovisioning-autoscale | 2024-12-09 07:18:14 +0000 UTC Normal Pod auto-provisioned-query-57fc868cd4-n2gfz.spec.containers{oauth-proxy} Pulled Successfully pulled image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" in 3.858s (3.858s including waiting). Image size: 339954870 bytes. kubelet logger.go:42: 07:18:26 | streaming-with-autoprovisioning-autoscale | 2024-12-09 07:18:14 +0000 UTC Normal Pod auto-provisioned-query-57fc868cd4-n2gfz.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:18:26 | streaming-with-autoprovisioning-autoscale | 2024-12-09 07:18:14 +0000 UTC Normal Pod auto-provisioned-query-57fc868cd4-n2gfz.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:18:26 | streaming-with-autoprovisioning-autoscale | 2024-12-09 07:18:14 +0000 UTC Normal Pod auto-provisioned-query-57fc868cd4-n2gfz.spec.containers{jaeger-agent} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" kubelet logger.go:42: 07:18:26 | streaming-with-autoprovisioning-autoscale | 2024-12-09 07:18:16 +0000 UTC Normal Pod auto-provisioned-query-57fc868cd4-n2gfz.spec.containers{jaeger-agent} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" in 1.801s (1.801s including waiting). Image size: 115697219 bytes. kubelet logger.go:42: 07:18:26 | streaming-with-autoprovisioning-autoscale | 2024-12-09 07:18:16 +0000 UTC Normal Pod auto-provisioned-query-57fc868cd4-n2gfz.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:18:26 | streaming-with-autoprovisioning-autoscale | 2024-12-09 07:18:16 +0000 UTC Normal Pod auto-provisioned-query-57fc868cd4-n2gfz.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:18:26 | streaming-with-autoprovisioning-autoscale | 2024-12-09 07:18:19 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling auto-provisioned-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:18:26 | streaming-with-autoprovisioning-autoscale | 2024-12-09 07:18:19 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling auto-provisioned-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:18:26 | streaming-with-autoprovisioning-autoscale | 2024-12-09 07:18:19 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling auto-provisioned-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:18:26 | streaming-with-autoprovisioning-autoscale | 2024-12-09 07:18:19 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling auto-provisioned-ingester FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:18:26 | streaming-with-autoprovisioning-autoscale | 2024-12-09 07:18:19 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling auto-provisioned-ingester FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:18:26 | streaming-with-autoprovisioning-autoscale | 2024-12-09 07:18:19 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling auto-provisioned-ingester FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:18:26 | streaming-with-autoprovisioning-autoscale | Deleting namespace: kuttl-test-singular-heron === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (176.41s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (5.92s) --- PASS: kuttl/harness/streaming-with-autoprovisioning-autoscale (170.30s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name streaming --report --output /logs/artifacts/streaming.xml ./artifacts/kuttl-report.xml time="2024-12-09T07:18:44Z" level=debug msg="Setting a new name for the test suites" time="2024-12-09T07:18:44Z" level=debug msg="Removing 'artifacts' TestCase" time="2024-12-09T07:18:44Z" level=debug msg="normalizing test case names" time="2024-12-09T07:18:44Z" level=debug msg="streaming/artifacts -> streaming_artifacts" time="2024-12-09T07:18:44Z" level=debug msg="streaming/streaming-with-autoprovisioning-autoscale -> streaming_streaming_with_autoprovisioning_autoscale" +-----------------------------------------------------+--------+ | NAME | RESULT | +-----------------------------------------------------+--------+ | streaming_artifacts | passed | | streaming_streaming_with_autoprovisioning_autoscale | passed | +-----------------------------------------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh ui false true + '[' 3 -ne 3 ']' + test_suite_name=ui + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/ui.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-ui make[2]: Entering directory '/tmp/jaeger-tests' >>>> Elasticsearch image not loaded because SKIP_ES_EXTERNAL is true SKIP_ES_EXTERNAL=true ./tests/e2e/ui/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.18.0-0.nightly-2024-12-05-191355 True False 13m Cluster version is 4.18.0-0.nightly-2024-12-05-191355' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.18.0-0.nightly-2024-12-05-191355 True False 13m Cluster version is 4.18.0-0.nightly-2024-12-05-191355' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/ui/render.sh ++ export SUITE_DIR=./tests/e2e/ui ++ SUITE_DIR=./tests/e2e/ui ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/ui ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + start_test allinone + '[' 1 -ne 1 ']' + test_name=allinone + echo =========================================================================== =========================================================================== + info 'Rendering files for test allinone' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test allinone\e[0m' Rendering files for test allinone + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/ui/_build + '[' _build '!=' _build ']' + mkdir -p allinone + cd allinone + export GET_URL_COMMAND + export URL + export JAEGER_NAME=all-in-one-ui + JAEGER_NAME=all-in-one-ui + '[' true = true ']' + GET_URL_COMMAND='kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE' + URL='https://$(kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE)/search' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/ensure-ingress-host.sh.template -o ./ensure-ingress-host.sh + chmod +x ./ensure-ingress-host.sh + EXPECTED_CODE=200 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./01-curl.yaml + ASSERT_PRESENT=true + TRACKING_ID=MyTrackingId + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/test-ui-config.yaml.template -o ./04-test-ui-config.yaml + start_test production + '[' 1 -ne 1 ']' + test_name=production + echo =========================================================================== =========================================================================== + info 'Rendering files for test production' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test production\e[0m' Rendering files for test production + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/ui/_build/allinone + '[' allinone '!=' _build ']' + cd .. + mkdir -p production + cd production + export JAEGER_NAME=production-ui + JAEGER_NAME=production-ui + [[ true = true ]] + [[ true = true ]] + render_install_jaeger production-ui production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=production-ui + JAEGER_NAME=production-ui + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/ensure-ingress-host.sh.template -o ./ensure-ingress-host.sh + chmod +x ./ensure-ingress-host.sh + '[' true = true ']' + INSECURE=true + EXPECTED_CODE=403 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./02-check-forbbiden-access.yaml + EXPECTED_CODE=200 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./03-curl.yaml + INSECURE=true + EXPECTED_CODE=200 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./05-check-disabled-security.yaml + ASSERT_PRESENT=false + TRACKING_ID=MyTrackingId + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/test-ui-config.yaml.template -o ./06-check-NO-gaID.yaml + ASSERT_PRESENT=true + TRACKING_ID=MyTrackingId + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/test-ui-config.yaml.template -o ./08-check-gaID.yaml make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running ui E2E tests' Running ui E2E tests + cd tests/e2e/ui/_build + set +e + KUBECONFIG=/tmp/kubeconfig-591533165 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-54q9kfc5-9210b.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 3 tests === RUN kuttl/harness === RUN kuttl/harness/allinone === PAUSE kuttl/harness/allinone === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/production === PAUSE kuttl/harness/production === CONT kuttl/harness/allinone logger.go:42: 07:18:51 | allinone | Ignoring ensure-ingress-host.sh as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:18:51 | allinone | Creating namespace: kuttl-test-notable-shad logger.go:42: 07:18:51 | allinone/0-install | starting test step 0-install logger.go:42: 07:18:51 | allinone/0-install | Jaeger:kuttl-test-notable-shad/all-in-one-ui created logger.go:42: 07:18:55 | allinone/0-install | test step completed 0-install logger.go:42: 07:18:55 | allinone/1-curl | starting test step 1-curl logger.go:42: 07:18:55 | allinone/1-curl | running command: [./ensure-ingress-host.sh] logger.go:42: 07:18:55 | allinone/1-curl | Checking the Ingress host value was populated logger.go:42: 07:18:55 | allinone/1-curl | Try number 0 logger.go:42: 07:18:55 | allinone/1-curl | Hostname is all-in-one-ui-kuttl-test-notable-shad.apps.ci-op-54q9kfc5-9210b.cspilp.interop.ccitredhat.com logger.go:42: 07:18:55 | allinone/1-curl | running command: [sh -c ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 200 true $NAMESPACE all-in-one-ui] logger.go:42: 07:18:55 | allinone/1-curl | Checking an expected HTTP response logger.go:42: 07:18:55 | allinone/1-curl | Running in OpenShift logger.go:42: 07:18:55 | allinone/1-curl | User not provided. Getting the token... logger.go:42: 07:18:57 | allinone/1-curl | Warning: resource jaegers/all-in-one-ui is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:19:10 | allinone/1-curl | Try number 1/30 the https://all-in-one-ui-kuttl-test-notable-shad.apps.ci-op-54q9kfc5-9210b.cspilp.interop.ccitredhat.com/search logger.go:42: 07:19:10 | allinone/1-curl | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 07:19:10 | allinone/1-curl | Try number 2/30 the https://all-in-one-ui-kuttl-test-notable-shad.apps.ci-op-54q9kfc5-9210b.cspilp.interop.ccitredhat.com/search logger.go:42: 07:19:10 | allinone/1-curl | curl response asserted properly logger.go:42: 07:19:10 | allinone/1-curl | test step completed 1-curl logger.go:42: 07:19:10 | allinone/2-delete | starting test step 2-delete logger.go:42: 07:19:10 | allinone/2-delete | Jaeger:kuttl-test-notable-shad/all-in-one-ui created logger.go:42: 07:19:10 | allinone/2-delete | test step completed 2-delete logger.go:42: 07:19:10 | allinone/3-install | starting test step 3-install logger.go:42: 07:19:10 | allinone/3-install | Jaeger:kuttl-test-notable-shad/all-in-one-ui updated logger.go:42: 07:19:10 | allinone/3-install | test step completed 3-install logger.go:42: 07:19:10 | allinone/4-test-ui-config | starting test step 4-test-ui-config logger.go:42: 07:19:10 | allinone/4-test-ui-config | running command: [./ensure-ingress-host.sh] logger.go:42: 07:19:10 | allinone/4-test-ui-config | Checking the Ingress host value was populated logger.go:42: 07:19:10 | allinone/4-test-ui-config | Try number 0 logger.go:42: 07:19:11 | allinone/4-test-ui-config | error: error executing jsonpath "{.items[0].status.ingress[0].host}": Error executing template: array index out of bounds: index 0, length 0. Printing more information for debugging the template: logger.go:42: 07:19:11 | allinone/4-test-ui-config | template was: logger.go:42: 07:19:11 | allinone/4-test-ui-config | {.items[0].status.ingress[0].host} logger.go:42: 07:19:11 | allinone/4-test-ui-config | object given to jsonpath engine was: logger.go:42: 07:19:11 | allinone/4-test-ui-config | map[string]interface {}{"apiVersion":"v1", "items":[]interface {}{}, "kind":"List", "metadata":map[string]interface {}{"resourceVersion":""}} logger.go:42: 07:19:11 | allinone/4-test-ui-config | logger.go:42: 07:19:11 | allinone/4-test-ui-config | logger.go:42: 07:19:21 | allinone/4-test-ui-config | Try number 1 logger.go:42: 07:19:21 | allinone/4-test-ui-config | Hostname is all-in-one-ui-kuttl-test-notable-shad.apps.ci-op-54q9kfc5-9210b.cspilp.interop.ccitredhat.com logger.go:42: 07:19:21 | allinone/4-test-ui-config | running command: [sh -c ASSERT_PRESENT=true EXPECTED_CONTENT=MyTrackingId QUERY_HOSTNAME=https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search go run ../../../../cmd-utils/uiconfig/main.go] logger.go:42: 07:19:22 | allinone/4-test-ui-config | time="2024-12-09T07:19:22Z" level=info msg="Querying https://all-in-one-ui-kuttl-test-notable-shad.apps.ci-op-54q9kfc5-9210b.cspilp.interop.ccitredhat.com/search..." logger.go:42: 07:19:22 | allinone/4-test-ui-config | time="2024-12-09T07:19:22Z" level=info msg="No secret provided for the Authorization header" logger.go:42: 07:19:22 | allinone/4-test-ui-config | time="2024-12-09T07:19:22Z" level=info msg="Polling to https://all-in-one-ui-kuttl-test-notable-shad.apps.ci-op-54q9kfc5-9210b.cspilp.interop.ccitredhat.com/search" logger.go:42: 07:19:22 | allinone/4-test-ui-config | time="2024-12-09T07:19:22Z" level=info msg="Doing request number 0" logger.go:42: 07:19:22 | allinone/4-test-ui-config | time="2024-12-09T07:19:22Z" level=info msg="Content found and asserted!" logger.go:42: 07:19:22 | allinone/4-test-ui-config | time="2024-12-09T07:19:22Z" level=info msg="Success!" logger.go:42: 07:19:22 | allinone/4-test-ui-config | test step completed 4-test-ui-config logger.go:42: 07:19:22 | allinone | allinone events from ns kuttl-test-notable-shad: logger.go:42: 07:19:22 | allinone | 2024-12-09 07:18:55 +0000 UTC Normal Pod all-in-one-ui-df49b4b6-97zrr Binding Scheduled Successfully assigned kuttl-test-notable-shad/all-in-one-ui-df49b4b6-97zrr to ip-10-0-39-223.ec2.internal default-scheduler logger.go:42: 07:19:22 | allinone | 2024-12-09 07:18:55 +0000 UTC Normal Pod all-in-one-ui-df49b4b6-97zrr AddedInterface Add eth0 [10.128.2.24/23] from ovn-kubernetes multus logger.go:42: 07:19:22 | allinone | 2024-12-09 07:18:55 +0000 UTC Normal ReplicaSet.apps all-in-one-ui-df49b4b6 SuccessfulCreate Created pod: all-in-one-ui-df49b4b6-97zrr replicaset-controller logger.go:42: 07:19:22 | allinone | 2024-12-09 07:18:55 +0000 UTC Normal Deployment.apps all-in-one-ui ScalingReplicaSet Scaled up replica set all-in-one-ui-df49b4b6 to 1 deployment-controller logger.go:42: 07:19:22 | allinone | 2024-12-09 07:18:56 +0000 UTC Normal Pod all-in-one-ui-df49b4b6-97zrr.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1fa5ed13cc8a023f1e987f6cafe86adc195c373cc2b774539df6d0fd02b780a7" already present on machine kubelet logger.go:42: 07:19:22 | allinone | 2024-12-09 07:18:56 +0000 UTC Normal Pod all-in-one-ui-df49b4b6-97zrr.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:19:22 | allinone | 2024-12-09 07:18:56 +0000 UTC Normal Pod all-in-one-ui-df49b4b6-97zrr.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:19:22 | allinone | 2024-12-09 07:18:56 +0000 UTC Normal Pod all-in-one-ui-df49b4b6-97zrr.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:19:22 | allinone | 2024-12-09 07:18:56 +0000 UTC Normal Pod all-in-one-ui-df49b4b6-97zrr.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:19:22 | allinone | 2024-12-09 07:18:56 +0000 UTC Normal Pod all-in-one-ui-df49b4b6-97zrr.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:19:22 | allinone | 2024-12-09 07:19:03 +0000 UTC Normal Pod all-in-one-ui-67c68b875c-mjsk4 Binding Scheduled Successfully assigned kuttl-test-notable-shad/all-in-one-ui-67c68b875c-mjsk4 to ip-10-0-79-227.ec2.internal default-scheduler logger.go:42: 07:19:22 | allinone | 2024-12-09 07:19:03 +0000 UTC Normal ReplicaSet.apps all-in-one-ui-67c68b875c SuccessfulCreate Created pod: all-in-one-ui-67c68b875c-mjsk4 replicaset-controller logger.go:42: 07:19:22 | allinone | 2024-12-09 07:19:03 +0000 UTC Normal Pod all-in-one-ui-df49b4b6-97zrr.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:19:22 | allinone | 2024-12-09 07:19:03 +0000 UTC Normal Pod all-in-one-ui-df49b4b6-97zrr.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:19:22 | allinone | 2024-12-09 07:19:03 +0000 UTC Normal ReplicaSet.apps all-in-one-ui-df49b4b6 SuccessfulDelete Deleted pod: all-in-one-ui-df49b4b6-97zrr replicaset-controller logger.go:42: 07:19:22 | allinone | 2024-12-09 07:19:03 +0000 UTC Normal Deployment.apps all-in-one-ui ScalingReplicaSet Scaled down replica set all-in-one-ui-df49b4b6 to 0 from 1 deployment-controller logger.go:42: 07:19:22 | allinone | 2024-12-09 07:19:03 +0000 UTC Normal Deployment.apps all-in-one-ui ScalingReplicaSet Scaled up replica set all-in-one-ui-67c68b875c to 1 deployment-controller logger.go:42: 07:19:22 | allinone | 2024-12-09 07:19:04 +0000 UTC Normal Pod all-in-one-ui-67c68b875c-mjsk4 AddedInterface Add eth0 [10.129.2.37/23] from ovn-kubernetes multus logger.go:42: 07:19:22 | allinone | 2024-12-09 07:19:04 +0000 UTC Normal Pod all-in-one-ui-67c68b875c-mjsk4.spec.containers{jaeger} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1fa5ed13cc8a023f1e987f6cafe86adc195c373cc2b774539df6d0fd02b780a7" kubelet logger.go:42: 07:19:22 | allinone | 2024-12-09 07:19:07 +0000 UTC Normal Pod all-in-one-ui-67c68b875c-mjsk4.spec.containers{jaeger} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1fa5ed13cc8a023f1e987f6cafe86adc195c373cc2b774539df6d0fd02b780a7" in 3.228s (3.229s including waiting). Image size: 147411949 bytes. kubelet logger.go:42: 07:19:22 | allinone | 2024-12-09 07:19:07 +0000 UTC Normal Pod all-in-one-ui-67c68b875c-mjsk4.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:19:22 | allinone | 2024-12-09 07:19:07 +0000 UTC Normal Pod all-in-one-ui-67c68b875c-mjsk4.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:19:22 | allinone | 2024-12-09 07:19:07 +0000 UTC Normal Pod all-in-one-ui-67c68b875c-mjsk4.spec.containers{oauth-proxy} Pulling Pulling image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" kubelet logger.go:42: 07:19:22 | allinone | 2024-12-09 07:19:10 +0000 UTC Normal Pod all-in-one-ui-67c68b875c-mjsk4.spec.containers{oauth-proxy} Pulled Successfully pulled image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" in 2.39s (2.39s including waiting). Image size: 339954870 bytes. kubelet logger.go:42: 07:19:22 | allinone | 2024-12-09 07:19:10 +0000 UTC Normal Pod all-in-one-ui-67c68b875c-mjsk4.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:19:22 | allinone | 2024-12-09 07:19:10 +0000 UTC Normal Pod all-in-one-ui-67c68b875c-mjsk4.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:19:22 | allinone | 2024-12-09 07:19:11 +0000 UTC Warning Pod all-in-one-ui-67c68b875c-mjsk4 FailedToRetrieveImagePullSecret Unable to retrieve some image pull secrets (all-in-one-ui-ui-proxy-dockercfg-lmn6k); attempting to pull the image may not succeed. kubelet logger.go:42: 07:19:22 | allinone | 2024-12-09 07:19:11 +0000 UTC Warning Pod all-in-one-ui-67c68b875c-mjsk4 FailedMount MountVolume.SetUp failed for volume "all-in-one-ui-collector-tls-config-volume" : secret "all-in-one-ui-collector-headless-tls" not found kubelet logger.go:42: 07:19:22 | allinone | 2024-12-09 07:19:11 +0000 UTC Warning Pod all-in-one-ui-67c68b875c-mjsk4 FailedMount MountVolume.SetUp failed for volume "all-in-one-ui-ui-oauth-proxy-tls" : secret "all-in-one-ui-ui-oauth-proxy-tls" not found kubelet logger.go:42: 07:19:22 | allinone | 2024-12-09 07:19:11 +0000 UTC Warning Pod all-in-one-ui-67c68b875c-mjsk4 FailedMount MountVolume.SetUp failed for volume "all-in-one-ui-sampling-configuration-volume" : configmap "all-in-one-ui-sampling-configuration" not found kubelet logger.go:42: 07:19:22 | allinone | 2024-12-09 07:19:11 +0000 UTC Warning Pod all-in-one-ui-67c68b875c-mjsk4 FailedMount MountVolume.SetUp failed for volume "all-in-one-ui-ui-configuration-volume" : configmap "all-in-one-ui-ui-configuration" not found kubelet logger.go:42: 07:19:22 | allinone | 2024-12-09 07:19:12 +0000 UTC Normal Pod all-in-one-ui-67c68b875c-mjsk4.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:19:22 | allinone | 2024-12-09 07:19:12 +0000 UTC Normal Pod all-in-one-ui-67c68b875c-mjsk4.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:19:22 | allinone | 2024-12-09 07:19:17 +0000 UTC Normal Pod all-in-one-ui-c77c5fd8f-c7t7x Binding Scheduled Successfully assigned kuttl-test-notable-shad/all-in-one-ui-c77c5fd8f-c7t7x to ip-10-0-95-250.ec2.internal default-scheduler logger.go:42: 07:19:22 | allinone | 2024-12-09 07:19:17 +0000 UTC Normal ReplicaSet.apps all-in-one-ui-c77c5fd8f SuccessfulCreate Created pod: all-in-one-ui-c77c5fd8f-c7t7x replicaset-controller logger.go:42: 07:19:22 | allinone | 2024-12-09 07:19:17 +0000 UTC Normal Deployment.apps all-in-one-ui ScalingReplicaSet Scaled up replica set all-in-one-ui-c77c5fd8f to 1 deployment-controller logger.go:42: 07:19:22 | allinone | 2024-12-09 07:19:18 +0000 UTC Normal Pod all-in-one-ui-c77c5fd8f-c7t7x AddedInterface Add eth0 [10.131.0.25/23] from ovn-kubernetes multus logger.go:42: 07:19:22 | allinone | 2024-12-09 07:19:18 +0000 UTC Normal Pod all-in-one-ui-c77c5fd8f-c7t7x.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1fa5ed13cc8a023f1e987f6cafe86adc195c373cc2b774539df6d0fd02b780a7" already present on machine kubelet logger.go:42: 07:19:22 | allinone | 2024-12-09 07:19:18 +0000 UTC Normal Pod all-in-one-ui-c77c5fd8f-c7t7x.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:19:22 | allinone | 2024-12-09 07:19:18 +0000 UTC Normal Pod all-in-one-ui-c77c5fd8f-c7t7x.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:19:22 | allinone | Deleting namespace: kuttl-test-notable-shad === CONT kuttl/harness/production logger.go:42: 07:19:28 | production | Ignoring add-tracking-id.yaml as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:19:28 | production | Ignoring ensure-ingress-host.sh as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:19:28 | production | Creating namespace: kuttl-test-new-kangaroo logger.go:42: 07:19:28 | production/1-install | starting test step 1-install logger.go:42: 07:19:28 | production/1-install | Jaeger:kuttl-test-new-kangaroo/production-ui created logger.go:42: 07:20:18 | production/1-install | test step completed 1-install logger.go:42: 07:20:18 | production/2-check-forbbiden-access | starting test step 2-check-forbbiden-access logger.go:42: 07:20:18 | production/2-check-forbbiden-access | running command: [./ensure-ingress-host.sh] logger.go:42: 07:20:18 | production/2-check-forbbiden-access | Checking the Ingress host value was populated logger.go:42: 07:20:18 | production/2-check-forbbiden-access | Try number 0 logger.go:42: 07:20:18 | production/2-check-forbbiden-access | Hostname is production-ui-kuttl-test-new-kangaroo.apps.ci-op-54q9kfc5-9210b.cspilp.interop.ccitredhat.com logger.go:42: 07:20:18 | production/2-check-forbbiden-access | running command: [sh -c INSECURE=true ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 403 true $NAMESPACE production-ui] logger.go:42: 07:20:18 | production/2-check-forbbiden-access | Checking an expected HTTP response logger.go:42: 07:20:18 | production/2-check-forbbiden-access | Running in OpenShift logger.go:42: 07:20:18 | production/2-check-forbbiden-access | Not using any secret logger.go:42: 07:20:18 | production/2-check-forbbiden-access | Try number 1/30 the https://production-ui-kuttl-test-new-kangaroo.apps.ci-op-54q9kfc5-9210b.cspilp.interop.ccitredhat.com/search logger.go:42: 07:20:18 | production/2-check-forbbiden-access | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 07:20:18 | production/2-check-forbbiden-access | Try number 2/30 the https://production-ui-kuttl-test-new-kangaroo.apps.ci-op-54q9kfc5-9210b.cspilp.interop.ccitredhat.com/search logger.go:42: 07:20:18 | production/2-check-forbbiden-access | curl response asserted properly logger.go:42: 07:20:18 | production/2-check-forbbiden-access | test step completed 2-check-forbbiden-access logger.go:42: 07:20:18 | production/3-curl | starting test step 3-curl logger.go:42: 07:20:18 | production/3-curl | running command: [./ensure-ingress-host.sh] logger.go:42: 07:20:18 | production/3-curl | Checking the Ingress host value was populated logger.go:42: 07:20:18 | production/3-curl | Try number 0 logger.go:42: 07:20:19 | production/3-curl | Hostname is production-ui-kuttl-test-new-kangaroo.apps.ci-op-54q9kfc5-9210b.cspilp.interop.ccitredhat.com logger.go:42: 07:20:19 | production/3-curl | running command: [sh -c ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 200 true $NAMESPACE production-ui] logger.go:42: 07:20:19 | production/3-curl | Checking an expected HTTP response logger.go:42: 07:20:19 | production/3-curl | Running in OpenShift logger.go:42: 07:20:19 | production/3-curl | User not provided. Getting the token... logger.go:42: 07:20:20 | production/3-curl | Warning: resource jaegers/production-ui is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:20:28 | production/3-curl | Try number 1/30 the https://production-ui-kuttl-test-new-kangaroo.apps.ci-op-54q9kfc5-9210b.cspilp.interop.ccitredhat.com/search logger.go:42: 07:20:28 | production/3-curl | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 07:20:28 | production/3-curl | Try number 2/30 the https://production-ui-kuttl-test-new-kangaroo.apps.ci-op-54q9kfc5-9210b.cspilp.interop.ccitredhat.com/search logger.go:42: 07:20:28 | production/3-curl | HTTP response is 503. 200 expected. Waiting 10 s logger.go:42: 07:20:38 | production/3-curl | Try number 3/30 the https://production-ui-kuttl-test-new-kangaroo.apps.ci-op-54q9kfc5-9210b.cspilp.interop.ccitredhat.com/search logger.go:42: 07:20:38 | production/3-curl | curl response asserted properly logger.go:42: 07:20:38 | production/3-curl | test step completed 3-curl logger.go:42: 07:20:38 | production/4-install | starting test step 4-install logger.go:42: 07:20:38 | production/4-install | Jaeger:kuttl-test-new-kangaroo/production-ui updated logger.go:42: 07:20:38 | production/4-install | test step completed 4-install logger.go:42: 07:20:38 | production/5-check-disabled-security | starting test step 5-check-disabled-security logger.go:42: 07:20:38 | production/5-check-disabled-security | running command: [./ensure-ingress-host.sh] logger.go:42: 07:20:38 | production/5-check-disabled-security | Checking the Ingress host value was populated logger.go:42: 07:20:38 | production/5-check-disabled-security | Try number 0 logger.go:42: 07:20:39 | production/5-check-disabled-security | Hostname is production-ui-kuttl-test-new-kangaroo.apps.ci-op-54q9kfc5-9210b.cspilp.interop.ccitredhat.com logger.go:42: 07:20:39 | production/5-check-disabled-security | running command: [sh -c INSECURE=true ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 200 true $NAMESPACE production-ui] logger.go:42: 07:20:39 | production/5-check-disabled-security | Checking an expected HTTP response logger.go:42: 07:20:39 | production/5-check-disabled-security | Running in OpenShift logger.go:42: 07:20:39 | production/5-check-disabled-security | Not using any secret logger.go:42: 07:20:39 | production/5-check-disabled-security | Try number 1/30 the https://production-ui-kuttl-test-new-kangaroo.apps.ci-op-54q9kfc5-9210b.cspilp.interop.ccitredhat.com/search logger.go:42: 07:20:39 | production/5-check-disabled-security | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 07:20:39 | production/5-check-disabled-security | Try number 2/30 the https://production-ui-kuttl-test-new-kangaroo.apps.ci-op-54q9kfc5-9210b.cspilp.interop.ccitredhat.com/search logger.go:42: 07:20:39 | production/5-check-disabled-security | HTTP response is 403. 200 expected. Waiting 10 s logger.go:42: 07:20:49 | production/5-check-disabled-security | Try number 3/30 the https://production-ui-kuttl-test-new-kangaroo.apps.ci-op-54q9kfc5-9210b.cspilp.interop.ccitredhat.com/search logger.go:42: 07:20:49 | production/5-check-disabled-security | curl response asserted properly logger.go:42: 07:20:49 | production/5-check-disabled-security | test step completed 5-check-disabled-security logger.go:42: 07:20:49 | production/6-check-NO-gaID | starting test step 6-check-NO-gaID logger.go:42: 07:20:49 | production/6-check-NO-gaID | running command: [./ensure-ingress-host.sh] logger.go:42: 07:20:49 | production/6-check-NO-gaID | Checking the Ingress host value was populated logger.go:42: 07:20:49 | production/6-check-NO-gaID | Try number 0 logger.go:42: 07:20:49 | production/6-check-NO-gaID | Hostname is production-ui-kuttl-test-new-kangaroo.apps.ci-op-54q9kfc5-9210b.cspilp.interop.ccitredhat.com logger.go:42: 07:20:49 | production/6-check-NO-gaID | running command: [sh -c ASSERT_PRESENT=false EXPECTED_CONTENT=MyTrackingId QUERY_HOSTNAME=https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search go run ../../../../cmd-utils/uiconfig/main.go] logger.go:42: 07:20:49 | production/6-check-NO-gaID | time="2024-12-09T07:20:49Z" level=info msg="Querying https://production-ui-kuttl-test-new-kangaroo.apps.ci-op-54q9kfc5-9210b.cspilp.interop.ccitredhat.com/search..." logger.go:42: 07:20:49 | production/6-check-NO-gaID | time="2024-12-09T07:20:49Z" level=info msg="No secret provided for the Authorization header" logger.go:42: 07:20:49 | production/6-check-NO-gaID | time="2024-12-09T07:20:49Z" level=info msg="Polling to https://production-ui-kuttl-test-new-kangaroo.apps.ci-op-54q9kfc5-9210b.cspilp.interop.ccitredhat.com/search" logger.go:42: 07:20:49 | production/6-check-NO-gaID | time="2024-12-09T07:20:49Z" level=info msg="Doing request number 0" logger.go:42: 07:20:50 | production/6-check-NO-gaID | time="2024-12-09T07:20:50Z" level=info msg="Content not found and asserted it was not found!" logger.go:42: 07:20:50 | production/6-check-NO-gaID | time="2024-12-09T07:20:50Z" level=info msg="Success!" logger.go:42: 07:20:50 | production/6-check-NO-gaID | test step completed 6-check-NO-gaID logger.go:42: 07:20:50 | production/7-add-tracking-id | starting test step 7-add-tracking-id logger.go:42: 07:20:50 | production/7-add-tracking-id | running command: [sh -c kubectl apply -f add-tracking-id.yaml -n $NAMESPACE] logger.go:42: 07:20:50 | production/7-add-tracking-id | jaeger.jaegertracing.io/production-ui configured logger.go:42: 07:20:50 | production/7-add-tracking-id | test step completed 7-add-tracking-id logger.go:42: 07:20:50 | production/8-check-gaID | starting test step 8-check-gaID logger.go:42: 07:20:50 | production/8-check-gaID | running command: [./ensure-ingress-host.sh] logger.go:42: 07:20:50 | production/8-check-gaID | Checking the Ingress host value was populated logger.go:42: 07:20:50 | production/8-check-gaID | Try number 0 logger.go:42: 07:20:50 | production/8-check-gaID | Hostname is production-ui-kuttl-test-new-kangaroo.apps.ci-op-54q9kfc5-9210b.cspilp.interop.ccitredhat.com logger.go:42: 07:20:50 | production/8-check-gaID | running command: [sh -c ASSERT_PRESENT=true EXPECTED_CONTENT=MyTrackingId QUERY_HOSTNAME=https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search go run ../../../../cmd-utils/uiconfig/main.go] logger.go:42: 07:20:51 | production/8-check-gaID | time="2024-12-09T07:20:51Z" level=info msg="Querying https://production-ui-kuttl-test-new-kangaroo.apps.ci-op-54q9kfc5-9210b.cspilp.interop.ccitredhat.com/search..." logger.go:42: 07:20:51 | production/8-check-gaID | time="2024-12-09T07:20:51Z" level=info msg="No secret provided for the Authorization header" logger.go:42: 07:20:51 | production/8-check-gaID | time="2024-12-09T07:20:51Z" level=info msg="Polling to https://production-ui-kuttl-test-new-kangaroo.apps.ci-op-54q9kfc5-9210b.cspilp.interop.ccitredhat.com/search" logger.go:42: 07:20:51 | production/8-check-gaID | time="2024-12-09T07:20:51Z" level=info msg="Doing request number 0" logger.go:42: 07:20:51 | production/8-check-gaID | time="2024-12-09T07:20:51Z" level=warning msg="Found: false . Assert: true" logger.go:42: 07:20:51 | production/8-check-gaID | time="2024-12-09T07:20:51Z" level=warning msg="The condition of the test function was not accomplished" logger.go:42: 07:20:59 | production/8-check-gaID | time="2024-12-09T07:20:59Z" level=info msg="Doing request number 1" logger.go:42: 07:21:19 | production/8-check-gaID | time="2024-12-09T07:21:19Z" level=info msg="Content found and asserted!" logger.go:42: 07:21:19 | production/8-check-gaID | time="2024-12-09T07:21:19Z" level=info msg="Success!" logger.go:42: 07:21:19 | production/8-check-gaID | test step completed 8-check-gaID logger.go:42: 07:21:19 | production | production events from ns kuttl-test-new-kangaroo: logger.go:42: 07:21:19 | production | 2024-12-09 07:19:35 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestnewkangarooproductionui-1-568b6cf5bf SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestnewkangarooproductionui-1-568b6mrt49 replicaset-controller logger.go:42: 07:21:19 | production | 2024-12-09 07:19:35 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestnewkangarooproductionui-1-568b6mrt49 Binding Scheduled Successfully assigned kuttl-test-new-kangaroo/elasticsearch-cdm-kuttltestnewkangarooproductionui-1-568b6mrt49 to ip-10-0-95-250.ec2.internal default-scheduler logger.go:42: 07:21:19 | production | 2024-12-09 07:19:35 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestnewkangarooproductionui-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestnewkangarooproductionui-1-568b6cf5bf to 1 deployment-controller logger.go:42: 07:21:19 | production | 2024-12-09 07:19:36 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestnewkangarooproductionui-1-568b6mrt49 AddedInterface Add eth0 [10.131.0.26/23] from ovn-kubernetes multus logger.go:42: 07:21:19 | production | 2024-12-09 07:19:36 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestnewkangarooproductionui-1-568b6mrt49.spec.containers{elasticsearch} Pulling Pulling image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:8e4fbea4983cd58352349ca291383169b286bc166fad95a87807552ca43335e6" kubelet logger.go:42: 07:21:19 | production | 2024-12-09 07:19:43 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestnewkangarooproductionui-1-568b6mrt49.spec.containers{elasticsearch} Pulled Successfully pulled image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:8e4fbea4983cd58352349ca291383169b286bc166fad95a87807552ca43335e6" in 7.379s (7.379s including waiting). Image size: 523762405 bytes. kubelet logger.go:42: 07:21:19 | production | 2024-12-09 07:19:43 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestnewkangarooproductionui-1-568b6mrt49.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:21:19 | production | 2024-12-09 07:19:43 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestnewkangarooproductionui-1-568b6mrt49.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:21:19 | production | 2024-12-09 07:19:43 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestnewkangarooproductionui-1-568b6mrt49.spec.containers{proxy} Pulling Pulling image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:d68824b0b2c84db8e33edf9ab344eb684c4a7ebd7ef162bbc309043adcb28e6b" kubelet logger.go:42: 07:21:19 | production | 2024-12-09 07:19:46 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestnewkangarooproductionui-1-568b6mrt49.spec.containers{proxy} Pulled Successfully pulled image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:d68824b0b2c84db8e33edf9ab344eb684c4a7ebd7ef162bbc309043adcb28e6b" in 2.995s (2.995s including waiting). Image size: 272839959 bytes. kubelet logger.go:42: 07:21:19 | production | 2024-12-09 07:19:46 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestnewkangarooproductionui-1-568b6mrt49.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:21:19 | production | 2024-12-09 07:19:46 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestnewkangarooproductionui-1-568b6mrt49.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:21:19 | production | 2024-12-09 07:19:55 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestnewkangarooproductionui-1-568b6mrt49.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 07:21:19 | production | 2024-12-09 07:20:00 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestnewkangarooproductionui-1-568b6mrt49.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:21:19 | production | 2024-12-09 07:20:11 +0000 UTC Normal Pod production-ui-collector-75d64dbfb7-l6wkz Binding Scheduled Successfully assigned kuttl-test-new-kangaroo/production-ui-collector-75d64dbfb7-l6wkz to ip-10-0-39-223.ec2.internal default-scheduler logger.go:42: 07:21:19 | production | 2024-12-09 07:20:11 +0000 UTC Normal ReplicaSet.apps production-ui-collector-75d64dbfb7 SuccessfulCreate Created pod: production-ui-collector-75d64dbfb7-l6wkz replicaset-controller logger.go:42: 07:21:19 | production | 2024-12-09 07:20:11 +0000 UTC Normal Deployment.apps production-ui-collector ScalingReplicaSet Scaled up replica set production-ui-collector-75d64dbfb7 to 1 deployment-controller logger.go:42: 07:21:19 | production | 2024-12-09 07:20:11 +0000 UTC Normal Pod production-ui-query-845c66876-8wjv8 Binding Scheduled Successfully assigned kuttl-test-new-kangaroo/production-ui-query-845c66876-8wjv8 to ip-10-0-79-227.ec2.internal default-scheduler logger.go:42: 07:21:19 | production | 2024-12-09 07:20:11 +0000 UTC Normal ReplicaSet.apps production-ui-query-845c66876 SuccessfulCreate Created pod: production-ui-query-845c66876-8wjv8 replicaset-controller logger.go:42: 07:21:19 | production | 2024-12-09 07:20:11 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled up replica set production-ui-query-845c66876 to 1 deployment-controller logger.go:42: 07:21:19 | production | 2024-12-09 07:20:12 +0000 UTC Normal Pod production-ui-collector-75d64dbfb7-l6wkz AddedInterface Add eth0 [10.128.2.25/23] from ovn-kubernetes multus logger.go:42: 07:21:19 | production | 2024-12-09 07:20:12 +0000 UTC Normal Pod production-ui-collector-75d64dbfb7-l6wkz.spec.containers{jaeger-collector} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:44686d560aa27cae8ff8693f88c4cb6e2edf1737010ec1f80709cb42250b729d" kubelet logger.go:42: 07:21:19 | production | 2024-12-09 07:20:12 +0000 UTC Normal Pod production-ui-query-845c66876-8wjv8 AddedInterface Add eth0 [10.129.2.38/23] from ovn-kubernetes multus logger.go:42: 07:21:19 | production | 2024-12-09 07:20:12 +0000 UTC Normal Pod production-ui-query-845c66876-8wjv8.spec.containers{jaeger-query} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:24a83160aa930e1b72c2a2442a33b28af2b06c24b058e09afcd0d495a8066d6d" kubelet logger.go:42: 07:21:19 | production | 2024-12-09 07:20:14 +0000 UTC Normal Pod production-ui-collector-75d64dbfb7-l6wkz.spec.containers{jaeger-collector} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:44686d560aa27cae8ff8693f88c4cb6e2edf1737010ec1f80709cb42250b729d" in 2.384s (2.384s including waiting). Image size: 142489237 bytes. kubelet logger.go:42: 07:21:19 | production | 2024-12-09 07:20:14 +0000 UTC Normal Pod production-ui-collector-75d64dbfb7-l6wkz.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:21:19 | production | 2024-12-09 07:20:14 +0000 UTC Normal Pod production-ui-collector-75d64dbfb7-l6wkz.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:21:19 | production | 2024-12-09 07:20:16 +0000 UTC Normal Pod production-ui-query-845c66876-8wjv8.spec.containers{jaeger-query} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:24a83160aa930e1b72c2a2442a33b28af2b06c24b058e09afcd0d495a8066d6d" in 4.423s (4.423s including waiting). Image size: 192936692 bytes. kubelet logger.go:42: 07:21:19 | production | 2024-12-09 07:20:16 +0000 UTC Normal Pod production-ui-query-845c66876-8wjv8.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:21:19 | production | 2024-12-09 07:20:16 +0000 UTC Normal Pod production-ui-query-845c66876-8wjv8.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:21:19 | production | 2024-12-09 07:20:16 +0000 UTC Normal Pod production-ui-query-845c66876-8wjv8.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:21:19 | production | 2024-12-09 07:20:16 +0000 UTC Normal Pod production-ui-query-845c66876-8wjv8.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:21:19 | production | 2024-12-09 07:20:16 +0000 UTC Normal Pod production-ui-query-845c66876-8wjv8.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:21:19 | production | 2024-12-09 07:20:16 +0000 UTC Normal Pod production-ui-query-845c66876-8wjv8.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" already present on machine kubelet logger.go:42: 07:21:19 | production | 2024-12-09 07:20:16 +0000 UTC Normal Pod production-ui-query-845c66876-8wjv8.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:21:19 | production | 2024-12-09 07:20:16 +0000 UTC Normal Pod production-ui-query-845c66876-8wjv8.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:21:19 | production | 2024-12-09 07:20:24 +0000 UTC Normal Pod production-ui-query-7b48fffd85-ftr7t Binding Scheduled Successfully assigned kuttl-test-new-kangaroo/production-ui-query-7b48fffd85-ftr7t to ip-10-0-79-227.ec2.internal default-scheduler logger.go:42: 07:21:19 | production | 2024-12-09 07:20:24 +0000 UTC Normal ReplicaSet.apps production-ui-query-7b48fffd85 SuccessfulCreate Created pod: production-ui-query-7b48fffd85-ftr7t replicaset-controller logger.go:42: 07:21:19 | production | 2024-12-09 07:20:24 +0000 UTC Normal Pod production-ui-query-845c66876-8wjv8.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:21:19 | production | 2024-12-09 07:20:24 +0000 UTC Normal Pod production-ui-query-845c66876-8wjv8.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:21:19 | production | 2024-12-09 07:20:24 +0000 UTC Normal Pod production-ui-query-845c66876-8wjv8.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:21:19 | production | 2024-12-09 07:20:24 +0000 UTC Normal ReplicaSet.apps production-ui-query-845c66876 SuccessfulDelete Deleted pod: production-ui-query-845c66876-8wjv8 replicaset-controller logger.go:42: 07:21:19 | production | 2024-12-09 07:20:24 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled down replica set production-ui-query-845c66876 to 0 from 1 deployment-controller logger.go:42: 07:21:19 | production | 2024-12-09 07:20:24 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled up replica set production-ui-query-7b48fffd85 to 1 deployment-controller logger.go:42: 07:21:19 | production | 2024-12-09 07:20:25 +0000 UTC Normal Pod production-ui-query-7b48fffd85-ftr7t AddedInterface Add eth0 [10.129.2.39/23] from ovn-kubernetes multus logger.go:42: 07:21:19 | production | 2024-12-09 07:20:25 +0000 UTC Normal Pod production-ui-query-7b48fffd85-ftr7t.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:24a83160aa930e1b72c2a2442a33b28af2b06c24b058e09afcd0d495a8066d6d" already present on machine kubelet logger.go:42: 07:21:19 | production | 2024-12-09 07:20:25 +0000 UTC Normal Pod production-ui-query-7b48fffd85-ftr7t.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:21:19 | production | 2024-12-09 07:20:25 +0000 UTC Normal Pod production-ui-query-7b48fffd85-ftr7t.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:21:19 | production | 2024-12-09 07:20:25 +0000 UTC Normal Pod production-ui-query-7b48fffd85-ftr7t.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:21:19 | production | 2024-12-09 07:20:25 +0000 UTC Normal Pod production-ui-query-7b48fffd85-ftr7t.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:21:19 | production | 2024-12-09 07:20:25 +0000 UTC Normal Pod production-ui-query-7b48fffd85-ftr7t.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:21:19 | production | 2024-12-09 07:20:25 +0000 UTC Normal Pod production-ui-query-7b48fffd85-ftr7t.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" already present on machine kubelet logger.go:42: 07:21:19 | production | 2024-12-09 07:20:25 +0000 UTC Normal Pod production-ui-query-7b48fffd85-ftr7t.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:21:19 | production | 2024-12-09 07:20:25 +0000 UTC Normal Pod production-ui-query-7b48fffd85-ftr7t.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:21:19 | production | 2024-12-09 07:20:26 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling production-ui-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:21:19 | production | 2024-12-09 07:20:26 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling production-ui-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:21:19 | production | 2024-12-09 07:20:26 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling production-ui-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:21:19 | production | 2024-12-09 07:20:40 +0000 UTC Normal Pod production-ui-query-7b48fffd85-ftr7t.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:21:19 | production | 2024-12-09 07:20:40 +0000 UTC Normal Pod production-ui-query-7b48fffd85-ftr7t.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:21:19 | production | 2024-12-09 07:20:40 +0000 UTC Normal Pod production-ui-query-7b48fffd85-ftr7t.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:21:19 | production | 2024-12-09 07:20:40 +0000 UTC Normal ReplicaSet.apps production-ui-query-7b48fffd85 SuccessfulDelete Deleted pod: production-ui-query-7b48fffd85-ftr7t replicaset-controller logger.go:42: 07:21:19 | production | 2024-12-09 07:20:40 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled down replica set production-ui-query-7b48fffd85 to 0 from 1 deployment-controller logger.go:42: 07:21:19 | production | 2024-12-09 07:20:41 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling production-ui-collector FailedGetResourceMetric failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 07:21:19 | production | 2024-12-09 07:20:41 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling production-ui-collector FailedGetResourceMetric failed to get memory utilization: missing request for memory in container jaeger-collector of Pod production-ui-collector-75d64dbfb7-l6wkz horizontal-pod-autoscaler logger.go:42: 07:21:19 | production | 2024-12-09 07:20:41 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling production-ui-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 07:21:19 | production | 2024-12-09 07:20:41 +0000 UTC Normal Pod production-ui-query-55679b59b7-wbcq7 Binding Scheduled Successfully assigned kuttl-test-new-kangaroo/production-ui-query-55679b59b7-wbcq7 to ip-10-0-79-227.ec2.internal default-scheduler logger.go:42: 07:21:19 | production | 2024-12-09 07:20:41 +0000 UTC Normal ReplicaSet.apps production-ui-query-55679b59b7 SuccessfulCreate Created pod: production-ui-query-55679b59b7-wbcq7 replicaset-controller logger.go:42: 07:21:19 | production | 2024-12-09 07:20:41 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled up replica set production-ui-query-55679b59b7 to 1 deployment-controller logger.go:42: 07:21:19 | production | 2024-12-09 07:20:42 +0000 UTC Normal Pod production-ui-query-55679b59b7-wbcq7 AddedInterface Add eth0 [10.129.2.40/23] from ovn-kubernetes multus logger.go:42: 07:21:19 | production | 2024-12-09 07:20:42 +0000 UTC Normal Pod production-ui-query-55679b59b7-wbcq7.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:24a83160aa930e1b72c2a2442a33b28af2b06c24b058e09afcd0d495a8066d6d" already present on machine kubelet logger.go:42: 07:21:19 | production | 2024-12-09 07:20:42 +0000 UTC Normal Pod production-ui-query-55679b59b7-wbcq7.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:21:19 | production | 2024-12-09 07:20:42 +0000 UTC Normal Pod production-ui-query-55679b59b7-wbcq7.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:21:19 | production | 2024-12-09 07:20:42 +0000 UTC Normal Pod production-ui-query-55679b59b7-wbcq7.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" already present on machine kubelet logger.go:42: 07:21:19 | production | 2024-12-09 07:20:42 +0000 UTC Normal Pod production-ui-query-55679b59b7-wbcq7.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:21:19 | production | 2024-12-09 07:20:42 +0000 UTC Normal Pod production-ui-query-55679b59b7-wbcq7.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:21:19 | production | 2024-12-09 07:20:51 +0000 UTC Normal Pod production-ui-query-55679b59b7-wbcq7.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:21:19 | production | 2024-12-09 07:20:51 +0000 UTC Normal Pod production-ui-query-55679b59b7-wbcq7.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:21:19 | production | 2024-12-09 07:20:51 +0000 UTC Normal ReplicaSet.apps production-ui-query-55679b59b7 SuccessfulDelete Deleted pod: production-ui-query-55679b59b7-wbcq7 replicaset-controller logger.go:42: 07:21:19 | production | 2024-12-09 07:20:51 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled down replica set production-ui-query-55679b59b7 to 0 from 1 deployment-controller logger.go:42: 07:21:19 | production | 2024-12-09 07:20:52 +0000 UTC Normal Pod production-ui-query-5449587dc4-vv529 Binding Scheduled Successfully assigned kuttl-test-new-kangaroo/production-ui-query-5449587dc4-vv529 to ip-10-0-79-227.ec2.internal default-scheduler logger.go:42: 07:21:19 | production | 2024-12-09 07:20:52 +0000 UTC Normal ReplicaSet.apps production-ui-query-5449587dc4 SuccessfulCreate Created pod: production-ui-query-5449587dc4-vv529 replicaset-controller logger.go:42: 07:21:19 | production | 2024-12-09 07:20:52 +0000 UTC Normal Deployment.apps production-ui-query ScalingReplicaSet Scaled up replica set production-ui-query-5449587dc4 to 1 deployment-controller logger.go:42: 07:21:19 | production | 2024-12-09 07:20:53 +0000 UTC Normal Pod production-ui-query-5449587dc4-vv529 AddedInterface Add eth0 [10.129.2.41/23] from ovn-kubernetes multus logger.go:42: 07:21:19 | production | 2024-12-09 07:20:53 +0000 UTC Normal Pod production-ui-query-5449587dc4-vv529.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:24a83160aa930e1b72c2a2442a33b28af2b06c24b058e09afcd0d495a8066d6d" already present on machine kubelet logger.go:42: 07:21:19 | production | 2024-12-09 07:20:53 +0000 UTC Normal Pod production-ui-query-5449587dc4-vv529.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:21:19 | production | 2024-12-09 07:20:53 +0000 UTC Normal Pod production-ui-query-5449587dc4-vv529.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:21:19 | production | 2024-12-09 07:20:53 +0000 UTC Normal Pod production-ui-query-5449587dc4-vv529.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" already present on machine kubelet logger.go:42: 07:21:19 | production | 2024-12-09 07:20:53 +0000 UTC Normal Pod production-ui-query-5449587dc4-vv529.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:21:19 | production | 2024-12-09 07:20:53 +0000 UTC Normal Pod production-ui-query-5449587dc4-vv529.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:21:19 | production | 2024-12-09 07:21:11 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling production-ui-collector FailedGetResourceMetric failed to get cpu utilization: missing request for cpu in container jaeger-collector of Pod production-ui-collector-75d64dbfb7-l6wkz horizontal-pod-autoscaler logger.go:42: 07:21:19 | production | 2024-12-09 07:21:11 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling production-ui-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: missing request for cpu in container jaeger-collector of Pod production-ui-collector-75d64dbfb7-l6wkz horizontal-pod-autoscaler logger.go:42: 07:21:19 | production | Deleting namespace: kuttl-test-new-kangaroo === CONT kuttl/harness/artifacts logger.go:42: 07:21:25 | artifacts | Creating namespace: kuttl-test-singular-shark logger.go:42: 07:21:25 | artifacts | artifacts events from ns kuttl-test-singular-shark: logger.go:42: 07:21:25 | artifacts | Deleting namespace: kuttl-test-singular-shark === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (160.71s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/allinone (37.66s) --- PASS: kuttl/harness/production (116.74s) --- PASS: kuttl/harness/artifacts (6.12s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name ui --report --output /logs/artifacts/ui.xml ./artifacts/kuttl-report.xml time="2024-12-09T07:21:31Z" level=debug msg="Setting a new name for the test suites" time="2024-12-09T07:21:31Z" level=debug msg="Removing 'artifacts' TestCase" time="2024-12-09T07:21:31Z" level=debug msg="normalizing test case names" time="2024-12-09T07:21:31Z" level=debug msg="ui/allinone -> ui_allinone" time="2024-12-09T07:21:31Z" level=debug msg="ui/production -> ui_production" time="2024-12-09T07:21:31Z" level=debug msg="ui/artifacts -> ui_artifacts" +---------------+--------+ | NAME | RESULT | +---------------+--------+ | ui_allinone | passed | | ui_production | passed | | ui_artifacts | passed | +---------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh miscellaneous false true + '[' 3 -ne 3 ']' + test_suite_name=miscellaneous + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/miscellaneous.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-miscellaneous make[2]: Entering directory '/tmp/jaeger-tests' SKIP_ES_EXTERNAL=true ./tests/e2e/miscellaneous/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.18.0-0.nightly-2024-12-05-191355 True False 16m Cluster version is 4.18.0-0.nightly-2024-12-05-191355' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.18.0-0.nightly-2024-12-05-191355 True False 16m Cluster version is 4.18.0-0.nightly-2024-12-05-191355' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/miscellaneous/render.sh ++ export SUITE_DIR=./tests/e2e/miscellaneous ++ SUITE_DIR=./tests/e2e/miscellaneous ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/miscellaneous ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + start_test collector-autoscale + '[' 1 -ne 1 ']' + test_name=collector-autoscale + echo =========================================================================== =========================================================================== + info 'Rendering files for test collector-autoscale' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test collector-autoscale\e[0m' Rendering files for test collector-autoscale + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build + '[' _build '!=' _build ']' + mkdir -p collector-autoscale + cd collector-autoscale + jaeger_name=simple-prod + jaeger_deploy_mode=production + [[ true = true ]] + [[ true = true ]] + jaeger_deploy_mode=production_autoprovisioned + ELASTICSEARCH_NODECOUNT=1 + render_install_jaeger simple-prod production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=simple-prod + JAEGER_NAME=simple-prod + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.collector.resources.requests.memory="200m"' 01-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.collector.autoscale=true 01-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.collector.minReplicas=1 01-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.collector.maxReplicas=2 01-install.yaml + version_lt 1.30 1.23 ++ echo 1.30 1.23 ++ tr ' ' '\n' ++ sort -rV ++ head -n 1 + test 1.30 '!=' 1.30 + rm ./03-assert.yaml + generate_otlp_e2e_tests http + test_protocol=http + is_secured=false + '[' true = true ']' + is_secured=true + start_test collector-otlp-allinone-http + '[' 1 -ne 1 ']' + test_name=collector-otlp-allinone-http + echo =========================================================================== =========================================================================== + info 'Rendering files for test collector-otlp-allinone-http' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test collector-otlp-allinone-http\e[0m' Rendering files for test collector-otlp-allinone-http + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/collector-autoscale + '[' collector-autoscale '!=' _build ']' + cd .. + mkdir -p collector-otlp-allinone-http + cd collector-otlp-allinone-http + render_install_jaeger my-jaeger allInOne 00 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=allInOne + test_step=00 + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_otlp_smoke_test my-jaeger http true 01 + '[' 4 -ne 4 ']' + jaeger=my-jaeger + reporting_protocol=http + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template + '[' http = grpc ']' + reporting_port=:4318 + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 + OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + REPORTING_PROTOCOL=http + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset OTEL_EXPORTER_OTLP_ENDPOINT + start_test collector-otlp-production-http + '[' 1 -ne 1 ']' + test_name=collector-otlp-production-http + echo =========================================================================== =========================================================================== + info 'Rendering files for test collector-otlp-production-http' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test collector-otlp-production-http\e[0m' Rendering files for test collector-otlp-production-http + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/collector-otlp-allinone-http + '[' collector-otlp-allinone-http '!=' _build ']' + cd .. + mkdir -p collector-otlp-production-http + cd collector-otlp-production-http + jaeger_deploy_mode=production + [[ true = true ]] + [[ true = true ]] + jaeger_deploy_mode=production_autoprovisioned + render_install_jaeger my-jaeger production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + render_otlp_smoke_test my-jaeger http true 02 + '[' 4 -ne 4 ']' + jaeger=my-jaeger + reporting_protocol=http + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template + '[' http = grpc ']' + reporting_port=:4318 + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 + OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + REPORTING_PROTOCOL=http + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset OTEL_EXPORTER_OTLP_ENDPOINT + generate_otlp_e2e_tests grpc + test_protocol=grpc + is_secured=false + '[' true = true ']' + is_secured=true + start_test collector-otlp-allinone-grpc + '[' 1 -ne 1 ']' + test_name=collector-otlp-allinone-grpc + echo =========================================================================== =========================================================================== + info 'Rendering files for test collector-otlp-allinone-grpc' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test collector-otlp-allinone-grpc\e[0m' Rendering files for test collector-otlp-allinone-grpc + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/collector-otlp-production-http + '[' collector-otlp-production-http '!=' _build ']' + cd .. + mkdir -p collector-otlp-allinone-grpc + cd collector-otlp-allinone-grpc + render_install_jaeger my-jaeger allInOne 00 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=allInOne + test_step=00 + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_otlp_smoke_test my-jaeger grpc true 01 + '[' 4 -ne 4 ']' + jaeger=my-jaeger + reporting_protocol=grpc + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template + '[' grpc = grpc ']' + reporting_port=:4317 + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 + OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + REPORTING_PROTOCOL=grpc + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset OTEL_EXPORTER_OTLP_ENDPOINT + start_test collector-otlp-production-grpc + '[' 1 -ne 1 ']' + test_name=collector-otlp-production-grpc + echo =========================================================================== =========================================================================== + info 'Rendering files for test collector-otlp-production-grpc' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test collector-otlp-production-grpc\e[0m' Rendering files for test collector-otlp-production-grpc + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/collector-otlp-allinone-grpc + '[' collector-otlp-allinone-grpc '!=' _build ']' + cd .. + mkdir -p collector-otlp-production-grpc + cd collector-otlp-production-grpc + jaeger_deploy_mode=production + [[ true = true ]] + [[ true = true ]] + jaeger_deploy_mode=production_autoprovisioned + render_install_jaeger my-jaeger production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + render_otlp_smoke_test my-jaeger grpc true 02 + '[' 4 -ne 4 ']' + jaeger=my-jaeger + reporting_protocol=grpc + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template + '[' grpc = grpc ']' + reporting_port=:4317 + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 + OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + REPORTING_PROTOCOL=grpc + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/otlp-smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset OTEL_EXPORTER_OTLP_ENDPOINT + '[' true = true ']' + skip_test istio 'Test not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=istio + message='Test not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/collector-otlp-production-grpc + '[' collector-otlp-production-grpc '!=' _build ']' + cd .. + rm -rf istio + warning 'istio: Test not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: istio: Test not supported in OpenShift\e[0m' WAR: istio: Test not supported in OpenShift + '[' true = true ']' + skip_test outside-cluster 'Test not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=outside-cluster + message='Test not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build + '[' _build '!=' _build ']' + rm -rf outside-cluster + warning 'outside-cluster: Test not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: outside-cluster: Test not supported in OpenShift\e[0m' WAR: outside-cluster: Test not supported in OpenShift + start_test set-custom-img + '[' 1 -ne 1 ']' + test_name=set-custom-img + echo =========================================================================== =========================================================================== + info 'Rendering files for test set-custom-img' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test set-custom-img\e[0m' Rendering files for test set-custom-img + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build + '[' _build '!=' _build ']' + mkdir -p set-custom-img + cd set-custom-img + jaeger_name=my-jaeger + jaeger_deploy_mode=production + [[ true = true ]] + [[ true = true ]] + jaeger_deploy_mode=production_autoprovisioned + render_install_jaeger my-jaeger production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + cp ./01-install.yaml ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.collector.image="test"' ./02-install.yaml + '[' true = true ']' + skip_test non-cluster-wide 'Test not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=non-cluster-wide + message='Test not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/miscellaneous/_build/set-custom-img + '[' set-custom-img '!=' _build ']' + cd .. + rm -rf non-cluster-wide + warning 'non-cluster-wide: Test not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: non-cluster-wide: Test not supported in OpenShift\e[0m' WAR: non-cluster-wide: Test not supported in OpenShift make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running miscellaneous E2E tests' Running miscellaneous E2E tests + cd tests/e2e/miscellaneous/_build + set +e + KUBECONFIG=/tmp/kubeconfig-591533165 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-54q9kfc5-9210b.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 8 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/cassandra-spark === PAUSE kuttl/harness/cassandra-spark === RUN kuttl/harness/collector-autoscale === PAUSE kuttl/harness/collector-autoscale === RUN kuttl/harness/collector-otlp-allinone-grpc === PAUSE kuttl/harness/collector-otlp-allinone-grpc === RUN kuttl/harness/collector-otlp-allinone-http === PAUSE kuttl/harness/collector-otlp-allinone-http === RUN kuttl/harness/collector-otlp-production-grpc === PAUSE kuttl/harness/collector-otlp-production-grpc === RUN kuttl/harness/collector-otlp-production-http === PAUSE kuttl/harness/collector-otlp-production-http === RUN kuttl/harness/set-custom-img === PAUSE kuttl/harness/set-custom-img === CONT kuttl/harness/artifacts logger.go:42: 07:21:42 | artifacts | Creating namespace: kuttl-test-pleasant-weevil logger.go:42: 07:21:42 | artifacts | artifacts events from ns kuttl-test-pleasant-weevil: logger.go:42: 07:21:42 | artifacts | Deleting namespace: kuttl-test-pleasant-weevil === CONT kuttl/harness/collector-otlp-allinone-http logger.go:42: 07:21:48 | collector-otlp-allinone-http | Creating namespace: kuttl-test-wanted-emu logger.go:42: 07:21:48 | collector-otlp-allinone-http/0-install | starting test step 0-install logger.go:42: 07:21:49 | collector-otlp-allinone-http/0-install | Jaeger:kuttl-test-wanted-emu/my-jaeger created logger.go:42: 07:21:54 | collector-otlp-allinone-http/0-install | test step completed 0-install logger.go:42: 07:21:54 | collector-otlp-allinone-http/1-smoke-test | starting test step 1-smoke-test logger.go:42: 07:21:54 | collector-otlp-allinone-http/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:21:55 | collector-otlp-allinone-http/1-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:22:01 | collector-otlp-allinone-http/1-smoke-test | running command: [sh -c REPORTING_PROTOCOL=http ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/otlp-smoke-test.yaml.template -o otlp-smoke-test-job.yaml] logger.go:42: 07:22:02 | collector-otlp-allinone-http/1-smoke-test | running command: [sh -c kubectl create -f otlp-smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:22:02 | collector-otlp-allinone-http/1-smoke-test | job.batch/report-span created logger.go:42: 07:22:02 | collector-otlp-allinone-http/1-smoke-test | job.batch/check-span created logger.go:42: 07:22:16 | collector-otlp-allinone-http/1-smoke-test | test step completed 1-smoke-test logger.go:42: 07:22:16 | collector-otlp-allinone-http | collector-otlp-allinone-http events from ns kuttl-test-wanted-emu: logger.go:42: 07:22:16 | collector-otlp-allinone-http | 2024-12-09 07:21:53 +0000 UTC Normal Pod my-jaeger-6ffb7594d6-ptpqr Binding Scheduled Successfully assigned kuttl-test-wanted-emu/my-jaeger-6ffb7594d6-ptpqr to ip-10-0-79-227.ec2.internal default-scheduler logger.go:42: 07:22:16 | collector-otlp-allinone-http | 2024-12-09 07:21:53 +0000 UTC Normal Pod my-jaeger-6ffb7594d6-ptpqr AddedInterface Add eth0 [10.129.2.42/23] from ovn-kubernetes multus logger.go:42: 07:22:16 | collector-otlp-allinone-http | 2024-12-09 07:21:53 +0000 UTC Normal Pod my-jaeger-6ffb7594d6-ptpqr.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1fa5ed13cc8a023f1e987f6cafe86adc195c373cc2b774539df6d0fd02b780a7" already present on machine kubelet logger.go:42: 07:22:16 | collector-otlp-allinone-http | 2024-12-09 07:21:53 +0000 UTC Normal Pod my-jaeger-6ffb7594d6-ptpqr.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:22:16 | collector-otlp-allinone-http | 2024-12-09 07:21:53 +0000 UTC Normal Pod my-jaeger-6ffb7594d6-ptpqr.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:22:16 | collector-otlp-allinone-http | 2024-12-09 07:21:53 +0000 UTC Normal Pod my-jaeger-6ffb7594d6-ptpqr.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:22:16 | collector-otlp-allinone-http | 2024-12-09 07:21:53 +0000 UTC Normal Pod my-jaeger-6ffb7594d6-ptpqr.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:22:16 | collector-otlp-allinone-http | 2024-12-09 07:21:53 +0000 UTC Normal Pod my-jaeger-6ffb7594d6-ptpqr.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:22:16 | collector-otlp-allinone-http | 2024-12-09 07:21:53 +0000 UTC Normal ReplicaSet.apps my-jaeger-6ffb7594d6 SuccessfulCreate Created pod: my-jaeger-6ffb7594d6-ptpqr replicaset-controller logger.go:42: 07:22:16 | collector-otlp-allinone-http | 2024-12-09 07:21:53 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-6ffb7594d6 to 1 deployment-controller logger.go:42: 07:22:16 | collector-otlp-allinone-http | 2024-12-09 07:21:58 +0000 UTC Normal Pod my-jaeger-6ffb7594d6-ptpqr.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:22:16 | collector-otlp-allinone-http | 2024-12-09 07:21:58 +0000 UTC Normal Pod my-jaeger-6ffb7594d6-ptpqr.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:22:16 | collector-otlp-allinone-http | 2024-12-09 07:21:58 +0000 UTC Normal ReplicaSet.apps my-jaeger-6ffb7594d6 SuccessfulDelete Deleted pod: my-jaeger-6ffb7594d6-ptpqr replicaset-controller logger.go:42: 07:22:16 | collector-otlp-allinone-http | 2024-12-09 07:21:58 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled down replica set my-jaeger-6ffb7594d6 to 0 from 1 deployment-controller logger.go:42: 07:22:16 | collector-otlp-allinone-http | 2024-12-09 07:21:59 +0000 UTC Normal Pod my-jaeger-7b8695dbd6-mpvbs Binding Scheduled Successfully assigned kuttl-test-wanted-emu/my-jaeger-7b8695dbd6-mpvbs to ip-10-0-79-227.ec2.internal default-scheduler logger.go:42: 07:22:16 | collector-otlp-allinone-http | 2024-12-09 07:21:59 +0000 UTC Normal Pod my-jaeger-7b8695dbd6-mpvbs AddedInterface Add eth0 [10.129.2.43/23] from ovn-kubernetes multus logger.go:42: 07:22:16 | collector-otlp-allinone-http | 2024-12-09 07:21:59 +0000 UTC Normal Pod my-jaeger-7b8695dbd6-mpvbs.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1fa5ed13cc8a023f1e987f6cafe86adc195c373cc2b774539df6d0fd02b780a7" already present on machine kubelet logger.go:42: 07:22:16 | collector-otlp-allinone-http | 2024-12-09 07:21:59 +0000 UTC Normal Pod my-jaeger-7b8695dbd6-mpvbs.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:22:16 | collector-otlp-allinone-http | 2024-12-09 07:21:59 +0000 UTC Normal Pod my-jaeger-7b8695dbd6-mpvbs.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:22:16 | collector-otlp-allinone-http | 2024-12-09 07:21:59 +0000 UTC Normal Pod my-jaeger-7b8695dbd6-mpvbs.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:22:16 | collector-otlp-allinone-http | 2024-12-09 07:21:59 +0000 UTC Normal Pod my-jaeger-7b8695dbd6-mpvbs.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:22:16 | collector-otlp-allinone-http | 2024-12-09 07:21:59 +0000 UTC Normal Pod my-jaeger-7b8695dbd6-mpvbs.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:22:16 | collector-otlp-allinone-http | 2024-12-09 07:21:59 +0000 UTC Normal ReplicaSet.apps my-jaeger-7b8695dbd6 SuccessfulCreate Created pod: my-jaeger-7b8695dbd6-mpvbs replicaset-controller logger.go:42: 07:22:16 | collector-otlp-allinone-http | 2024-12-09 07:21:59 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-7b8695dbd6 to 1 deployment-controller logger.go:42: 07:22:16 | collector-otlp-allinone-http | 2024-12-09 07:22:02 +0000 UTC Normal Pod check-span-bk5nx Binding Scheduled Successfully assigned kuttl-test-wanted-emu/check-span-bk5nx to ip-10-0-39-223.ec2.internal default-scheduler logger.go:42: 07:22:16 | collector-otlp-allinone-http | 2024-12-09 07:22:02 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-bk5nx job-controller logger.go:42: 07:22:16 | collector-otlp-allinone-http | 2024-12-09 07:22:02 +0000 UTC Normal Pod report-span-8b52g Binding Scheduled Successfully assigned kuttl-test-wanted-emu/report-span-8b52g to ip-10-0-95-250.ec2.internal default-scheduler logger.go:42: 07:22:16 | collector-otlp-allinone-http | 2024-12-09 07:22:02 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-8b52g job-controller logger.go:42: 07:22:16 | collector-otlp-allinone-http | 2024-12-09 07:22:03 +0000 UTC Normal Pod check-span-bk5nx AddedInterface Add eth0 [10.128.2.26/23] from ovn-kubernetes multus logger.go:42: 07:22:16 | collector-otlp-allinone-http | 2024-12-09 07:22:03 +0000 UTC Normal Pod check-span-bk5nx.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:22:16 | collector-otlp-allinone-http | 2024-12-09 07:22:03 +0000 UTC Normal Pod report-span-8b52g AddedInterface Add eth0 [10.131.0.27/23] from ovn-kubernetes multus logger.go:42: 07:22:16 | collector-otlp-allinone-http | 2024-12-09 07:22:03 +0000 UTC Normal Pod report-span-8b52g.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:22:16 | collector-otlp-allinone-http | 2024-12-09 07:22:04 +0000 UTC Normal Pod report-span-8b52g.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 1.801s (1.801s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:22:16 | collector-otlp-allinone-http | 2024-12-09 07:22:04 +0000 UTC Normal Pod report-span-8b52g.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:22:16 | collector-otlp-allinone-http | 2024-12-09 07:22:04 +0000 UTC Normal Pod report-span-8b52g.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:22:16 | collector-otlp-allinone-http | 2024-12-09 07:22:05 +0000 UTC Normal Pod check-span-bk5nx.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 2.179s (2.179s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:22:16 | collector-otlp-allinone-http | 2024-12-09 07:22:05 +0000 UTC Normal Pod check-span-bk5nx.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:22:16 | collector-otlp-allinone-http | 2024-12-09 07:22:05 +0000 UTC Normal Pod check-span-bk5nx.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:22:16 | collector-otlp-allinone-http | 2024-12-09 07:22:16 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:22:16 | collector-otlp-allinone-http | Deleting namespace: kuttl-test-wanted-emu === CONT kuttl/harness/set-custom-img logger.go:42: 07:22:29 | set-custom-img | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:22:29 | set-custom-img | Ignoring check-collector-img.sh as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:22:29 | set-custom-img | Creating namespace: kuttl-test-relative-badger logger.go:42: 07:22:29 | set-custom-img/1-install | starting test step 1-install logger.go:42: 07:22:29 | set-custom-img/1-install | Jaeger:kuttl-test-relative-badger/my-jaeger created logger.go:42: 07:23:05 | set-custom-img/1-install | test step completed 1-install logger.go:42: 07:23:05 | set-custom-img/2-install | starting test step 2-install logger.go:42: 07:23:05 | set-custom-img/2-install | Jaeger:kuttl-test-relative-badger/my-jaeger updated logger.go:42: 07:23:05 | set-custom-img/2-install | test step completed 2-install logger.go:42: 07:23:05 | set-custom-img/3-check-image | starting test step 3-check-image logger.go:42: 07:23:05 | set-custom-img/3-check-image | running command: [sh -c ./check-collector-img.sh] logger.go:42: 07:23:06 | set-custom-img/3-check-image | Collector image missmatch. Expected: test. Has: registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:44686d560aa27cae8ff8693f88c4cb6e2edf1737010ec1f80709cb42250b729d logger.go:42: 07:23:11 | set-custom-img/3-check-image | Collector image asserted properly! logger.go:42: 07:23:11 | set-custom-img/3-check-image | test step completed 3-check-image logger.go:42: 07:23:11 | set-custom-img | set-custom-img events from ns kuttl-test-relative-badger: logger.go:42: 07:23:11 | set-custom-img | 2024-12-09 07:22:36 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestrelativebadgermyjaeger-1-999cc5c5c SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestrelativebadgermyjaeger-1-999cc5pmd7q replicaset-controller logger.go:42: 07:23:11 | set-custom-img | 2024-12-09 07:22:36 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestrelativebadgermyjaeger-1-999cc5pmd7q Binding Scheduled Successfully assigned kuttl-test-relative-badger/elasticsearch-cdm-kuttltestrelativebadgermyjaeger-1-999cc5pmd7q to ip-10-0-95-250.ec2.internal default-scheduler logger.go:42: 07:23:11 | set-custom-img | 2024-12-09 07:22:36 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestrelativebadgermyjaeger-1-999cc5pmd7q AddedInterface Add eth0 [10.131.0.28/23] from ovn-kubernetes multus logger.go:42: 07:23:11 | set-custom-img | 2024-12-09 07:22:36 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestrelativebadgermyjaeger-1-999cc5pmd7q.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:8e4fbea4983cd58352349ca291383169b286bc166fad95a87807552ca43335e6" already present on machine kubelet logger.go:42: 07:23:11 | set-custom-img | 2024-12-09 07:22:36 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestrelativebadgermyjaeger-1-999cc5pmd7q.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:23:11 | set-custom-img | 2024-12-09 07:22:36 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestrelativebadgermyjaeger-1-999cc5pmd7q.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:23:11 | set-custom-img | 2024-12-09 07:22:36 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestrelativebadgermyjaeger-1-999cc5pmd7q.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:d68824b0b2c84db8e33edf9ab344eb684c4a7ebd7ef162bbc309043adcb28e6b" already present on machine kubelet logger.go:42: 07:23:11 | set-custom-img | 2024-12-09 07:22:36 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestrelativebadgermyjaeger-1-999cc5pmd7q.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:23:11 | set-custom-img | 2024-12-09 07:22:36 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestrelativebadgermyjaeger-1-999cc5pmd7q.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:23:11 | set-custom-img | 2024-12-09 07:22:36 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestrelativebadgermyjaeger-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestrelativebadgermyjaeger-1-999cc5c5c to 1 deployment-controller logger.go:42: 07:23:11 | set-custom-img | 2024-12-09 07:22:46 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestrelativebadgermyjaeger-1-999cc5pmd7q.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 07:23:11 | set-custom-img | 2024-12-09 07:22:51 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestrelativebadgermyjaeger-1-999cc5pmd7q.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:23:11 | set-custom-img | 2024-12-09 07:23:02 +0000 UTC Normal Pod my-jaeger-collector-7cdf765fd-q6tqb Binding Scheduled Successfully assigned kuttl-test-relative-badger/my-jaeger-collector-7cdf765fd-q6tqb to ip-10-0-79-227.ec2.internal default-scheduler logger.go:42: 07:23:11 | set-custom-img | 2024-12-09 07:23:02 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-7cdf765fd SuccessfulCreate Created pod: my-jaeger-collector-7cdf765fd-q6tqb replicaset-controller logger.go:42: 07:23:11 | set-custom-img | 2024-12-09 07:23:02 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-7cdf765fd to 1 deployment-controller logger.go:42: 07:23:11 | set-custom-img | 2024-12-09 07:23:03 +0000 UTC Normal Pod my-jaeger-collector-7cdf765fd-q6tqb AddedInterface Add eth0 [10.129.2.44/23] from ovn-kubernetes multus logger.go:42: 07:23:11 | set-custom-img | 2024-12-09 07:23:03 +0000 UTC Normal Pod my-jaeger-collector-7cdf765fd-q6tqb.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:44686d560aa27cae8ff8693f88c4cb6e2edf1737010ec1f80709cb42250b729d" already present on machine kubelet logger.go:42: 07:23:11 | set-custom-img | 2024-12-09 07:23:03 +0000 UTC Normal Pod my-jaeger-collector-7cdf765fd-q6tqb.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:23:11 | set-custom-img | 2024-12-09 07:23:03 +0000 UTC Normal Pod my-jaeger-collector-7cdf765fd-q6tqb.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:23:11 | set-custom-img | 2024-12-09 07:23:03 +0000 UTC Normal Pod my-jaeger-query-dc998d4fd-rzxrh Binding Scheduled Successfully assigned kuttl-test-relative-badger/my-jaeger-query-dc998d4fd-rzxrh to ip-10-0-79-227.ec2.internal default-scheduler logger.go:42: 07:23:11 | set-custom-img | 2024-12-09 07:23:03 +0000 UTC Normal Pod my-jaeger-query-dc998d4fd-rzxrh AddedInterface Add eth0 [10.129.2.45/23] from ovn-kubernetes multus logger.go:42: 07:23:11 | set-custom-img | 2024-12-09 07:23:03 +0000 UTC Normal Pod my-jaeger-query-dc998d4fd-rzxrh.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:24a83160aa930e1b72c2a2442a33b28af2b06c24b058e09afcd0d495a8066d6d" already present on machine kubelet logger.go:42: 07:23:11 | set-custom-img | 2024-12-09 07:23:03 +0000 UTC Normal Pod my-jaeger-query-dc998d4fd-rzxrh.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:23:11 | set-custom-img | 2024-12-09 07:23:03 +0000 UTC Normal Pod my-jaeger-query-dc998d4fd-rzxrh.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:23:11 | set-custom-img | 2024-12-09 07:23:03 +0000 UTC Normal Pod my-jaeger-query-dc998d4fd-rzxrh.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:23:11 | set-custom-img | 2024-12-09 07:23:03 +0000 UTC Normal Pod my-jaeger-query-dc998d4fd-rzxrh.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:23:11 | set-custom-img | 2024-12-09 07:23:03 +0000 UTC Normal Pod my-jaeger-query-dc998d4fd-rzxrh.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:23:11 | set-custom-img | 2024-12-09 07:23:03 +0000 UTC Normal Pod my-jaeger-query-dc998d4fd-rzxrh.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" already present on machine kubelet logger.go:42: 07:23:11 | set-custom-img | 2024-12-09 07:23:03 +0000 UTC Normal Pod my-jaeger-query-dc998d4fd-rzxrh.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:23:11 | set-custom-img | 2024-12-09 07:23:03 +0000 UTC Normal Pod my-jaeger-query-dc998d4fd-rzxrh.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:23:11 | set-custom-img | 2024-12-09 07:23:03 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-dc998d4fd SuccessfulCreate Created pod: my-jaeger-query-dc998d4fd-rzxrh replicaset-controller logger.go:42: 07:23:11 | set-custom-img | 2024-12-09 07:23:03 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-dc998d4fd to 1 deployment-controller logger.go:42: 07:23:11 | set-custom-img | 2024-12-09 07:23:09 +0000 UTC Normal Pod my-jaeger-collector-7cdf765fd-q6tqb.spec.containers{jaeger-collector} Killing Stopping container jaeger-collector kubelet logger.go:42: 07:23:11 | set-custom-img | 2024-12-09 07:23:09 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-7cdf765fd SuccessfulDelete Deleted pod: my-jaeger-collector-7cdf765fd-q6tqb replicaset-controller logger.go:42: 07:23:11 | set-custom-img | 2024-12-09 07:23:09 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled down replica set my-jaeger-collector-7cdf765fd to 0 from 1 deployment-controller logger.go:42: 07:23:11 | set-custom-img | 2024-12-09 07:23:10 +0000 UTC Normal Pod my-jaeger-collector-5cdbbb585f-hlml7 Binding Scheduled Successfully assigned kuttl-test-relative-badger/my-jaeger-collector-5cdbbb585f-hlml7 to ip-10-0-39-223.ec2.internal default-scheduler logger.go:42: 07:23:11 | set-custom-img | 2024-12-09 07:23:10 +0000 UTC Normal Pod my-jaeger-collector-5cdbbb585f-hlml7 AddedInterface Add eth0 [10.128.2.27/23] from ovn-kubernetes multus logger.go:42: 07:23:11 | set-custom-img | 2024-12-09 07:23:10 +0000 UTC Normal Pod my-jaeger-collector-5cdbbb585f-hlml7.spec.containers{jaeger-collector} Pulling Pulling image "test" kubelet logger.go:42: 07:23:11 | set-custom-img | 2024-12-09 07:23:10 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-5cdbbb585f SuccessfulCreate Created pod: my-jaeger-collector-5cdbbb585f-hlml7 replicaset-controller logger.go:42: 07:23:11 | set-custom-img | 2024-12-09 07:23:10 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-5cdbbb585f to 1 deployment-controller logger.go:42: 07:23:11 | set-custom-img | 2024-12-09 07:23:11 +0000 UTC Warning Pod my-jaeger-collector-5cdbbb585f-hlml7.spec.containers{jaeger-collector} Failed Failed to pull image "test": initializing source docker://test:latest: reading manifest latest in docker.io/library/test: requested access to the resource is denied kubelet logger.go:42: 07:23:11 | set-custom-img | 2024-12-09 07:23:11 +0000 UTC Warning Pod my-jaeger-collector-5cdbbb585f-hlml7.spec.containers{jaeger-collector} Failed Error: ErrImagePull kubelet logger.go:42: 07:23:11 | set-custom-img | 2024-12-09 07:23:11 +0000 UTC Normal Pod my-jaeger-collector-5cdbbb585f-hlml7.spec.containers{jaeger-collector} BackOff Back-off pulling image "test" kubelet logger.go:42: 07:23:11 | set-custom-img | Deleting namespace: kuttl-test-relative-badger === CONT kuttl/harness/collector-otlp-production-http logger.go:42: 07:23:17 | collector-otlp-production-http | Creating namespace: kuttl-test-viable-swift logger.go:42: 07:23:17 | collector-otlp-production-http/1-install | starting test step 1-install logger.go:42: 07:23:18 | collector-otlp-production-http/1-install | Jaeger:kuttl-test-viable-swift/my-jaeger created logger.go:42: 07:23:54 | collector-otlp-production-http/1-install | test step completed 1-install logger.go:42: 07:23:54 | collector-otlp-production-http/2-smoke-test | starting test step 2-smoke-test logger.go:42: 07:23:54 | collector-otlp-production-http/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:23:55 | collector-otlp-production-http/2-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:24:03 | collector-otlp-production-http/2-smoke-test | running command: [sh -c REPORTING_PROTOCOL=http ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4318 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/otlp-smoke-test.yaml.template -o otlp-smoke-test-job.yaml] logger.go:42: 07:24:04 | collector-otlp-production-http/2-smoke-test | running command: [sh -c kubectl create -f otlp-smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:24:04 | collector-otlp-production-http/2-smoke-test | job.batch/report-span created logger.go:42: 07:24:04 | collector-otlp-production-http/2-smoke-test | job.batch/check-span created logger.go:42: 07:24:16 | collector-otlp-production-http/2-smoke-test | test step completed 2-smoke-test logger.go:42: 07:24:16 | collector-otlp-production-http | collector-otlp-production-http events from ns kuttl-test-viable-swift: logger.go:42: 07:24:16 | collector-otlp-production-http | 2024-12-09 07:23:24 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestviableswiftmyjaeger-1-7959bcbb9 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestviableswiftmyjaeger-1-7959bcbb9qcvdt replicaset-controller logger.go:42: 07:24:16 | collector-otlp-production-http | 2024-12-09 07:23:24 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestviableswiftmyjaeger-1-7959bcbb9qcvdt Binding Scheduled Successfully assigned kuttl-test-viable-swift/elasticsearch-cdm-kuttltestviableswiftmyjaeger-1-7959bcbb9qcvdt to ip-10-0-95-250.ec2.internal default-scheduler logger.go:42: 07:24:16 | collector-otlp-production-http | 2024-12-09 07:23:24 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestviableswiftmyjaeger-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestviableswiftmyjaeger-1-7959bcbb9 to 1 deployment-controller logger.go:42: 07:24:16 | collector-otlp-production-http | 2024-12-09 07:23:25 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestviableswiftmyjaeger-1-7959bcbb9qcvdt AddedInterface Add eth0 [10.131.0.29/23] from ovn-kubernetes multus logger.go:42: 07:24:16 | collector-otlp-production-http | 2024-12-09 07:23:25 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestviableswiftmyjaeger-1-7959bcbb9qcvdt.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:8e4fbea4983cd58352349ca291383169b286bc166fad95a87807552ca43335e6" already present on machine kubelet logger.go:42: 07:24:16 | collector-otlp-production-http | 2024-12-09 07:23:25 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestviableswiftmyjaeger-1-7959bcbb9qcvdt.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:24:16 | collector-otlp-production-http | 2024-12-09 07:23:25 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestviableswiftmyjaeger-1-7959bcbb9qcvdt.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:24:16 | collector-otlp-production-http | 2024-12-09 07:23:25 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestviableswiftmyjaeger-1-7959bcbb9qcvdt.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:d68824b0b2c84db8e33edf9ab344eb684c4a7ebd7ef162bbc309043adcb28e6b" already present on machine kubelet logger.go:42: 07:24:16 | collector-otlp-production-http | 2024-12-09 07:23:25 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestviableswiftmyjaeger-1-7959bcbb9qcvdt.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:24:16 | collector-otlp-production-http | 2024-12-09 07:23:25 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestviableswiftmyjaeger-1-7959bcbb9qcvdt.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:24:16 | collector-otlp-production-http | 2024-12-09 07:23:35 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestviableswiftmyjaeger-1-7959bcbb9qcvdt.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 07:24:16 | collector-otlp-production-http | 2024-12-09 07:23:40 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestviableswiftmyjaeger-1-7959bcbb9qcvdt.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:24:16 | collector-otlp-production-http | 2024-12-09 07:23:51 +0000 UTC Normal Pod my-jaeger-collector-7f77965c5c-jnqnh Binding Scheduled Successfully assigned kuttl-test-viable-swift/my-jaeger-collector-7f77965c5c-jnqnh to ip-10-0-79-227.ec2.internal default-scheduler logger.go:42: 07:24:16 | collector-otlp-production-http | 2024-12-09 07:23:51 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-7f77965c5c SuccessfulCreate Created pod: my-jaeger-collector-7f77965c5c-jnqnh replicaset-controller logger.go:42: 07:24:16 | collector-otlp-production-http | 2024-12-09 07:23:51 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-7f77965c5c to 1 deployment-controller logger.go:42: 07:24:16 | collector-otlp-production-http | 2024-12-09 07:23:51 +0000 UTC Normal Pod my-jaeger-query-7fc7b76dd7-4s9pc Binding Scheduled Successfully assigned kuttl-test-viable-swift/my-jaeger-query-7fc7b76dd7-4s9pc to ip-10-0-79-227.ec2.internal default-scheduler logger.go:42: 07:24:16 | collector-otlp-production-http | 2024-12-09 07:23:51 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-7fc7b76dd7 SuccessfulCreate Created pod: my-jaeger-query-7fc7b76dd7-4s9pc replicaset-controller logger.go:42: 07:24:16 | collector-otlp-production-http | 2024-12-09 07:23:51 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-7fc7b76dd7 to 1 deployment-controller logger.go:42: 07:24:16 | collector-otlp-production-http | 2024-12-09 07:23:52 +0000 UTC Normal Pod my-jaeger-collector-7f77965c5c-jnqnh AddedInterface Add eth0 [10.129.2.46/23] from ovn-kubernetes multus logger.go:42: 07:24:16 | collector-otlp-production-http | 2024-12-09 07:23:52 +0000 UTC Normal Pod my-jaeger-collector-7f77965c5c-jnqnh.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:44686d560aa27cae8ff8693f88c4cb6e2edf1737010ec1f80709cb42250b729d" already present on machine kubelet logger.go:42: 07:24:16 | collector-otlp-production-http | 2024-12-09 07:23:52 +0000 UTC Normal Pod my-jaeger-collector-7f77965c5c-jnqnh.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:24:16 | collector-otlp-production-http | 2024-12-09 07:23:52 +0000 UTC Normal Pod my-jaeger-collector-7f77965c5c-jnqnh.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:24:16 | collector-otlp-production-http | 2024-12-09 07:23:52 +0000 UTC Normal Pod my-jaeger-query-7fc7b76dd7-4s9pc AddedInterface Add eth0 [10.129.2.47/23] from ovn-kubernetes multus logger.go:42: 07:24:16 | collector-otlp-production-http | 2024-12-09 07:23:52 +0000 UTC Normal Pod my-jaeger-query-7fc7b76dd7-4s9pc.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:24a83160aa930e1b72c2a2442a33b28af2b06c24b058e09afcd0d495a8066d6d" already present on machine kubelet logger.go:42: 07:24:16 | collector-otlp-production-http | 2024-12-09 07:23:52 +0000 UTC Normal Pod my-jaeger-query-7fc7b76dd7-4s9pc.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:24:16 | collector-otlp-production-http | 2024-12-09 07:23:52 +0000 UTC Normal Pod my-jaeger-query-7fc7b76dd7-4s9pc.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:24:16 | collector-otlp-production-http | 2024-12-09 07:23:52 +0000 UTC Normal Pod my-jaeger-query-7fc7b76dd7-4s9pc.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:24:16 | collector-otlp-production-http | 2024-12-09 07:23:52 +0000 UTC Normal Pod my-jaeger-query-7fc7b76dd7-4s9pc.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:24:16 | collector-otlp-production-http | 2024-12-09 07:23:52 +0000 UTC Normal Pod my-jaeger-query-7fc7b76dd7-4s9pc.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:24:16 | collector-otlp-production-http | 2024-12-09 07:23:52 +0000 UTC Normal Pod my-jaeger-query-7fc7b76dd7-4s9pc.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" already present on machine kubelet logger.go:42: 07:24:16 | collector-otlp-production-http | 2024-12-09 07:23:52 +0000 UTC Normal Pod my-jaeger-query-7fc7b76dd7-4s9pc.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:24:16 | collector-otlp-production-http | 2024-12-09 07:23:52 +0000 UTC Normal Pod my-jaeger-query-7fc7b76dd7-4s9pc.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:24:16 | collector-otlp-production-http | 2024-12-09 07:23:58 +0000 UTC Normal Pod my-jaeger-query-7fc7b76dd7-4s9pc.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:24:16 | collector-otlp-production-http | 2024-12-09 07:23:58 +0000 UTC Normal Pod my-jaeger-query-7fc7b76dd7-4s9pc.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:24:16 | collector-otlp-production-http | 2024-12-09 07:23:58 +0000 UTC Normal Pod my-jaeger-query-7fc7b76dd7-4s9pc.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:24:16 | collector-otlp-production-http | 2024-12-09 07:23:58 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-7fc7b76dd7 SuccessfulDelete Deleted pod: my-jaeger-query-7fc7b76dd7-4s9pc replicaset-controller logger.go:42: 07:24:16 | collector-otlp-production-http | 2024-12-09 07:23:58 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled down replica set my-jaeger-query-7fc7b76dd7 to 0 from 1 deployment-controller logger.go:42: 07:24:16 | collector-otlp-production-http | 2024-12-09 07:23:59 +0000 UTC Normal Pod my-jaeger-query-7574b555bb-8l8jk Binding Scheduled Successfully assigned kuttl-test-viable-swift/my-jaeger-query-7574b555bb-8l8jk to ip-10-0-79-227.ec2.internal default-scheduler logger.go:42: 07:24:16 | collector-otlp-production-http | 2024-12-09 07:23:59 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-7574b555bb SuccessfulCreate Created pod: my-jaeger-query-7574b555bb-8l8jk replicaset-controller logger.go:42: 07:24:16 | collector-otlp-production-http | 2024-12-09 07:23:59 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-7574b555bb to 1 deployment-controller logger.go:42: 07:24:16 | collector-otlp-production-http | 2024-12-09 07:24:00 +0000 UTC Normal Pod my-jaeger-query-7574b555bb-8l8jk AddedInterface Add eth0 [10.129.2.48/23] from ovn-kubernetes multus logger.go:42: 07:24:16 | collector-otlp-production-http | 2024-12-09 07:24:00 +0000 UTC Normal Pod my-jaeger-query-7574b555bb-8l8jk.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:24a83160aa930e1b72c2a2442a33b28af2b06c24b058e09afcd0d495a8066d6d" already present on machine kubelet logger.go:42: 07:24:16 | collector-otlp-production-http | 2024-12-09 07:24:00 +0000 UTC Normal Pod my-jaeger-query-7574b555bb-8l8jk.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:24:16 | collector-otlp-production-http | 2024-12-09 07:24:00 +0000 UTC Normal Pod my-jaeger-query-7574b555bb-8l8jk.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:24:16 | collector-otlp-production-http | 2024-12-09 07:24:00 +0000 UTC Normal Pod my-jaeger-query-7574b555bb-8l8jk.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:24:16 | collector-otlp-production-http | 2024-12-09 07:24:00 +0000 UTC Normal Pod my-jaeger-query-7574b555bb-8l8jk.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:24:16 | collector-otlp-production-http | 2024-12-09 07:24:00 +0000 UTC Normal Pod my-jaeger-query-7574b555bb-8l8jk.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:24:16 | collector-otlp-production-http | 2024-12-09 07:24:00 +0000 UTC Normal Pod my-jaeger-query-7574b555bb-8l8jk.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" already present on machine kubelet logger.go:42: 07:24:16 | collector-otlp-production-http | 2024-12-09 07:24:00 +0000 UTC Normal Pod my-jaeger-query-7574b555bb-8l8jk.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:24:16 | collector-otlp-production-http | 2024-12-09 07:24:00 +0000 UTC Normal Pod my-jaeger-query-7574b555bb-8l8jk.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:24:16 | collector-otlp-production-http | 2024-12-09 07:24:04 +0000 UTC Normal Pod check-span-fldg2 Binding Scheduled Successfully assigned kuttl-test-viable-swift/check-span-fldg2 to ip-10-0-39-223.ec2.internal default-scheduler logger.go:42: 07:24:16 | collector-otlp-production-http | 2024-12-09 07:24:04 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-fldg2 job-controller logger.go:42: 07:24:16 | collector-otlp-production-http | 2024-12-09 07:24:04 +0000 UTC Normal Pod report-span-smwpw Binding Scheduled Successfully assigned kuttl-test-viable-swift/report-span-smwpw to ip-10-0-39-223.ec2.internal default-scheduler logger.go:42: 07:24:16 | collector-otlp-production-http | 2024-12-09 07:24:04 +0000 UTC Normal Pod report-span-smwpw AddedInterface Add eth0 [10.128.2.28/23] from ovn-kubernetes multus logger.go:42: 07:24:16 | collector-otlp-production-http | 2024-12-09 07:24:04 +0000 UTC Normal Pod report-span-smwpw.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:24:16 | collector-otlp-production-http | 2024-12-09 07:24:04 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-smwpw job-controller logger.go:42: 07:24:16 | collector-otlp-production-http | 2024-12-09 07:24:05 +0000 UTC Normal Pod check-span-fldg2 AddedInterface Add eth0 [10.128.2.29/23] from ovn-kubernetes multus logger.go:42: 07:24:16 | collector-otlp-production-http | 2024-12-09 07:24:05 +0000 UTC Normal Pod check-span-fldg2.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:24:16 | collector-otlp-production-http | 2024-12-09 07:24:05 +0000 UTC Normal Pod check-span-fldg2.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 757ms (757ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:24:16 | collector-otlp-production-http | 2024-12-09 07:24:05 +0000 UTC Normal Pod check-span-fldg2.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:24:16 | collector-otlp-production-http | 2024-12-09 07:24:05 +0000 UTC Normal Pod check-span-fldg2.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:24:16 | collector-otlp-production-http | 2024-12-09 07:24:06 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:24:16 | collector-otlp-production-http | 2024-12-09 07:24:06 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:24:16 | collector-otlp-production-http | 2024-12-09 07:24:06 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:24:16 | collector-otlp-production-http | 2024-12-09 07:24:06 +0000 UTC Normal Pod report-span-smwpw.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 1.849s (1.849s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:24:16 | collector-otlp-production-http | 2024-12-09 07:24:06 +0000 UTC Normal Pod report-span-smwpw.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:24:16 | collector-otlp-production-http | 2024-12-09 07:24:06 +0000 UTC Normal Pod report-span-smwpw.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:24:16 | collector-otlp-production-http | 2024-12-09 07:24:16 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:24:16 | collector-otlp-production-http | Deleting namespace: kuttl-test-viable-swift === CONT kuttl/harness/collector-otlp-production-grpc logger.go:42: 07:24:28 | collector-otlp-production-grpc | Creating namespace: kuttl-test-merry-manatee logger.go:42: 07:24:28 | collector-otlp-production-grpc/1-install | starting test step 1-install logger.go:42: 07:24:29 | collector-otlp-production-grpc/1-install | Jaeger:kuttl-test-merry-manatee/my-jaeger created logger.go:42: 07:25:04 | collector-otlp-production-grpc/1-install | test step completed 1-install logger.go:42: 07:25:04 | collector-otlp-production-grpc/2-smoke-test | starting test step 2-smoke-test logger.go:42: 07:25:04 | collector-otlp-production-grpc/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:25:05 | collector-otlp-production-grpc/2-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:25:11 | collector-otlp-production-grpc/2-smoke-test | running command: [sh -c REPORTING_PROTOCOL=grpc ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/otlp-smoke-test.yaml.template -o otlp-smoke-test-job.yaml] logger.go:42: 07:25:12 | collector-otlp-production-grpc/2-smoke-test | running command: [sh -c kubectl create -f otlp-smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:25:12 | collector-otlp-production-grpc/2-smoke-test | job.batch/report-span created logger.go:42: 07:25:12 | collector-otlp-production-grpc/2-smoke-test | job.batch/check-span created logger.go:42: 07:25:32 | collector-otlp-production-grpc/2-smoke-test | test step completed 2-smoke-test logger.go:42: 07:25:33 | collector-otlp-production-grpc | collector-otlp-production-grpc events from ns kuttl-test-merry-manatee: logger.go:42: 07:25:33 | collector-otlp-production-grpc | 2024-12-09 07:24:34 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestmerrymanateemyjaeger-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestmerrymanateemyjaeger-1-777949c86c to 1 deployment-controller logger.go:42: 07:25:33 | collector-otlp-production-grpc | 2024-12-09 07:24:35 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmerrymanateemyjaeger-1-777949c84vhz7 Binding Scheduled Successfully assigned kuttl-test-merry-manatee/elasticsearch-cdm-kuttltestmerrymanateemyjaeger-1-777949c84vhz7 to ip-10-0-95-250.ec2.internal default-scheduler logger.go:42: 07:25:33 | collector-otlp-production-grpc | 2024-12-09 07:24:35 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmerrymanateemyjaeger-1-777949c84vhz7 AddedInterface Add eth0 [10.131.0.30/23] from ovn-kubernetes multus logger.go:42: 07:25:33 | collector-otlp-production-grpc | 2024-12-09 07:24:35 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmerrymanateemyjaeger-1-777949c84vhz7.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:8e4fbea4983cd58352349ca291383169b286bc166fad95a87807552ca43335e6" already present on machine kubelet logger.go:42: 07:25:33 | collector-otlp-production-grpc | 2024-12-09 07:24:35 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmerrymanateemyjaeger-1-777949c84vhz7.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:25:33 | collector-otlp-production-grpc | 2024-12-09 07:24:35 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmerrymanateemyjaeger-1-777949c84vhz7.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:25:33 | collector-otlp-production-grpc | 2024-12-09 07:24:35 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmerrymanateemyjaeger-1-777949c84vhz7.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:d68824b0b2c84db8e33edf9ab344eb684c4a7ebd7ef162bbc309043adcb28e6b" already present on machine kubelet logger.go:42: 07:25:33 | collector-otlp-production-grpc | 2024-12-09 07:24:35 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmerrymanateemyjaeger-1-777949c84vhz7.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:25:33 | collector-otlp-production-grpc | 2024-12-09 07:24:35 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmerrymanateemyjaeger-1-777949c84vhz7.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:25:33 | collector-otlp-production-grpc | 2024-12-09 07:24:35 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestmerrymanateemyjaeger-1-777949c86c SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestmerrymanateemyjaeger-1-777949c84vhz7 replicaset-controller logger.go:42: 07:25:33 | collector-otlp-production-grpc | 2024-12-09 07:24:45 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestmerrymanateemyjaeger-1-777949c84vhz7.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 07:25:33 | collector-otlp-production-grpc | 2024-12-09 07:24:50 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestmerrymanateemyjaeger-1-777949c84vhz7.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:25:33 | collector-otlp-production-grpc | 2024-12-09 07:25:01 +0000 UTC Normal Pod my-jaeger-collector-64fcd4f4c9-b5q8k Binding Scheduled Successfully assigned kuttl-test-merry-manatee/my-jaeger-collector-64fcd4f4c9-b5q8k to ip-10-0-79-227.ec2.internal default-scheduler logger.go:42: 07:25:33 | collector-otlp-production-grpc | 2024-12-09 07:25:01 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-64fcd4f4c9 SuccessfulCreate Created pod: my-jaeger-collector-64fcd4f4c9-b5q8k replicaset-controller logger.go:42: 07:25:33 | collector-otlp-production-grpc | 2024-12-09 07:25:01 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-64fcd4f4c9 to 1 deployment-controller logger.go:42: 07:25:33 | collector-otlp-production-grpc | 2024-12-09 07:25:01 +0000 UTC Normal Pod my-jaeger-query-8679f8b77d-2wvrw Binding Scheduled Successfully assigned kuttl-test-merry-manatee/my-jaeger-query-8679f8b77d-2wvrw to ip-10-0-79-227.ec2.internal default-scheduler logger.go:42: 07:25:33 | collector-otlp-production-grpc | 2024-12-09 07:25:01 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-8679f8b77d SuccessfulCreate Created pod: my-jaeger-query-8679f8b77d-2wvrw replicaset-controller logger.go:42: 07:25:33 | collector-otlp-production-grpc | 2024-12-09 07:25:01 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-8679f8b77d to 1 deployment-controller logger.go:42: 07:25:33 | collector-otlp-production-grpc | 2024-12-09 07:25:02 +0000 UTC Normal Pod my-jaeger-collector-64fcd4f4c9-b5q8k AddedInterface Add eth0 [10.129.2.49/23] from ovn-kubernetes multus logger.go:42: 07:25:33 | collector-otlp-production-grpc | 2024-12-09 07:25:02 +0000 UTC Normal Pod my-jaeger-collector-64fcd4f4c9-b5q8k.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:44686d560aa27cae8ff8693f88c4cb6e2edf1737010ec1f80709cb42250b729d" already present on machine kubelet logger.go:42: 07:25:33 | collector-otlp-production-grpc | 2024-12-09 07:25:02 +0000 UTC Normal Pod my-jaeger-collector-64fcd4f4c9-b5q8k.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:25:33 | collector-otlp-production-grpc | 2024-12-09 07:25:02 +0000 UTC Normal Pod my-jaeger-collector-64fcd4f4c9-b5q8k.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:25:33 | collector-otlp-production-grpc | 2024-12-09 07:25:02 +0000 UTC Normal Pod my-jaeger-query-8679f8b77d-2wvrw AddedInterface Add eth0 [10.129.2.50/23] from ovn-kubernetes multus logger.go:42: 07:25:33 | collector-otlp-production-grpc | 2024-12-09 07:25:02 +0000 UTC Normal Pod my-jaeger-query-8679f8b77d-2wvrw.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:24a83160aa930e1b72c2a2442a33b28af2b06c24b058e09afcd0d495a8066d6d" already present on machine kubelet logger.go:42: 07:25:33 | collector-otlp-production-grpc | 2024-12-09 07:25:02 +0000 UTC Normal Pod my-jaeger-query-8679f8b77d-2wvrw.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:25:33 | collector-otlp-production-grpc | 2024-12-09 07:25:02 +0000 UTC Normal Pod my-jaeger-query-8679f8b77d-2wvrw.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:25:33 | collector-otlp-production-grpc | 2024-12-09 07:25:02 +0000 UTC Normal Pod my-jaeger-query-8679f8b77d-2wvrw.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:25:33 | collector-otlp-production-grpc | 2024-12-09 07:25:02 +0000 UTC Normal Pod my-jaeger-query-8679f8b77d-2wvrw.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:25:33 | collector-otlp-production-grpc | 2024-12-09 07:25:02 +0000 UTC Normal Pod my-jaeger-query-8679f8b77d-2wvrw.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:25:33 | collector-otlp-production-grpc | 2024-12-09 07:25:02 +0000 UTC Normal Pod my-jaeger-query-8679f8b77d-2wvrw.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" already present on machine kubelet logger.go:42: 07:25:33 | collector-otlp-production-grpc | 2024-12-09 07:25:02 +0000 UTC Normal Pod my-jaeger-query-8679f8b77d-2wvrw.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:25:33 | collector-otlp-production-grpc | 2024-12-09 07:25:02 +0000 UTC Normal Pod my-jaeger-query-8679f8b77d-2wvrw.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:25:33 | collector-otlp-production-grpc | 2024-12-09 07:25:07 +0000 UTC Normal Pod my-jaeger-query-68b69bbfcf-rw4r8 Binding Scheduled Successfully assigned kuttl-test-merry-manatee/my-jaeger-query-68b69bbfcf-rw4r8 to ip-10-0-79-227.ec2.internal default-scheduler logger.go:42: 07:25:33 | collector-otlp-production-grpc | 2024-12-09 07:25:07 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-68b69bbfcf SuccessfulCreate Created pod: my-jaeger-query-68b69bbfcf-rw4r8 replicaset-controller logger.go:42: 07:25:33 | collector-otlp-production-grpc | 2024-12-09 07:25:07 +0000 UTC Normal Pod my-jaeger-query-8679f8b77d-2wvrw.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:25:33 | collector-otlp-production-grpc | 2024-12-09 07:25:07 +0000 UTC Normal Pod my-jaeger-query-8679f8b77d-2wvrw.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:25:33 | collector-otlp-production-grpc | 2024-12-09 07:25:07 +0000 UTC Normal Pod my-jaeger-query-8679f8b77d-2wvrw.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:25:33 | collector-otlp-production-grpc | 2024-12-09 07:25:07 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-8679f8b77d SuccessfulDelete Deleted pod: my-jaeger-query-8679f8b77d-2wvrw replicaset-controller logger.go:42: 07:25:33 | collector-otlp-production-grpc | 2024-12-09 07:25:07 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled down replica set my-jaeger-query-8679f8b77d to 0 from 1 deployment-controller logger.go:42: 07:25:33 | collector-otlp-production-grpc | 2024-12-09 07:25:07 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-68b69bbfcf to 1 deployment-controller logger.go:42: 07:25:33 | collector-otlp-production-grpc | 2024-12-09 07:25:08 +0000 UTC Normal Pod my-jaeger-query-68b69bbfcf-rw4r8 AddedInterface Add eth0 [10.129.2.51/23] from ovn-kubernetes multus logger.go:42: 07:25:33 | collector-otlp-production-grpc | 2024-12-09 07:25:08 +0000 UTC Normal Pod my-jaeger-query-68b69bbfcf-rw4r8.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:24a83160aa930e1b72c2a2442a33b28af2b06c24b058e09afcd0d495a8066d6d" already present on machine kubelet logger.go:42: 07:25:33 | collector-otlp-production-grpc | 2024-12-09 07:25:08 +0000 UTC Normal Pod my-jaeger-query-68b69bbfcf-rw4r8.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:25:33 | collector-otlp-production-grpc | 2024-12-09 07:25:08 +0000 UTC Normal Pod my-jaeger-query-68b69bbfcf-rw4r8.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:25:33 | collector-otlp-production-grpc | 2024-12-09 07:25:08 +0000 UTC Normal Pod my-jaeger-query-68b69bbfcf-rw4r8.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:25:33 | collector-otlp-production-grpc | 2024-12-09 07:25:08 +0000 UTC Normal Pod my-jaeger-query-68b69bbfcf-rw4r8.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:25:33 | collector-otlp-production-grpc | 2024-12-09 07:25:08 +0000 UTC Normal Pod my-jaeger-query-68b69bbfcf-rw4r8.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:25:33 | collector-otlp-production-grpc | 2024-12-09 07:25:08 +0000 UTC Normal Pod my-jaeger-query-68b69bbfcf-rw4r8.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" already present on machine kubelet logger.go:42: 07:25:33 | collector-otlp-production-grpc | 2024-12-09 07:25:08 +0000 UTC Normal Pod my-jaeger-query-68b69bbfcf-rw4r8.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:25:33 | collector-otlp-production-grpc | 2024-12-09 07:25:08 +0000 UTC Normal Pod my-jaeger-query-68b69bbfcf-rw4r8.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:25:33 | collector-otlp-production-grpc | 2024-12-09 07:25:12 +0000 UTC Normal Pod check-span-s78vd Binding Scheduled Successfully assigned kuttl-test-merry-manatee/check-span-s78vd to ip-10-0-39-223.ec2.internal default-scheduler logger.go:42: 07:25:33 | collector-otlp-production-grpc | 2024-12-09 07:25:12 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-s78vd job-controller logger.go:42: 07:25:33 | collector-otlp-production-grpc | 2024-12-09 07:25:12 +0000 UTC Normal Pod report-span-bwr9l Binding Scheduled Successfully assigned kuttl-test-merry-manatee/report-span-bwr9l to ip-10-0-39-223.ec2.internal default-scheduler logger.go:42: 07:25:33 | collector-otlp-production-grpc | 2024-12-09 07:25:12 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-bwr9l job-controller logger.go:42: 07:25:33 | collector-otlp-production-grpc | 2024-12-09 07:25:13 +0000 UTC Normal Pod check-span-s78vd AddedInterface Add eth0 [10.128.2.31/23] from ovn-kubernetes multus logger.go:42: 07:25:33 | collector-otlp-production-grpc | 2024-12-09 07:25:13 +0000 UTC Normal Pod check-span-s78vd.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:25:33 | collector-otlp-production-grpc | 2024-12-09 07:25:13 +0000 UTC Normal Pod report-span-bwr9l AddedInterface Add eth0 [10.128.2.30/23] from ovn-kubernetes multus logger.go:42: 07:25:33 | collector-otlp-production-grpc | 2024-12-09 07:25:13 +0000 UTC Normal Pod report-span-bwr9l.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:25:33 | collector-otlp-production-grpc | 2024-12-09 07:25:14 +0000 UTC Normal Pod check-span-s78vd.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 875ms (875ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:25:33 | collector-otlp-production-grpc | 2024-12-09 07:25:14 +0000 UTC Normal Pod check-span-s78vd.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:25:33 | collector-otlp-production-grpc | 2024-12-09 07:25:14 +0000 UTC Normal Pod check-span-s78vd.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:25:33 | collector-otlp-production-grpc | 2024-12-09 07:25:14 +0000 UTC Normal Pod report-span-bwr9l.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 895ms (895ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:25:33 | collector-otlp-production-grpc | 2024-12-09 07:25:14 +0000 UTC Normal Pod report-span-bwr9l.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:25:33 | collector-otlp-production-grpc | 2024-12-09 07:25:14 +0000 UTC Normal Pod report-span-bwr9l.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:25:33 | collector-otlp-production-grpc | 2024-12-09 07:25:16 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:25:33 | collector-otlp-production-grpc | 2024-12-09 07:25:17 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get memory utilization: missing request for memory in container jaeger-collector of Pod my-jaeger-collector-64fcd4f4c9-b5q8k horizontal-pod-autoscaler logger.go:42: 07:25:33 | collector-otlp-production-grpc | 2024-12-09 07:25:17 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:25:33 | collector-otlp-production-grpc | 2024-12-09 07:25:32 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:25:33 | collector-otlp-production-grpc | 2024-12-09 07:25:32 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 07:25:33 | collector-otlp-production-grpc | 2024-12-09 07:25:32 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 07:25:33 | collector-otlp-production-grpc | Deleting namespace: kuttl-test-merry-manatee === CONT kuttl/harness/collector-autoscale logger.go:42: 07:25:45 | collector-autoscale | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:25:45 | collector-autoscale | Creating namespace: kuttl-test-present-perch logger.go:42: 07:25:45 | collector-autoscale/1-install | starting test step 1-install logger.go:42: 07:25:46 | collector-autoscale/1-install | Jaeger:kuttl-test-present-perch/simple-prod created logger.go:42: 07:26:21 | collector-autoscale/1-install | test step completed 1-install logger.go:42: 07:26:21 | collector-autoscale/2- | starting test step 2- logger.go:42: 07:26:22 | collector-autoscale/2- | test step completed 2- logger.go:42: 07:26:22 | collector-autoscale | collector-autoscale events from ns kuttl-test-present-perch: logger.go:42: 07:26:22 | collector-autoscale | 2024-12-09 07:25:51 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpresentperchsimpleprod-1-78754ccvbpm Binding Scheduled Successfully assigned kuttl-test-present-perch/elasticsearch-cdm-kuttltestpresentperchsimpleprod-1-78754ccvbpm to ip-10-0-95-250.ec2.internal default-scheduler logger.go:42: 07:26:22 | collector-autoscale | 2024-12-09 07:25:51 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestpresentperchsimpleprod-1-78754cf4cf SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestpresentperchsimpleprod-1-78754ccvbpm replicaset-controller logger.go:42: 07:26:22 | collector-autoscale | 2024-12-09 07:25:51 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestpresentperchsimpleprod-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestpresentperchsimpleprod-1-78754cf4cf to 1 deployment-controller logger.go:42: 07:26:22 | collector-autoscale | 2024-12-09 07:25:52 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpresentperchsimpleprod-1-78754ccvbpm AddedInterface Add eth0 [10.131.0.31/23] from ovn-kubernetes multus logger.go:42: 07:26:22 | collector-autoscale | 2024-12-09 07:25:52 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpresentperchsimpleprod-1-78754ccvbpm.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:8e4fbea4983cd58352349ca291383169b286bc166fad95a87807552ca43335e6" already present on machine kubelet logger.go:42: 07:26:22 | collector-autoscale | 2024-12-09 07:25:52 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpresentperchsimpleprod-1-78754ccvbpm.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:26:22 | collector-autoscale | 2024-12-09 07:25:52 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpresentperchsimpleprod-1-78754ccvbpm.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:26:22 | collector-autoscale | 2024-12-09 07:25:52 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpresentperchsimpleprod-1-78754ccvbpm.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:d68824b0b2c84db8e33edf9ab344eb684c4a7ebd7ef162bbc309043adcb28e6b" already present on machine kubelet logger.go:42: 07:26:22 | collector-autoscale | 2024-12-09 07:25:52 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpresentperchsimpleprod-1-78754ccvbpm.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:26:22 | collector-autoscale | 2024-12-09 07:25:52 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestpresentperchsimpleprod-1-78754ccvbpm.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:26:22 | collector-autoscale | 2024-12-09 07:26:02 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestpresentperchsimpleprod-1-78754ccvbpm.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 07:26:22 | collector-autoscale | 2024-12-09 07:26:07 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestpresentperchsimpleprod-1-78754ccvbpm.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:26:22 | collector-autoscale | 2024-12-09 07:26:18 +0000 UTC Normal Pod simple-prod-collector-76f84c77d-22dk5 Binding Scheduled Successfully assigned kuttl-test-present-perch/simple-prod-collector-76f84c77d-22dk5 to ip-10-0-39-223.ec2.internal default-scheduler logger.go:42: 07:26:22 | collector-autoscale | 2024-12-09 07:26:18 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-76f84c77d SuccessfulCreate Created pod: simple-prod-collector-76f84c77d-22dk5 replicaset-controller logger.go:42: 07:26:22 | collector-autoscale | 2024-12-09 07:26:18 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-76f84c77d to 1 deployment-controller logger.go:42: 07:26:22 | collector-autoscale | 2024-12-09 07:26:18 +0000 UTC Normal Pod simple-prod-query-58fd99995c-wbn72 Binding Scheduled Successfully assigned kuttl-test-present-perch/simple-prod-query-58fd99995c-wbn72 to ip-10-0-79-227.ec2.internal default-scheduler logger.go:42: 07:26:22 | collector-autoscale | 2024-12-09 07:26:18 +0000 UTC Normal ReplicaSet.apps simple-prod-query-58fd99995c SuccessfulCreate Created pod: simple-prod-query-58fd99995c-wbn72 replicaset-controller logger.go:42: 07:26:22 | collector-autoscale | 2024-12-09 07:26:18 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-58fd99995c to 1 deployment-controller logger.go:42: 07:26:22 | collector-autoscale | 2024-12-09 07:26:19 +0000 UTC Normal Pod simple-prod-collector-76f84c77d-22dk5 AddedInterface Add eth0 [10.128.2.32/23] from ovn-kubernetes multus logger.go:42: 07:26:22 | collector-autoscale | 2024-12-09 07:26:19 +0000 UTC Normal Pod simple-prod-collector-76f84c77d-22dk5.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:44686d560aa27cae8ff8693f88c4cb6e2edf1737010ec1f80709cb42250b729d" already present on machine kubelet logger.go:42: 07:26:22 | collector-autoscale | 2024-12-09 07:26:19 +0000 UTC Normal Pod simple-prod-collector-76f84c77d-22dk5.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:26:22 | collector-autoscale | 2024-12-09 07:26:19 +0000 UTC Normal Pod simple-prod-collector-76f84c77d-22dk5.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:26:22 | collector-autoscale | 2024-12-09 07:26:19 +0000 UTC Normal Pod simple-prod-query-58fd99995c-wbn72 AddedInterface Add eth0 [10.129.2.52/23] from ovn-kubernetes multus logger.go:42: 07:26:22 | collector-autoscale | 2024-12-09 07:26:19 +0000 UTC Normal Pod simple-prod-query-58fd99995c-wbn72.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:24a83160aa930e1b72c2a2442a33b28af2b06c24b058e09afcd0d495a8066d6d" already present on machine kubelet logger.go:42: 07:26:22 | collector-autoscale | 2024-12-09 07:26:19 +0000 UTC Normal Pod simple-prod-query-58fd99995c-wbn72.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:26:22 | collector-autoscale | 2024-12-09 07:26:19 +0000 UTC Normal Pod simple-prod-query-58fd99995c-wbn72.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:26:22 | collector-autoscale | 2024-12-09 07:26:19 +0000 UTC Normal Pod simple-prod-query-58fd99995c-wbn72.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:26:22 | collector-autoscale | 2024-12-09 07:26:19 +0000 UTC Normal Pod simple-prod-query-58fd99995c-wbn72.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:26:22 | collector-autoscale | 2024-12-09 07:26:19 +0000 UTC Normal Pod simple-prod-query-58fd99995c-wbn72.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:26:22 | collector-autoscale | 2024-12-09 07:26:19 +0000 UTC Normal Pod simple-prod-query-58fd99995c-wbn72.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" already present on machine kubelet logger.go:42: 07:26:22 | collector-autoscale | 2024-12-09 07:26:19 +0000 UTC Normal Pod simple-prod-query-58fd99995c-wbn72.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:26:22 | collector-autoscale | 2024-12-09 07:26:19 +0000 UTC Normal Pod simple-prod-query-58fd99995c-wbn72.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:26:22 | collector-autoscale | Deleting namespace: kuttl-test-present-perch === CONT kuttl/harness/collector-otlp-allinone-grpc logger.go:42: 07:26:28 | collector-otlp-allinone-grpc | Creating namespace: kuttl-test-hardy-quail logger.go:42: 07:26:28 | collector-otlp-allinone-grpc/0-install | starting test step 0-install logger.go:42: 07:26:28 | collector-otlp-allinone-grpc/0-install | Jaeger:kuttl-test-hardy-quail/my-jaeger created logger.go:42: 07:26:36 | collector-otlp-allinone-grpc/0-install | test step completed 0-install logger.go:42: 07:26:36 | collector-otlp-allinone-grpc/1-smoke-test | starting test step 1-smoke-test logger.go:42: 07:26:36 | collector-otlp-allinone-grpc/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:26:37 | collector-otlp-allinone-grpc/1-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:26:43 | collector-otlp-allinone-grpc/1-smoke-test | running command: [sh -c REPORTING_PROTOCOL=grpc ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest OTEL_EXPORTER_OTLP_ENDPOINT=http://my-jaeger-collector-headless:4317 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/otlp-smoke-test.yaml.template -o otlp-smoke-test-job.yaml] logger.go:42: 07:26:44 | collector-otlp-allinone-grpc/1-smoke-test | running command: [sh -c kubectl create -f otlp-smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:26:44 | collector-otlp-allinone-grpc/1-smoke-test | job.batch/report-span created logger.go:42: 07:26:44 | collector-otlp-allinone-grpc/1-smoke-test | job.batch/check-span created logger.go:42: 07:27:05 | collector-otlp-allinone-grpc/1-smoke-test | test step completed 1-smoke-test logger.go:42: 07:27:05 | collector-otlp-allinone-grpc | collector-otlp-allinone-grpc events from ns kuttl-test-hardy-quail: logger.go:42: 07:27:05 | collector-otlp-allinone-grpc | 2024-12-09 07:26:32 +0000 UTC Normal Pod my-jaeger-76f8cc548d-f5pxv Binding Scheduled Successfully assigned kuttl-test-hardy-quail/my-jaeger-76f8cc548d-f5pxv to ip-10-0-79-227.ec2.internal default-scheduler logger.go:42: 07:27:05 | collector-otlp-allinone-grpc | 2024-12-09 07:26:32 +0000 UTC Normal ReplicaSet.apps my-jaeger-76f8cc548d SuccessfulCreate Created pod: my-jaeger-76f8cc548d-f5pxv replicaset-controller logger.go:42: 07:27:05 | collector-otlp-allinone-grpc | 2024-12-09 07:26:32 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-76f8cc548d to 1 deployment-controller logger.go:42: 07:27:05 | collector-otlp-allinone-grpc | 2024-12-09 07:26:33 +0000 UTC Normal Pod my-jaeger-76f8cc548d-f5pxv AddedInterface Add eth0 [10.129.2.53/23] from ovn-kubernetes multus logger.go:42: 07:27:05 | collector-otlp-allinone-grpc | 2024-12-09 07:26:33 +0000 UTC Normal Pod my-jaeger-76f8cc548d-f5pxv.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1fa5ed13cc8a023f1e987f6cafe86adc195c373cc2b774539df6d0fd02b780a7" already present on machine kubelet logger.go:42: 07:27:05 | collector-otlp-allinone-grpc | 2024-12-09 07:26:33 +0000 UTC Normal Pod my-jaeger-76f8cc548d-f5pxv.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:27:05 | collector-otlp-allinone-grpc | 2024-12-09 07:26:33 +0000 UTC Normal Pod my-jaeger-76f8cc548d-f5pxv.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:27:05 | collector-otlp-allinone-grpc | 2024-12-09 07:26:33 +0000 UTC Normal Pod my-jaeger-76f8cc548d-f5pxv.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:27:05 | collector-otlp-allinone-grpc | 2024-12-09 07:26:33 +0000 UTC Normal Pod my-jaeger-76f8cc548d-f5pxv.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:27:05 | collector-otlp-allinone-grpc | 2024-12-09 07:26:33 +0000 UTC Normal Pod my-jaeger-76f8cc548d-f5pxv.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:27:05 | collector-otlp-allinone-grpc | 2024-12-09 07:26:40 +0000 UTC Normal Pod my-jaeger-76f8cc548d-f5pxv.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:27:05 | collector-otlp-allinone-grpc | 2024-12-09 07:26:40 +0000 UTC Normal Pod my-jaeger-76f8cc548d-f5pxv.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:27:05 | collector-otlp-allinone-grpc | 2024-12-09 07:26:40 +0000 UTC Normal ReplicaSet.apps my-jaeger-76f8cc548d SuccessfulDelete Deleted pod: my-jaeger-76f8cc548d-f5pxv replicaset-controller logger.go:42: 07:27:05 | collector-otlp-allinone-grpc | 2024-12-09 07:26:40 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled down replica set my-jaeger-76f8cc548d to 0 from 1 deployment-controller logger.go:42: 07:27:05 | collector-otlp-allinone-grpc | 2024-12-09 07:26:42 +0000 UTC Normal Pod my-jaeger-6884cf8d7b-67j9r Binding Scheduled Successfully assigned kuttl-test-hardy-quail/my-jaeger-6884cf8d7b-67j9r to ip-10-0-79-227.ec2.internal default-scheduler logger.go:42: 07:27:05 | collector-otlp-allinone-grpc | 2024-12-09 07:26:42 +0000 UTC Normal Pod my-jaeger-6884cf8d7b-67j9r AddedInterface Add eth0 [10.129.2.54/23] from ovn-kubernetes multus logger.go:42: 07:27:05 | collector-otlp-allinone-grpc | 2024-12-09 07:26:42 +0000 UTC Normal Pod my-jaeger-6884cf8d7b-67j9r.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1fa5ed13cc8a023f1e987f6cafe86adc195c373cc2b774539df6d0fd02b780a7" already present on machine kubelet logger.go:42: 07:27:05 | collector-otlp-allinone-grpc | 2024-12-09 07:26:42 +0000 UTC Normal Pod my-jaeger-6884cf8d7b-67j9r.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:27:05 | collector-otlp-allinone-grpc | 2024-12-09 07:26:42 +0000 UTC Normal Pod my-jaeger-6884cf8d7b-67j9r.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:27:05 | collector-otlp-allinone-grpc | 2024-12-09 07:26:42 +0000 UTC Normal Pod my-jaeger-6884cf8d7b-67j9r.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:27:05 | collector-otlp-allinone-grpc | 2024-12-09 07:26:42 +0000 UTC Normal Pod my-jaeger-6884cf8d7b-67j9r.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:27:05 | collector-otlp-allinone-grpc | 2024-12-09 07:26:42 +0000 UTC Normal Pod my-jaeger-6884cf8d7b-67j9r.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:27:05 | collector-otlp-allinone-grpc | 2024-12-09 07:26:42 +0000 UTC Normal ReplicaSet.apps my-jaeger-6884cf8d7b SuccessfulCreate Created pod: my-jaeger-6884cf8d7b-67j9r replicaset-controller logger.go:42: 07:27:05 | collector-otlp-allinone-grpc | 2024-12-09 07:26:42 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-6884cf8d7b to 1 deployment-controller logger.go:42: 07:27:05 | collector-otlp-allinone-grpc | 2024-12-09 07:26:44 +0000 UTC Normal Pod check-span-j4djp Binding Scheduled Successfully assigned kuttl-test-hardy-quail/check-span-j4djp to ip-10-0-39-223.ec2.internal default-scheduler logger.go:42: 07:27:05 | collector-otlp-allinone-grpc | 2024-12-09 07:26:44 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-j4djp job-controller logger.go:42: 07:27:05 | collector-otlp-allinone-grpc | 2024-12-09 07:26:44 +0000 UTC Normal Pod report-span-zkt7s Binding Scheduled Successfully assigned kuttl-test-hardy-quail/report-span-zkt7s to ip-10-0-95-250.ec2.internal default-scheduler logger.go:42: 07:27:05 | collector-otlp-allinone-grpc | 2024-12-09 07:26:44 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-zkt7s job-controller logger.go:42: 07:27:05 | collector-otlp-allinone-grpc | 2024-12-09 07:26:45 +0000 UTC Normal Pod check-span-j4djp AddedInterface Add eth0 [10.128.2.33/23] from ovn-kubernetes multus logger.go:42: 07:27:05 | collector-otlp-allinone-grpc | 2024-12-09 07:26:45 +0000 UTC Normal Pod check-span-j4djp.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:27:05 | collector-otlp-allinone-grpc | 2024-12-09 07:26:45 +0000 UTC Normal Pod report-span-zkt7s AddedInterface Add eth0 [10.131.0.32/23] from ovn-kubernetes multus logger.go:42: 07:27:05 | collector-otlp-allinone-grpc | 2024-12-09 07:26:45 +0000 UTC Normal Pod report-span-zkt7s.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:27:05 | collector-otlp-allinone-grpc | 2024-12-09 07:26:46 +0000 UTC Normal Pod check-span-j4djp.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 1.107s (1.107s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:27:05 | collector-otlp-allinone-grpc | 2024-12-09 07:26:46 +0000 UTC Normal Pod check-span-j4djp.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:27:05 | collector-otlp-allinone-grpc | 2024-12-09 07:26:46 +0000 UTC Normal Pod check-span-j4djp.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:27:05 | collector-otlp-allinone-grpc | 2024-12-09 07:26:46 +0000 UTC Normal Pod report-span-zkt7s.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 1.182s (1.182s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:27:05 | collector-otlp-allinone-grpc | 2024-12-09 07:26:46 +0000 UTC Normal Pod report-span-zkt7s.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:27:05 | collector-otlp-allinone-grpc | 2024-12-09 07:26:46 +0000 UTC Normal Pod report-span-zkt7s.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:27:05 | collector-otlp-allinone-grpc | 2024-12-09 07:27:04 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:27:05 | collector-otlp-allinone-grpc | Deleting namespace: kuttl-test-hardy-quail === CONT kuttl/harness/cassandra-spark logger.go:42: 07:27:18 | cassandra-spark | Ignoring 01-assert.yaml.template as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:27:18 | cassandra-spark | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:27:18 | cassandra-spark | Creating namespace: kuttl-test-subtle-tortoise logger.go:42: 07:27:18 | cassandra-spark | cassandra-spark events from ns kuttl-test-subtle-tortoise: logger.go:42: 07:27:18 | cassandra-spark | Deleting namespace: kuttl-test-subtle-tortoise === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (341.94s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (6.06s) --- PASS: kuttl/harness/collector-otlp-allinone-http (40.57s) --- PASS: kuttl/harness/set-custom-img (48.38s) --- PASS: kuttl/harness/collector-otlp-production-http (71.06s) --- PASS: kuttl/harness/collector-otlp-production-grpc (76.93s) --- PASS: kuttl/harness/collector-autoscale (42.89s) --- PASS: kuttl/harness/collector-otlp-allinone-grpc (49.61s) --- PASS: kuttl/harness/cassandra-spark (6.26s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name miscellaneous --report --output /logs/artifacts/miscellaneous.xml ./artifacts/kuttl-report.xml time="2024-12-09T07:27:24Z" level=debug msg="Setting a new name for the test suites" time="2024-12-09T07:27:24Z" level=debug msg="Removing 'artifacts' TestCase" time="2024-12-09T07:27:24Z" level=debug msg="normalizing test case names" time="2024-12-09T07:27:24Z" level=debug msg="miscellaneous/artifacts -> miscellaneous_artifacts" time="2024-12-09T07:27:24Z" level=debug msg="miscellaneous/collector-otlp-allinone-http -> miscellaneous_collector_otlp_allinone_http" time="2024-12-09T07:27:24Z" level=debug msg="miscellaneous/set-custom-img -> miscellaneous_set_custom_img" time="2024-12-09T07:27:24Z" level=debug msg="miscellaneous/collector-otlp-production-http -> miscellaneous_collector_otlp_production_http" time="2024-12-09T07:27:24Z" level=debug msg="miscellaneous/collector-otlp-production-grpc -> miscellaneous_collector_otlp_production_grpc" time="2024-12-09T07:27:24Z" level=debug msg="miscellaneous/collector-autoscale -> miscellaneous_collector_autoscale" time="2024-12-09T07:27:24Z" level=debug msg="miscellaneous/collector-otlp-allinone-grpc -> miscellaneous_collector_otlp_allinone_grpc" time="2024-12-09T07:27:24Z" level=debug msg="miscellaneous/cassandra-spark -> miscellaneous_cassandra_spark" +----------------------------------------------+--------+ | NAME | RESULT | +----------------------------------------------+--------+ | miscellaneous_artifacts | passed | | miscellaneous_collector_otlp_allinone_http | passed | | miscellaneous_set_custom_img | passed | | miscellaneous_collector_otlp_production_http | passed | | miscellaneous_collector_otlp_production_grpc | passed | | miscellaneous_collector_autoscale | passed | | miscellaneous_collector_otlp_allinone_grpc | passed | | miscellaneous_cassandra_spark | passed | +----------------------------------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh elasticsearch false true + '[' 3 -ne 3 ']' + test_suite_name=elasticsearch + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/elasticsearch.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-elasticsearch make[2]: Entering directory '/tmp/jaeger-tests' >>>> Elasticsearch image not loaded because SKIP_ES_EXTERNAL is true SKIP_ES_EXTERNAL=true \ KAFKA_VERSION=3.6.0 \ SKIP_KAFKA=false \ ./tests/e2e/elasticsearch/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.18.0-0.nightly-2024-12-05-191355 True False 21m Cluster version is 4.18.0-0.nightly-2024-12-05-191355' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.18.0-0.nightly-2024-12-05-191355 True False 21m Cluster version is 4.18.0-0.nightly-2024-12-05-191355' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/elasticsearch/render.sh ++ export SUITE_DIR=./tests/e2e/elasticsearch ++ SUITE_DIR=./tests/e2e/elasticsearch ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/elasticsearch ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + is_secured=false + '[' true = true ']' + is_secured=true + start_test es-from-aio-to-production + '[' 1 -ne 1 ']' + test_name=es-from-aio-to-production + echo =========================================================================== =========================================================================== + info 'Rendering files for test es-from-aio-to-production' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test es-from-aio-to-production\e[0m' Rendering files for test es-from-aio-to-production + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build + '[' _build '!=' _build ']' + mkdir -p es-from-aio-to-production + cd es-from-aio-to-production + jaeger_name=my-jaeger + render_install_jaeger my-jaeger allInOne 00 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=allInOne + test_step=00 + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test my-jaeger true 01 + '[' 3 -ne 3 ']' + jaeger=my-jaeger + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + jaeger_deploy_mode=production + [[ true = true ]] + [[ true = true ]] + jaeger_deploy_mode=production_autoprovisioned + render_install_jaeger my-jaeger production_autoprovisioned 03 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=production_autoprovisioned + test_step=03 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./03-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./03-assert.yaml + [[ true = true ]] + [[ true = true ]] + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.elasticsearch.redundancyPolicy="ZeroRedundancy"' ./03-install.yaml + render_smoke_test my-jaeger true 04 + '[' 3 -ne 3 ']' + jaeger=my-jaeger + is_secured=true + test_step=04 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./04-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./04-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test es-increasing-replicas + '[' 1 -ne 1 ']' + test_name=es-increasing-replicas + echo =========================================================================== =========================================================================== + info 'Rendering files for test es-increasing-replicas' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test es-increasing-replicas\e[0m' Rendering files for test es-increasing-replicas + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build/es-from-aio-to-production + '[' es-from-aio-to-production '!=' _build ']' + cd .. + mkdir -p es-increasing-replicas + cd es-increasing-replicas + jaeger_name=simple-prod + '[' true = true ']' + jaeger_deployment_mode=production_autoprovisioned + render_install_jaeger simple-prod production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=simple-prod + JAEGER_NAME=simple-prod + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + cp ./01-install.yaml ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.collector.replicas=2 ./02-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.query.replicas=2 ./02-install.yaml + cp ./01-assert.yaml ./02-assert.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.replicas=2 ./02-assert.yaml + /tmp/jaeger-tests/bin/yq e -i .status.readyReplicas=2 ./02-assert.yaml + render_smoke_test simple-prod true 03 + '[' 3 -ne 3 ']' + jaeger=simple-prod + is_secured=true + test_step=03 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + export JAEGER_NAME=simple-prod + JAEGER_NAME=simple-prod + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./03-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./03-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + '[' true = true ']' + cp ./02-install.yaml ./04-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.storage.elasticsearch.nodeCount=2 ./04-install.yaml + /tmp/jaeger-tests/bin/gomplate -f ./openshift-check-es-nodes.yaml.template -o ./05-check-es-nodes.yaml + '[' true = true ']' + skip_test es-index-cleaner-upstream 'SKIP_ES_EXTERNAL is true' + '[' 2 -ne 2 ']' + test_name=es-index-cleaner-upstream + message='SKIP_ES_EXTERNAL is true' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build/es-increasing-replicas + '[' es-increasing-replicas '!=' _build ']' + cd .. + rm -rf es-index-cleaner-upstream + warning 'es-index-cleaner-upstream: SKIP_ES_EXTERNAL is true' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: es-index-cleaner-upstream: SKIP_ES_EXTERNAL is true\e[0m' WAR: es-index-cleaner-upstream: SKIP_ES_EXTERNAL is true + '[' true = true ']' + es_index_cleaner -autoprov production_autoprovisioned + '[' 2 -ne 2 ']' + postfix=-autoprov + jaeger_deployment_strategy=production_autoprovisioned + start_test es-index-cleaner-autoprov + '[' 1 -ne 1 ']' + test_name=es-index-cleaner-autoprov + echo =========================================================================== =========================================================================== + info 'Rendering files for test es-index-cleaner-autoprov' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test es-index-cleaner-autoprov\e[0m' Rendering files for test es-index-cleaner-autoprov + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build + '[' _build '!=' _build ']' + mkdir -p es-index-cleaner-autoprov + cd es-index-cleaner-autoprov + jaeger_name=test-es-index-cleaner-with-prefix + cronjob_name=test-es-index-cleaner-with-prefix-es-index-cleaner + secured_es_connection=false + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_managed_es ']' + ELASTICSEARCH_URL=https://elasticsearch + secured_es_connection=true + cp ../../es-index-cleaner-upstream/04-assert.yaml ../../es-index-cleaner-upstream/README.md . + render_install_jaeger test-es-index-cleaner-with-prefix production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=test-es-index-cleaner-with-prefix + JAEGER_NAME=test-es-index-cleaner-with-prefix + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.options.es.index-prefix=""' ./01-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.storage.esIndexCleaner.enabled=false ./01-install.yaml + /tmp/jaeger-tests/bin/yq e -i .spec.storage.esIndexCleaner.numberOfDays=0 ./01-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.esIndexCleaner.schedule="*/1 * * * *"' ./01-install.yaml + render_report_spans test-es-index-cleaner-with-prefix true 5 00 true 02 + '[' 6 -ne 6 ']' + jaeger=test-es-index-cleaner-with-prefix + is_secured=true + number_of_spans=5 + job_number=00 + ensure_reported_spans=true + test_step=02 + export JAEGER_NAME=test-es-index-cleaner-with-prefix + JAEGER_NAME=test-es-index-cleaner-with-prefix + export JAEGER_COLLECTOR_ENDPOINT=http://test-es-index-cleaner-with-prefix-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://test-es-index-cleaner-with-prefix-collector-headless:14268 + export JOB_NUMBER=00 + JOB_NUMBER=00 + export DAYS=5 + DAYS=5 + '[' true = true ']' + protocol=https:// + query_port= + template=/tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template + '[' true = true ']' + export ENSURE_REPORTED_SPANS=true + ENSURE_REPORTED_SPANS=true + export JAEGER_QUERY_ENDPOINT=https://test-es-index-cleaner-with-prefix-query + JAEGER_QUERY_ENDPOINT=https://test-es-index-cleaner-with-prefix-query + params= + '[' true = true ']' + '[' true = true ']' + '[' '' '!=' allInOne ']' + params='-t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template -t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template -o ./02-report-spans.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-report-spans.yaml.template -o ./02-assert.yaml + unset JAEGER_COLLECTOR_ENDPOINT + unset JAEGER_QUERY_ENDPOINT + unset JOB_NUMBER + unset DAYS + unset ENSURE_REPORTED_SPANS + sed 's~enabled: false~enabled: true~gi' ./01-install.yaml + CRONJOB_NAME=test-es-index-cleaner-with-prefix-es-index-cleaner + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/wait-for-cronjob-execution.yaml.template -o ./04-wait-es-index-cleaner.yaml + /tmp/jaeger-tests/bin/gomplate -f ./01-install.yaml -o ./05-install.yaml + render_check_indices true ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' 00 06 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + job_number=00 + test_step=06 + escape_command ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + '[' 1 -ne 1 ']' + command=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' ++ echo ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=test-es-index-cleaner-with-prefix-curator + JOB_NUMBER=00 + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + MOUNT_SECRET=test-es-index-cleaner-with-prefix-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./06-check-indices.yaml + JOB_NUMBER=00 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./06-assert.yaml + '[' true = true ']' + get_elasticsearch_openshift_operator_version + export ESO_OPERATOR_VERSION + '[' true = true ']' ++ kubectl get pods -l name=elasticsearch-operator --all-namespaces '-o=jsonpath={.items[0].metadata.annotations.operatorframework\.io/properties}' + properties='{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.18},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.15"}}]}' + '[' -z '{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.18},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.15"}}]}' ']' ++ echo '{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.18},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.15"}}]}' ++ /tmp/jaeger-tests/bin/yq e -P '.properties.[] | select(.value.packageName == "elasticsearch-operator") | .value.version' + ESO_OPERATOR_VERSION=5.8.15 ++ version_ge 5.8.15 5.4 +++ echo 5.8.15 5.4 +++ tr ' ' '\n' +++ sort -rV +++ head -n 1 ++ test 5.8.15 == 5.8.15 + '[' -n '' ']' + skip_test es-index-cleaner-managed 'Test only supported with Elasticsearch OpenShift Operator >= 5.4' + '[' 2 -ne 2 ']' + test_name=es-index-cleaner-managed + message='Test only supported with Elasticsearch OpenShift Operator >= 5.4' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build/es-index-cleaner-autoprov + '[' es-index-cleaner-autoprov '!=' _build ']' + cd .. + rm -rf es-index-cleaner-managed + warning 'es-index-cleaner-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: es-index-cleaner-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4\e[0m' WAR: es-index-cleaner-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4 + '[' true = true ']' + start_test es-multiinstance + '[' 1 -ne 1 ']' + test_name=es-multiinstance + echo =========================================================================== =========================================================================== + info 'Rendering files for test es-multiinstance' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test es-multiinstance\e[0m' Rendering files for test es-multiinstance + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build + '[' _build '!=' _build ']' + mkdir -p es-multiinstance + cd es-multiinstance + jaeger_name=instance-1 + render_install_jaeger instance-1 production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=instance-1 + JAEGER_NAME=instance-1 + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + /tmp/jaeger-tests/bin/gomplate -f ./03-create-second-instance.yaml.template -o 03-create-second-instance.yaml + '[' true = true ']' + skip_test es-rollover-upstream 'SKIP_ES_EXTERNAL is true' + '[' 2 -ne 2 ']' + test_name=es-rollover-upstream + message='SKIP_ES_EXTERNAL is true' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build/es-multiinstance + '[' es-multiinstance '!=' _build ']' + cd .. + rm -rf es-rollover-upstream + warning 'es-rollover-upstream: SKIP_ES_EXTERNAL is true' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: es-rollover-upstream: SKIP_ES_EXTERNAL is true\e[0m' WAR: es-rollover-upstream: SKIP_ES_EXTERNAL is true + '[' true = true ']' + es_rollover -autoprov production_autoprovisioned + '[' 2 -ne 2 ']' + postfix=-autoprov + jaeger_deployment_strategy=production_autoprovisioned + start_test es-rollover-autoprov + '[' 1 -ne 1 ']' + test_name=es-rollover-autoprov + echo =========================================================================== =========================================================================== + info 'Rendering files for test es-rollover-autoprov' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test es-rollover-autoprov\e[0m' Rendering files for test es-rollover-autoprov + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build + '[' _build '!=' _build ']' + mkdir -p es-rollover-autoprov + cd es-rollover-autoprov + cp ../../es-rollover-upstream/05-assert.yaml ../../es-rollover-upstream/05-install.yaml ../../es-rollover-upstream/README.md . + jaeger_name=my-jaeger + secured_es_connection=false + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_managed_es ']' + ELASTICSEARCH_URL=https://elasticsearch + secured_es_connection=true + render_install_jaeger my-jaeger production_autoprovisioned 01 + '[' 3 -ne 3 ']' + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + deploy_mode=production_autoprovisioned + test_step=01 + '[' production_autoprovisioned = allInOne ']' + '[' production_autoprovisioned = production ']' + '[' production_autoprovisioned = production_cassandra ']' + '[' production_autoprovisioned = production_autoprovisioned ']' + '[' true '!=' true ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/production-jaeger-autoprovisioned-install.yaml.template -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + render_report_spans my-jaeger true 2 00 true 02 + '[' 6 -ne 6 ']' + jaeger=my-jaeger + is_secured=true + number_of_spans=2 + job_number=00 + ensure_reported_spans=true + test_step=02 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JOB_NUMBER=00 + JOB_NUMBER=00 + export DAYS=2 + DAYS=2 + '[' true = true ']' + protocol=https:// + query_port= + template=/tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template + '[' true = true ']' + export ENSURE_REPORTED_SPANS=true + ENSURE_REPORTED_SPANS=true + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + params= + '[' true = true ']' + '[' true = true ']' + '[' '' '!=' allInOne ']' + params='-t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template -t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template -o ./02-report-spans.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-report-spans.yaml.template -o ./02-assert.yaml + unset JAEGER_COLLECTOR_ENDPOINT + unset JAEGER_QUERY_ENDPOINT + unset JOB_NUMBER + unset DAYS + unset ENSURE_REPORTED_SPANS + render_check_indices true ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' 00 03 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + job_number=00 + test_step=03 + escape_command ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + '[' 1 -ne 1 ']' + command=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' ++ echo ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-exist'\'',' + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-exist'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=00 + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./03-check-indices.yaml + JOB_NUMBER=00 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./03-assert.yaml + render_check_indices true ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' 01 04 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + job_number=01 + test_step=04 + escape_command ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + '[' 1 -ne 1 ']' + command=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' ++ echo ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--pattern'\'', '\''jaeger-span-\\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=01 + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-count-indices'\'', '\''0'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./04-check-indices.yaml + JOB_NUMBER=01 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./04-assert.yaml + render_report_spans my-jaeger true 2 02 true 06 + '[' 6 -ne 6 ']' + jaeger=my-jaeger + is_secured=true + number_of_spans=2 + job_number=02 + ensure_reported_spans=true + test_step=06 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JOB_NUMBER=02 + JOB_NUMBER=02 + export DAYS=2 + DAYS=2 + '[' true = true ']' + protocol=https:// + query_port= + template=/tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template + '[' true = true ']' + export ENSURE_REPORTED_SPANS=true + ENSURE_REPORTED_SPANS=true + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + params= + '[' true = true ']' + '[' true = true ']' + '[' '' '!=' allInOne ']' + params='-t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template -t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template -o ./06-report-spans.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-report-spans.yaml.template -o ./06-assert.yaml + unset JAEGER_COLLECTOR_ENDPOINT + unset JAEGER_QUERY_ENDPOINT + unset JOB_NUMBER + unset DAYS + unset ENSURE_REPORTED_SPANS + render_check_indices true ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' 02 07 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + job_number=02 + test_step=07 + escape_command ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + '[' 1 -ne 1 ']' + command=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' ++ echo ''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-exist'\'',' + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\\d{4}-\\d{2}-\\d{2}'\'', '\''--assert-exist'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=02 + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\d{4}-\d{2}-\d{2}'\'', '\''--assert-exist'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./07-check-indices.yaml + JOB_NUMBER=02 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./07-assert.yaml + render_check_indices true ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' 03 08 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' + job_number=03 + test_step=08 + escape_command ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' + '[' 1 -ne 1 ']' + command=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' ++ echo ''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--pattern'\'', '\''jaeger-span-\\d{6}'\'', '\''--assert-exist'\'',' + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\\d{6}'\'', '\''--assert-exist'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=03 + CMD_PARAMETERS=''\''--pattern'\'', '\''jaeger-span-\d{6}'\'', '\''--assert-exist'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./08-check-indices.yaml + JOB_NUMBER=03 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./08-assert.yaml + render_check_indices true ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' 04 09 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' + job_number=04 + test_step=09 + escape_command ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' + '[' 1 -ne 1 ']' + command=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' ++ echo ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=04 + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-exist'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./09-check-indices.yaml + JOB_NUMBER=04 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./09-assert.yaml + render_report_spans my-jaeger true 2 03 true 10 + '[' 6 -ne 6 ']' + jaeger=my-jaeger + is_secured=true + number_of_spans=2 + job_number=03 + ensure_reported_spans=true + test_step=10 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JOB_NUMBER=03 + JOB_NUMBER=03 + export DAYS=2 + DAYS=2 + '[' true = true ']' + protocol=https:// + query_port= + template=/tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template + '[' true = true ']' + export ENSURE_REPORTED_SPANS=true + ENSURE_REPORTED_SPANS=true + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query + params= + '[' true = true ']' + '[' true = true ']' + '[' '' '!=' allInOne ']' + params='-t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/report-spans.yaml.template -t /tmp/jaeger-tests/tests/templates/openshift/configure-api-query-oauth.yaml.template -o ./10-report-spans.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-report-spans.yaml.template -o ./10-assert.yaml + unset JAEGER_COLLECTOR_ENDPOINT + unset JAEGER_QUERY_ENDPOINT + unset JOB_NUMBER + unset DAYS + unset ENSURE_REPORTED_SPANS + CRONJOB_NAME=my-jaeger-es-rollover + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/wait-for-cronjob-execution.yaml.template -o ./11-wait-rollover.yaml + render_check_indices true ''\''--name'\'', '\''jaeger-span-000002'\'',' 05 11 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--name'\'', '\''jaeger-span-000002'\'',' + job_number=05 + test_step=11 + escape_command ''\''--name'\'', '\''jaeger-span-000002'\'',' + '[' 1 -ne 1 ']' + command=''\''--name'\'', '\''jaeger-span-000002'\'',' ++ echo ''\''--name'\'', '\''jaeger-span-000002'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--name'\'', '\''jaeger-span-000002'\'',' + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-000002'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=05 + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-000002'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./11-check-indices.yaml + JOB_NUMBER=05 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./11-assert.yaml + render_check_indices true ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' 06 12 + '[' 4 -ne 4 ']' + secured=true + cmd_parameters=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' + job_number=06 + test_step=12 + escape_command ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' + '[' 1 -ne 1 ']' + command=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' ++ echo ''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' ++ sed 's/\\/\\\\/g' + export 'CMD_PARAMETERS='\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' + mount_secret= + '[' true = true ']' + '[' true = true ']' + mount_secret=my-jaeger-curator + JOB_NUMBER=06 + CMD_PARAMETERS=''\''--name'\'', '\''jaeger-span-read'\'', '\''--assert-count-docs'\'', '\''4'\'', '\''--jaeger-service'\'', '\''smoke-test-service'\'',' + MOUNT_SECRET=my-jaeger-curator + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/check-indices.yaml.template -o ./12-check-indices.yaml + JOB_NUMBER=06 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-check-indices.yaml.template -o ./12-assert.yaml + '[' true = true ']' + get_elasticsearch_openshift_operator_version + export ESO_OPERATOR_VERSION + '[' true = true ']' ++ kubectl get pods -l name=elasticsearch-operator --all-namespaces '-o=jsonpath={.items[0].metadata.annotations.operatorframework\.io/properties}' + properties='{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.18},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.15"}}]}' + '[' -z '{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.18},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.15"}}]}' ']' ++ echo '{"properties":[{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Elasticsearch","version":"v1"}},{"type":"olm.gvk","value":{"group":"logging.openshift.io","kind":"Kibana","version":"v1"}},{"type":"olm.maxOpenShiftVersion","value":4.18},{"type":"olm.package","value":{"packageName":"elasticsearch-operator","version":"5.8.15"}}]}' ++ /tmp/jaeger-tests/bin/yq e -P '.properties.[] | select(.value.packageName == "elasticsearch-operator") | .value.version' + ESO_OPERATOR_VERSION=5.8.15 ++ version_ge 5.8.15 5.4 +++ echo 5.8.15 5.4 +++ tr ' ' '\n' +++ sort -rV +++ head -n 1 ++ test 5.8.15 == 5.8.15 + '[' -n '' ']' + skip_test es-rollover-managed 'Test only supported with Elasticsearch OpenShift Operator >= 5.4' + '[' 2 -ne 2 ']' + test_name=es-rollover-managed + message='Test only supported with Elasticsearch OpenShift Operator >= 5.4' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build/es-rollover-autoprov + '[' es-rollover-autoprov '!=' _build ']' + cd .. + rm -rf es-rollover-managed + warning 'es-rollover-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: es-rollover-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4\e[0m' WAR: es-rollover-managed: Test only supported with Elasticsearch OpenShift Operator >= 5.4 + '[' true = true ']' + skip_test es-spark-dependencies 'This test is not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=es-spark-dependencies + message='This test is not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/elasticsearch/_build + '[' _build '!=' _build ']' + rm -rf es-spark-dependencies + warning 'es-spark-dependencies: This test is not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: es-spark-dependencies: This test is not supported in OpenShift\e[0m' WAR: es-spark-dependencies: This test is not supported in OpenShift make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running elasticsearch E2E tests' Running elasticsearch E2E tests + cd tests/e2e/elasticsearch/_build + set +e + KUBECONFIG=/tmp/kubeconfig-591533165 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-54q9kfc5-9210b.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 7 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/es-from-aio-to-production === PAUSE kuttl/harness/es-from-aio-to-production === RUN kuttl/harness/es-increasing-replicas === PAUSE kuttl/harness/es-increasing-replicas === RUN kuttl/harness/es-index-cleaner-autoprov === PAUSE kuttl/harness/es-index-cleaner-autoprov === RUN kuttl/harness/es-multiinstance === PAUSE kuttl/harness/es-multiinstance === RUN kuttl/harness/es-rollover-autoprov === PAUSE kuttl/harness/es-rollover-autoprov === RUN kuttl/harness/es-simple-prod === PAUSE kuttl/harness/es-simple-prod === CONT kuttl/harness/artifacts logger.go:42: 07:27:48 | artifacts | Creating namespace: kuttl-test-deep-buck logger.go:42: 07:27:48 | artifacts | artifacts events from ns kuttl-test-deep-buck: logger.go:42: 07:27:48 | artifacts | Deleting namespace: kuttl-test-deep-buck === CONT kuttl/harness/es-multiinstance logger.go:42: 07:27:54 | es-multiinstance | Ignoring 03-create-second-instance.yaml.template as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:27:54 | es-multiinstance | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:27:54 | es-multiinstance | Creating namespace: kuttl-test-humble-earwig logger.go:42: 07:27:54 | es-multiinstance/0-clear-namespace | starting test step 0-clear-namespace logger.go:42: 07:27:54 | es-multiinstance/0-clear-namespace | running command: [sh -c kubectl delete namespace jaeger-e2e-multiinstance-test --ignore-not-found=true] logger.go:42: 07:27:54 | es-multiinstance/0-clear-namespace | test step completed 0-clear-namespace logger.go:42: 07:27:54 | es-multiinstance/1-install | starting test step 1-install logger.go:42: 07:27:54 | es-multiinstance/1-install | Jaeger:kuttl-test-humble-earwig/instance-1 created logger.go:42: 07:28:33 | es-multiinstance/1-install | test step completed 1-install logger.go:42: 07:28:33 | es-multiinstance/2-create-namespace | starting test step 2-create-namespace logger.go:42: 07:28:33 | es-multiinstance/2-create-namespace | running command: [sh -c kubectl create namespace jaeger-e2e-multiinstance-test] logger.go:42: 07:28:33 | es-multiinstance/2-create-namespace | namespace/jaeger-e2e-multiinstance-test created logger.go:42: 07:28:33 | es-multiinstance/2-create-namespace | test step completed 2-create-namespace logger.go:42: 07:28:33 | es-multiinstance/3-create-second-instance | starting test step 3-create-second-instance logger.go:42: 07:28:33 | es-multiinstance/3-create-second-instance | running command: [sh -c kubectl apply -f ./01-install.yaml -n jaeger-e2e-multiinstance-test] logger.go:42: 07:28:33 | es-multiinstance/3-create-second-instance | jaeger.jaegertracing.io/instance-1 created logger.go:42: 07:28:33 | es-multiinstance/3-create-second-instance | running command: [sh -c /tmp/jaeger-tests/bin/kubectl-kuttl assert ./01-assert.yaml -n jaeger-e2e-multiinstance-test --timeout 1000] logger.go:42: 07:29:25 | es-multiinstance/3-create-second-instance | assert is valid logger.go:42: 07:29:25 | es-multiinstance/3-create-second-instance | test step completed 3-create-second-instance logger.go:42: 07:29:25 | es-multiinstance/4-check-secrets | starting test step 4-check-secrets logger.go:42: 07:29:25 | es-multiinstance/4-check-secrets | running command: [sh -c kubectl get secrets elasticsearch -o jsonpath='{.data.logging-es\.crt}' -n $NAMESPACE > secret1] logger.go:42: 07:29:26 | es-multiinstance/4-check-secrets | running command: [sh -c kubectl get secrets elasticsearch -o jsonpath='{.data.logging-es\.crt}' -n jaeger-e2e-multiinstance-test > secret2] logger.go:42: 07:29:26 | es-multiinstance/4-check-secrets | running command: [sh -c cmp --silent secret1 secret2 || exit 0] logger.go:42: 07:29:26 | es-multiinstance/4-check-secrets | test step completed 4-check-secrets logger.go:42: 07:29:26 | es-multiinstance/5-delete | starting test step 5-delete logger.go:42: 07:29:26 | es-multiinstance/5-delete | running command: [sh -c kubectl delete namespace jaeger-e2e-multiinstance-test --wait=false] logger.go:42: 07:29:26 | es-multiinstance/5-delete | namespace "jaeger-e2e-multiinstance-test" deleted logger.go:42: 07:29:26 | es-multiinstance/5-delete | test step completed 5-delete logger.go:42: 07:29:26 | es-multiinstance | es-multiinstance events from ns kuttl-test-humble-earwig: logger.go:42: 07:29:26 | es-multiinstance | 2024-12-09 07:28:03 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltesthumbleearwiginstance1-1-774574c6f7 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltesthumbleearwiginstance1-1-774574crxn4d replicaset-controller logger.go:42: 07:29:26 | es-multiinstance | 2024-12-09 07:28:03 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesthumbleearwiginstance1-1-774574crxn4d Binding Scheduled Successfully assigned kuttl-test-humble-earwig/elasticsearch-cdm-kuttltesthumbleearwiginstance1-1-774574crxn4d to ip-10-0-95-250.ec2.internal default-scheduler logger.go:42: 07:29:26 | es-multiinstance | 2024-12-09 07:28:03 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltesthumbleearwiginstance1-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltesthumbleearwiginstance1-1-774574c6f7 to 1 deployment-controller logger.go:42: 07:29:26 | es-multiinstance | 2024-12-09 07:28:04 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesthumbleearwiginstance1-1-774574crxn4d AddedInterface Add eth0 [10.131.0.33/23] from ovn-kubernetes multus logger.go:42: 07:29:26 | es-multiinstance | 2024-12-09 07:28:04 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesthumbleearwiginstance1-1-774574crxn4d.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:8e4fbea4983cd58352349ca291383169b286bc166fad95a87807552ca43335e6" already present on machine kubelet logger.go:42: 07:29:26 | es-multiinstance | 2024-12-09 07:28:04 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesthumbleearwiginstance1-1-774574crxn4d.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:29:26 | es-multiinstance | 2024-12-09 07:28:04 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesthumbleearwiginstance1-1-774574crxn4d.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:29:26 | es-multiinstance | 2024-12-09 07:28:04 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesthumbleearwiginstance1-1-774574crxn4d.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:d68824b0b2c84db8e33edf9ab344eb684c4a7ebd7ef162bbc309043adcb28e6b" already present on machine kubelet logger.go:42: 07:29:26 | es-multiinstance | 2024-12-09 07:28:04 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesthumbleearwiginstance1-1-774574crxn4d.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:29:26 | es-multiinstance | 2024-12-09 07:28:04 +0000 UTC Normal Pod elasticsearch-cdm-kuttltesthumbleearwiginstance1-1-774574crxn4d.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:29:26 | es-multiinstance | 2024-12-09 07:28:18 +0000 UTC Warning Pod elasticsearch-cdm-kuttltesthumbleearwiginstance1-1-774574crxn4d.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:29:26 | es-multiinstance | 2024-12-09 07:28:30 +0000 UTC Normal Pod instance-1-collector-74fc68b785-l2kkj Binding Scheduled Successfully assigned kuttl-test-humble-earwig/instance-1-collector-74fc68b785-l2kkj to ip-10-0-39-223.ec2.internal default-scheduler logger.go:42: 07:29:26 | es-multiinstance | 2024-12-09 07:28:30 +0000 UTC Normal Pod instance-1-collector-74fc68b785-l2kkj AddedInterface Add eth0 [10.128.2.34/23] from ovn-kubernetes multus logger.go:42: 07:29:26 | es-multiinstance | 2024-12-09 07:28:30 +0000 UTC Normal Pod instance-1-collector-74fc68b785-l2kkj.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:44686d560aa27cae8ff8693f88c4cb6e2edf1737010ec1f80709cb42250b729d" already present on machine kubelet logger.go:42: 07:29:26 | es-multiinstance | 2024-12-09 07:28:30 +0000 UTC Normal Pod instance-1-collector-74fc68b785-l2kkj.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:29:26 | es-multiinstance | 2024-12-09 07:28:30 +0000 UTC Normal Pod instance-1-collector-74fc68b785-l2kkj.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:29:26 | es-multiinstance | 2024-12-09 07:28:30 +0000 UTC Normal ReplicaSet.apps instance-1-collector-74fc68b785 SuccessfulCreate Created pod: instance-1-collector-74fc68b785-l2kkj replicaset-controller logger.go:42: 07:29:26 | es-multiinstance | 2024-12-09 07:28:30 +0000 UTC Normal Deployment.apps instance-1-collector ScalingReplicaSet Scaled up replica set instance-1-collector-74fc68b785 to 1 deployment-controller logger.go:42: 07:29:26 | es-multiinstance | 2024-12-09 07:28:30 +0000 UTC Normal Pod instance-1-query-87477f48d-2pkh9 Binding Scheduled Successfully assigned kuttl-test-humble-earwig/instance-1-query-87477f48d-2pkh9 to ip-10-0-79-227.ec2.internal default-scheduler logger.go:42: 07:29:26 | es-multiinstance | 2024-12-09 07:28:30 +0000 UTC Normal Pod instance-1-query-87477f48d-2pkh9 AddedInterface Add eth0 [10.129.2.55/23] from ovn-kubernetes multus logger.go:42: 07:29:26 | es-multiinstance | 2024-12-09 07:28:30 +0000 UTC Normal Pod instance-1-query-87477f48d-2pkh9.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:24a83160aa930e1b72c2a2442a33b28af2b06c24b058e09afcd0d495a8066d6d" already present on machine kubelet logger.go:42: 07:29:26 | es-multiinstance | 2024-12-09 07:28:30 +0000 UTC Normal Pod instance-1-query-87477f48d-2pkh9.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:29:26 | es-multiinstance | 2024-12-09 07:28:30 +0000 UTC Normal Pod instance-1-query-87477f48d-2pkh9.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:29:26 | es-multiinstance | 2024-12-09 07:28:30 +0000 UTC Normal Pod instance-1-query-87477f48d-2pkh9.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:29:26 | es-multiinstance | 2024-12-09 07:28:30 +0000 UTC Normal ReplicaSet.apps instance-1-query-87477f48d SuccessfulCreate Created pod: instance-1-query-87477f48d-2pkh9 replicaset-controller logger.go:42: 07:29:26 | es-multiinstance | 2024-12-09 07:28:30 +0000 UTC Normal Deployment.apps instance-1-query ScalingReplicaSet Scaled up replica set instance-1-query-87477f48d to 1 deployment-controller logger.go:42: 07:29:26 | es-multiinstance | 2024-12-09 07:28:31 +0000 UTC Normal Pod instance-1-query-87477f48d-2pkh9.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:29:26 | es-multiinstance | 2024-12-09 07:28:31 +0000 UTC Normal Pod instance-1-query-87477f48d-2pkh9.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:29:26 | es-multiinstance | 2024-12-09 07:28:31 +0000 UTC Normal Pod instance-1-query-87477f48d-2pkh9.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" already present on machine kubelet logger.go:42: 07:29:26 | es-multiinstance | 2024-12-09 07:28:31 +0000 UTC Normal Pod instance-1-query-87477f48d-2pkh9.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:29:26 | es-multiinstance | 2024-12-09 07:28:31 +0000 UTC Normal Pod instance-1-query-87477f48d-2pkh9.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:29:26 | es-multiinstance | 2024-12-09 07:28:45 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling instance-1-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:29:26 | es-multiinstance | 2024-12-09 07:28:45 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling instance-1-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:29:26 | es-multiinstance | 2024-12-09 07:28:45 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling instance-1-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:29:26 | es-multiinstance | 2024-12-09 07:29:00 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling instance-1-collector FailedGetResourceMetric failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 07:29:26 | es-multiinstance | 2024-12-09 07:29:00 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling instance-1-collector FailedGetResourceMetric failed to get memory utilization: missing request for memory in container jaeger-collector of Pod instance-1-collector-74fc68b785-l2kkj horizontal-pod-autoscaler logger.go:42: 07:29:26 | es-multiinstance | 2024-12-09 07:29:00 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling instance-1-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: did not receive metrics for targeted pods (pods might be unready) horizontal-pod-autoscaler logger.go:42: 07:29:26 | es-multiinstance | 2024-12-09 07:29:15 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling instance-1-collector FailedGetResourceMetric failed to get cpu utilization: missing request for cpu in container jaeger-collector of Pod instance-1-collector-74fc68b785-l2kkj horizontal-pod-autoscaler logger.go:42: 07:29:26 | es-multiinstance | 2024-12-09 07:29:15 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling instance-1-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: missing request for cpu in container jaeger-collector of Pod instance-1-collector-74fc68b785-l2kkj horizontal-pod-autoscaler logger.go:42: 07:29:26 | es-multiinstance | Deleting namespace: kuttl-test-humble-earwig === CONT kuttl/harness/es-simple-prod logger.go:42: 07:29:33 | es-simple-prod | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:29:33 | es-simple-prod | Creating namespace: kuttl-test-hardy-chimp logger.go:42: 07:29:33 | es-simple-prod | es-simple-prod events from ns kuttl-test-hardy-chimp: logger.go:42: 07:29:33 | es-simple-prod | Deleting namespace: kuttl-test-hardy-chimp === CONT kuttl/harness/es-rollover-autoprov logger.go:42: 07:29:40 | es-rollover-autoprov | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:29:40 | es-rollover-autoprov | Creating namespace: kuttl-test-ace-dory logger.go:42: 07:29:40 | es-rollover-autoprov/1-install | starting test step 1-install logger.go:42: 07:29:40 | es-rollover-autoprov/1-install | Jaeger:kuttl-test-ace-dory/my-jaeger created logger.go:42: 07:30:16 | es-rollover-autoprov/1-install | test step completed 1-install logger.go:42: 07:30:16 | es-rollover-autoprov/2-report-spans | starting test step 2-report-spans logger.go:42: 07:30:16 | es-rollover-autoprov/2-report-spans | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:30:17 | es-rollover-autoprov/2-report-spans | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:30:26 | es-rollover-autoprov/2-report-spans | running command: [sh -c DAYS=2 ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JOB_NUMBER=00 JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/report-spans.yaml.template -o report-span-00-job.yaml] logger.go:42: 07:30:27 | es-rollover-autoprov/2-report-spans | running command: [sh -c kubectl apply -f report-span-00-job.yaml -n $NAMESPACE] logger.go:42: 07:30:27 | es-rollover-autoprov/2-report-spans | job.batch/00-report-span created logger.go:42: 07:30:53 | es-rollover-autoprov/2-report-spans | test step completed 2-report-spans logger.go:42: 07:30:53 | es-rollover-autoprov/3-check-indices | starting test step 3-check-indices logger.go:42: 07:30:53 | es-rollover-autoprov/3-check-indices | Job:kuttl-test-ace-dory/00-check-indices created logger.go:42: 07:30:58 | es-rollover-autoprov/3-check-indices | test step completed 3-check-indices logger.go:42: 07:30:58 | es-rollover-autoprov/4-check-indices | starting test step 4-check-indices logger.go:42: 07:30:58 | es-rollover-autoprov/4-check-indices | Job:kuttl-test-ace-dory/01-check-indices created logger.go:42: 07:31:03 | es-rollover-autoprov/4-check-indices | test step completed 4-check-indices logger.go:42: 07:31:03 | es-rollover-autoprov/5-install | starting test step 5-install logger.go:42: 07:31:03 | es-rollover-autoprov/5-install | Jaeger:kuttl-test-ace-dory/my-jaeger updated logger.go:42: 07:31:11 | es-rollover-autoprov/5-install | test step completed 5-install logger.go:42: 07:31:11 | es-rollover-autoprov/6-report-spans | starting test step 6-report-spans logger.go:42: 07:31:11 | es-rollover-autoprov/6-report-spans | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:31:19 | es-rollover-autoprov/6-report-spans | running command: [sh -c DAYS=2 ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JOB_NUMBER=02 JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/report-spans.yaml.template -o report-span-02-job.yaml] logger.go:42: 07:31:20 | es-rollover-autoprov/6-report-spans | running command: [sh -c kubectl apply -f report-span-02-job.yaml -n $NAMESPACE] logger.go:42: 07:31:20 | es-rollover-autoprov/6-report-spans | job.batch/02-report-span created logger.go:42: 07:31:44 | es-rollover-autoprov/6-report-spans | test step completed 6-report-spans logger.go:42: 07:31:44 | es-rollover-autoprov/7-check-indices | starting test step 7-check-indices logger.go:42: 07:31:44 | es-rollover-autoprov/7-check-indices | Job:kuttl-test-ace-dory/02-check-indices created logger.go:42: 07:31:50 | es-rollover-autoprov/7-check-indices | test step completed 7-check-indices logger.go:42: 07:31:50 | es-rollover-autoprov/8-check-indices | starting test step 8-check-indices logger.go:42: 07:31:50 | es-rollover-autoprov/8-check-indices | Job:kuttl-test-ace-dory/03-check-indices created logger.go:42: 07:31:55 | es-rollover-autoprov/8-check-indices | test step completed 8-check-indices logger.go:42: 07:31:55 | es-rollover-autoprov/9-check-indices | starting test step 9-check-indices logger.go:42: 07:31:55 | es-rollover-autoprov/9-check-indices | Job:kuttl-test-ace-dory/04-check-indices created logger.go:42: 07:31:59 | es-rollover-autoprov/9-check-indices | test step completed 9-check-indices logger.go:42: 07:31:59 | es-rollover-autoprov/10-report-spans | starting test step 10-report-spans logger.go:42: 07:31:59 | es-rollover-autoprov/10-report-spans | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:32:07 | es-rollover-autoprov/10-report-spans | running command: [sh -c DAYS=2 ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JOB_NUMBER=03 JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/report-spans.yaml.template -o report-span-03-job.yaml] logger.go:42: 07:32:08 | es-rollover-autoprov/10-report-spans | running command: [sh -c kubectl apply -f report-span-03-job.yaml -n $NAMESPACE] logger.go:42: 07:32:08 | es-rollover-autoprov/10-report-spans | job.batch/03-report-span created logger.go:42: 07:32:33 | es-rollover-autoprov/10-report-spans | test step completed 10-report-spans logger.go:42: 07:32:33 | es-rollover-autoprov/11-check-indices | starting test step 11-check-indices logger.go:42: 07:32:33 | es-rollover-autoprov/11-check-indices | running command: [sh -c go run ../../../../cmd-utils/wait-cronjob/main.go --cronjob my-jaeger-es-rollover --namespace $NAMESPACE] logger.go:42: 07:32:43 | es-rollover-autoprov/11-check-indices | time="2024-12-09T07:32:43Z" level=debug msg="Checking if the my-jaeger-es-rollover CronJob exists" logger.go:42: 07:32:43 | es-rollover-autoprov/11-check-indices | time="2024-12-09T07:32:43Z" level=debug msg="No BatchV1beta1/Cronjobs were found" logger.go:42: 07:32:43 | es-rollover-autoprov/11-check-indices | time="2024-12-09T07:32:43Z" level=info msg="Cronjob my-jaeger-es-rollover found successfully" logger.go:42: 07:32:43 | es-rollover-autoprov/11-check-indices | time="2024-12-09T07:32:43Z" level=debug msg="Waiting for the next scheduled job from my-jaeger-es-rollover cronjob" logger.go:42: 07:32:43 | es-rollover-autoprov/11-check-indices | time="2024-12-09T07:32:43Z" level=debug msg="Waiting for next job from my-jaeger-es-rollover to succeed" logger.go:42: 07:32:53 | es-rollover-autoprov/11-check-indices | time="2024-12-09T07:32:53Z" level=debug msg="Waiting for next job from my-jaeger-es-rollover to succeed" logger.go:42: 07:33:03 | es-rollover-autoprov/11-check-indices | time="2024-12-09T07:33:03Z" level=debug msg="Waiting for next job from my-jaeger-es-rollover to succeed" logger.go:42: 07:33:13 | es-rollover-autoprov/11-check-indices | time="2024-12-09T07:33:13Z" level=info msg="Job of owner my-jaeger-es-rollover succeeded after my-jaeger-es-rollover 30.044581277s" logger.go:42: 07:33:14 | es-rollover-autoprov/11-check-indices | Job:kuttl-test-ace-dory/05-check-indices created logger.go:42: 07:33:18 | es-rollover-autoprov/11-check-indices | test step completed 11-check-indices logger.go:42: 07:33:18 | es-rollover-autoprov/12-check-indices | starting test step 12-check-indices logger.go:42: 07:33:18 | es-rollover-autoprov/12-check-indices | Job:kuttl-test-ace-dory/06-check-indices created logger.go:42: 07:33:23 | es-rollover-autoprov/12-check-indices | test step completed 12-check-indices logger.go:42: 07:33:23 | es-rollover-autoprov | es-rollover-autoprov events from ns kuttl-test-ace-dory: logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:29:46 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestacedorymyjaeger-1-9854b467-qn4zr Binding Scheduled Successfully assigned kuttl-test-ace-dory/elasticsearch-cdm-kuttltestacedorymyjaeger-1-9854b467-qn4zr to ip-10-0-95-250.ec2.internal default-scheduler logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:29:46 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestacedorymyjaeger-1-9854b467-qn4zr AddedInterface Add eth0 [10.131.0.34/23] from ovn-kubernetes multus logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:29:46 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestacedorymyjaeger-1-9854b467-qn4zr.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:8e4fbea4983cd58352349ca291383169b286bc166fad95a87807552ca43335e6" already present on machine kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:29:46 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestacedorymyjaeger-1-9854b467-qn4zr.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:29:46 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestacedorymyjaeger-1-9854b467-qn4zr.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:29:46 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestacedorymyjaeger-1-9854b467-qn4zr.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:d68824b0b2c84db8e33edf9ab344eb684c4a7ebd7ef162bbc309043adcb28e6b" already present on machine kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:29:46 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestacedorymyjaeger-1-9854b467 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestacedorymyjaeger-1-9854b467-qn4zr replicaset-controller logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:29:46 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestacedorymyjaeger-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestacedorymyjaeger-1-9854b467 to 1 deployment-controller logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:29:47 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestacedorymyjaeger-1-9854b467-qn4zr.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:29:47 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestacedorymyjaeger-1-9854b467-qn4zr.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:29:56 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestacedorymyjaeger-1-9854b467-qn4zr.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:30:01 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestacedorymyjaeger-1-9854b467-qn4zr.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:30:12 +0000 UTC Normal Pod my-jaeger-collector-9b5c76b54-fbnb8 Binding Scheduled Successfully assigned kuttl-test-ace-dory/my-jaeger-collector-9b5c76b54-fbnb8 to ip-10-0-79-227.ec2.internal default-scheduler logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:30:12 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-9b5c76b54 SuccessfulCreate Created pod: my-jaeger-collector-9b5c76b54-fbnb8 replicaset-controller logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:30:12 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-9b5c76b54 to 1 deployment-controller logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:30:12 +0000 UTC Normal Pod my-jaeger-query-7d9b66f887-nsk75 Binding Scheduled Successfully assigned kuttl-test-ace-dory/my-jaeger-query-7d9b66f887-nsk75 to ip-10-0-79-227.ec2.internal default-scheduler logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:30:12 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-7d9b66f887 SuccessfulCreate Created pod: my-jaeger-query-7d9b66f887-nsk75 replicaset-controller logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:30:12 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-7d9b66f887 to 1 deployment-controller logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:30:13 +0000 UTC Normal Pod my-jaeger-collector-9b5c76b54-fbnb8 AddedInterface Add eth0 [10.129.2.59/23] from ovn-kubernetes multus logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:30:13 +0000 UTC Normal Pod my-jaeger-collector-9b5c76b54-fbnb8.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:44686d560aa27cae8ff8693f88c4cb6e2edf1737010ec1f80709cb42250b729d" already present on machine kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:30:13 +0000 UTC Normal Pod my-jaeger-collector-9b5c76b54-fbnb8.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:30:13 +0000 UTC Normal Pod my-jaeger-collector-9b5c76b54-fbnb8.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:30:13 +0000 UTC Normal Pod my-jaeger-query-7d9b66f887-nsk75 AddedInterface Add eth0 [10.129.2.60/23] from ovn-kubernetes multus logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:30:13 +0000 UTC Normal Pod my-jaeger-query-7d9b66f887-nsk75.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:24a83160aa930e1b72c2a2442a33b28af2b06c24b058e09afcd0d495a8066d6d" already present on machine kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:30:13 +0000 UTC Normal Pod my-jaeger-query-7d9b66f887-nsk75.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:30:13 +0000 UTC Normal Pod my-jaeger-query-7d9b66f887-nsk75.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:30:13 +0000 UTC Normal Pod my-jaeger-query-7d9b66f887-nsk75.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:30:13 +0000 UTC Normal Pod my-jaeger-query-7d9b66f887-nsk75.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:30:13 +0000 UTC Normal Pod my-jaeger-query-7d9b66f887-nsk75.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:30:13 +0000 UTC Normal Pod my-jaeger-query-7d9b66f887-nsk75.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" already present on machine kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:30:13 +0000 UTC Normal Pod my-jaeger-query-7d9b66f887-nsk75.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:30:13 +0000 UTC Normal Pod my-jaeger-query-7d9b66f887-nsk75.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:30:22 +0000 UTC Normal Pod my-jaeger-query-7d9b66f887-nsk75.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:30:22 +0000 UTC Normal Pod my-jaeger-query-7d9b66f887-nsk75.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:30:22 +0000 UTC Normal Pod my-jaeger-query-7d9b66f887-nsk75.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:30:22 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-7d9b66f887 SuccessfulDelete Deleted pod: my-jaeger-query-7d9b66f887-nsk75 replicaset-controller logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:30:22 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled down replica set my-jaeger-query-7d9b66f887 to 0 from 1 deployment-controller logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:30:23 +0000 UTC Normal Pod my-jaeger-query-5c6f58c779-m6zrk Binding Scheduled Successfully assigned kuttl-test-ace-dory/my-jaeger-query-5c6f58c779-m6zrk to ip-10-0-79-227.ec2.internal default-scheduler logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:30:23 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-5c6f58c779 SuccessfulCreate Created pod: my-jaeger-query-5c6f58c779-m6zrk replicaset-controller logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:30:23 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-5c6f58c779 to 1 deployment-controller logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:30:24 +0000 UTC Normal Pod my-jaeger-query-5c6f58c779-m6zrk AddedInterface Add eth0 [10.129.2.61/23] from ovn-kubernetes multus logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:30:24 +0000 UTC Normal Pod my-jaeger-query-5c6f58c779-m6zrk.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:24a83160aa930e1b72c2a2442a33b28af2b06c24b058e09afcd0d495a8066d6d" already present on machine kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:30:24 +0000 UTC Normal Pod my-jaeger-query-5c6f58c779-m6zrk.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:30:24 +0000 UTC Normal Pod my-jaeger-query-5c6f58c779-m6zrk.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:30:24 +0000 UTC Normal Pod my-jaeger-query-5c6f58c779-m6zrk.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:30:24 +0000 UTC Normal Pod my-jaeger-query-5c6f58c779-m6zrk.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:30:24 +0000 UTC Normal Pod my-jaeger-query-5c6f58c779-m6zrk.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:30:24 +0000 UTC Normal Pod my-jaeger-query-5c6f58c779-m6zrk.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" already present on machine kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:30:24 +0000 UTC Normal Pod my-jaeger-query-5c6f58c779-m6zrk.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:30:24 +0000 UTC Normal Pod my-jaeger-query-5c6f58c779-m6zrk.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:30:27 +0000 UTC Normal Pod 00-report-span-9wtv7 Binding Scheduled Successfully assigned kuttl-test-ace-dory/00-report-span-9wtv7 to ip-10-0-39-223.ec2.internal default-scheduler logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:30:27 +0000 UTC Normal Job.batch 00-report-span SuccessfulCreate Created pod: 00-report-span-9wtv7 job-controller logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:30:28 +0000 UTC Normal Pod 00-report-span-9wtv7 AddedInterface Add eth0 [10.128.2.36/23] from ovn-kubernetes multus logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:30:28 +0000 UTC Normal Pod 00-report-span-9wtv7.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:30:29 +0000 UTC Normal Pod 00-report-span-9wtv7.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 1.304s (1.304s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:30:29 +0000 UTC Normal Pod 00-report-span-9wtv7.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:30:29 +0000 UTC Normal Pod 00-report-span-9wtv7.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:30:30 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:30:30 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:30:31 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:30:52 +0000 UTC Normal Job.batch 00-report-span Completed Job completed job-controller logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:30:53 +0000 UTC Normal Pod 00-check-indices-xgbtp Binding Scheduled Successfully assigned kuttl-test-ace-dory/00-check-indices-xgbtp to ip-10-0-39-223.ec2.internal default-scheduler logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:30:53 +0000 UTC Normal Pod 00-check-indices-xgbtp AddedInterface Add eth0 [10.128.2.37/23] from ovn-kubernetes multus logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:30:53 +0000 UTC Normal Pod 00-check-indices-xgbtp.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:30:53 +0000 UTC Normal Job.batch 00-check-indices SuccessfulCreate Created pod: 00-check-indices-xgbtp job-controller logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:30:54 +0000 UTC Normal Pod 00-check-indices-xgbtp.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 829ms (829ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:30:54 +0000 UTC Normal Pod 00-check-indices-xgbtp.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:30:54 +0000 UTC Normal Pod 00-check-indices-xgbtp.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:30:57 +0000 UTC Normal Job.batch 00-check-indices Completed Job completed job-controller logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:30:58 +0000 UTC Normal Pod 01-check-indices-c9dw2 Binding Scheduled Successfully assigned kuttl-test-ace-dory/01-check-indices-c9dw2 to ip-10-0-39-223.ec2.internal default-scheduler logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:30:58 +0000 UTC Normal Pod 01-check-indices-c9dw2 AddedInterface Add eth0 [10.128.2.38/23] from ovn-kubernetes multus logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:30:58 +0000 UTC Normal Pod 01-check-indices-c9dw2.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:30:58 +0000 UTC Normal Job.batch 01-check-indices SuccessfulCreate Created pod: 01-check-indices-c9dw2 job-controller logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:30:59 +0000 UTC Normal Pod 01-check-indices-c9dw2.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 849ms (849ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:30:59 +0000 UTC Normal Pod 01-check-indices-c9dw2.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:30:59 +0000 UTC Normal Pod 01-check-indices-c9dw2.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:31:01 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get cpu utilization: missing request for cpu in container jaeger-collector of Pod my-jaeger-collector-9b5c76b54-fbnb8 horizontal-pod-autoscaler logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:31:01 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get memory utilization: missing request for memory in container jaeger-collector of Pod my-jaeger-collector-9b5c76b54-fbnb8 horizontal-pod-autoscaler logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:31:01 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: missing request for cpu in container jaeger-collector of Pod my-jaeger-collector-9b5c76b54-fbnb8 horizontal-pod-autoscaler logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:31:02 +0000 UTC Normal Job.batch 01-check-indices Completed Job completed job-controller logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:31:04 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-zqmt5 Binding Scheduled Successfully assigned kuttl-test-ace-dory/my-jaeger-es-rollover-create-mapping-zqmt5 to ip-10-0-39-223.ec2.internal default-scheduler logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:31:04 +0000 UTC Normal Job.batch my-jaeger-es-rollover-create-mapping SuccessfulCreate Created pod: my-jaeger-es-rollover-create-mapping-zqmt5 job-controller logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:31:05 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-zqmt5 AddedInterface Add eth0 [10.128.2.39/23] from ovn-kubernetes multus logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:31:05 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-zqmt5.spec.containers{my-jaeger-es-rollover-create-mapping} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:44283d1da2e4839c9a23cf2848ddc67d853a6d1fe532e04c642a6ecee08b25cb" kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:31:08 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-zqmt5.spec.containers{my-jaeger-es-rollover-create-mapping} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:44283d1da2e4839c9a23cf2848ddc67d853a6d1fe532e04c642a6ecee08b25cb" in 3.454s (3.454s including waiting). Image size: 111742913 bytes. kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:31:08 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-zqmt5.spec.containers{my-jaeger-es-rollover-create-mapping} Created Created container my-jaeger-es-rollover-create-mapping kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:31:09 +0000 UTC Normal Pod my-jaeger-es-rollover-create-mapping-zqmt5.spec.containers{my-jaeger-es-rollover-create-mapping} Started Started container my-jaeger-es-rollover-create-mapping kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:31:11 +0000 UTC Normal Pod my-jaeger-collector-9b5c76b54-fbnb8.spec.containers{jaeger-collector} Killing Stopping container jaeger-collector kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:31:11 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-9b5c76b54 SuccessfulDelete Deleted pod: my-jaeger-collector-9b5c76b54-fbnb8 replicaset-controller logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:31:11 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled down replica set my-jaeger-collector-9b5c76b54 to 0 from 1 deployment-controller logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:31:11 +0000 UTC Normal Job.batch my-jaeger-es-rollover-create-mapping Completed Job completed job-controller logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:31:11 +0000 UTC Normal Pod my-jaeger-query-5c6f58c779-m6zrk.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:31:11 +0000 UTC Normal Pod my-jaeger-query-5c6f58c779-m6zrk.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:31:11 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-5c6f58c779 SuccessfulDelete Deleted pod: my-jaeger-query-5c6f58c779-m6zrk replicaset-controller logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:31:11 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled down replica set my-jaeger-query-5c6f58c779 to 0 from 1 deployment-controller logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:31:12 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-f69797d79 to 1 deployment-controller logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:31:12 +0000 UTC Normal Pod my-jaeger-query-5c6f58c779-m6zrk.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:31:13 +0000 UTC Normal Pod my-jaeger-collector-f69797d79-789zx Binding Scheduled Successfully assigned kuttl-test-ace-dory/my-jaeger-collector-f69797d79-789zx to ip-10-0-79-227.ec2.internal default-scheduler logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:31:13 +0000 UTC Normal Pod my-jaeger-collector-f69797d79-789zx AddedInterface Add eth0 [10.129.2.62/23] from ovn-kubernetes multus logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:31:13 +0000 UTC Normal Pod my-jaeger-collector-f69797d79-789zx.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:44686d560aa27cae8ff8693f88c4cb6e2edf1737010ec1f80709cb42250b729d" already present on machine kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:31:13 +0000 UTC Normal Pod my-jaeger-collector-f69797d79-789zx.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:31:13 +0000 UTC Normal Pod my-jaeger-collector-f69797d79-789zx.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:31:13 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-f69797d79 SuccessfulCreate Created pod: my-jaeger-collector-f69797d79-789zx replicaset-controller logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:31:13 +0000 UTC Normal Pod my-jaeger-query-7975b5bb54-47pgl Binding Scheduled Successfully assigned kuttl-test-ace-dory/my-jaeger-query-7975b5bb54-47pgl to ip-10-0-79-227.ec2.internal default-scheduler logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:31:13 +0000 UTC Normal Pod my-jaeger-query-7975b5bb54-47pgl AddedInterface Add eth0 [10.129.2.63/23] from ovn-kubernetes multus logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:31:13 +0000 UTC Normal Pod my-jaeger-query-7975b5bb54-47pgl.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:24a83160aa930e1b72c2a2442a33b28af2b06c24b058e09afcd0d495a8066d6d" already present on machine kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:31:13 +0000 UTC Normal Pod my-jaeger-query-7975b5bb54-47pgl.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:31:13 +0000 UTC Normal Pod my-jaeger-query-7975b5bb54-47pgl.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:31:13 +0000 UTC Normal Pod my-jaeger-query-7975b5bb54-47pgl.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:31:13 +0000 UTC Normal Pod my-jaeger-query-7975b5bb54-47pgl.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:31:13 +0000 UTC Normal Pod my-jaeger-query-7975b5bb54-47pgl.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:31:13 +0000 UTC Normal Pod my-jaeger-query-7975b5bb54-47pgl.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" already present on machine kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:31:13 +0000 UTC Normal Pod my-jaeger-query-7975b5bb54-47pgl.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:31:13 +0000 UTC Normal Pod my-jaeger-query-7975b5bb54-47pgl.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:31:13 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-7975b5bb54 SuccessfulCreate Created pod: my-jaeger-query-7975b5bb54-47pgl replicaset-controller logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:31:13 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-7975b5bb54 to 1 deployment-controller logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:31:20 +0000 UTC Normal Pod 02-report-span-g4x9x Binding Scheduled Successfully assigned kuttl-test-ace-dory/02-report-span-g4x9x to ip-10-0-39-223.ec2.internal default-scheduler logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:31:20 +0000 UTC Normal Job.batch 02-report-span SuccessfulCreate Created pod: 02-report-span-g4x9x job-controller logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:31:21 +0000 UTC Normal Pod 02-report-span-g4x9x AddedInterface Add eth0 [10.128.2.40/23] from ovn-kubernetes multus logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:31:21 +0000 UTC Normal Pod 02-report-span-g4x9x.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:31:22 +0000 UTC Normal Pod 02-report-span-g4x9x.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 1.004s (1.004s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:31:22 +0000 UTC Normal Pod 02-report-span-g4x9x.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:31:22 +0000 UTC Normal Pod 02-report-span-g4x9x.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:31:44 +0000 UTC Normal Job.batch 02-report-span Completed Job completed job-controller logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:31:45 +0000 UTC Normal Pod 02-check-indices-lb5t8 Binding Scheduled Successfully assigned kuttl-test-ace-dory/02-check-indices-lb5t8 to ip-10-0-39-223.ec2.internal default-scheduler logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:31:45 +0000 UTC Normal Pod 02-check-indices-lb5t8 AddedInterface Add eth0 [10.128.2.41/23] from ovn-kubernetes multus logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:31:45 +0000 UTC Normal Pod 02-check-indices-lb5t8.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:31:45 +0000 UTC Normal Job.batch 02-check-indices SuccessfulCreate Created pod: 02-check-indices-lb5t8 job-controller logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:31:46 +0000 UTC Normal Pod 02-check-indices-lb5t8.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 931ms (931ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:31:46 +0000 UTC Normal Pod 02-check-indices-lb5t8.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:31:46 +0000 UTC Normal Pod 02-check-indices-lb5t8.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:31:49 +0000 UTC Normal Job.batch 02-check-indices Completed Job completed job-controller logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:31:50 +0000 UTC Normal Pod 03-check-indices-8bdnz Binding Scheduled Successfully assigned kuttl-test-ace-dory/03-check-indices-8bdnz to ip-10-0-39-223.ec2.internal default-scheduler logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:31:50 +0000 UTC Normal Pod 03-check-indices-8bdnz AddedInterface Add eth0 [10.128.2.42/23] from ovn-kubernetes multus logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:31:50 +0000 UTC Normal Pod 03-check-indices-8bdnz.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:31:50 +0000 UTC Normal Job.batch 03-check-indices SuccessfulCreate Created pod: 03-check-indices-8bdnz job-controller logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:31:51 +0000 UTC Normal Pod 03-check-indices-8bdnz.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 1.128s (1.128s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:31:52 +0000 UTC Normal Pod 03-check-indices-8bdnz.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:31:52 +0000 UTC Normal Pod 03-check-indices-8bdnz.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:31:54 +0000 UTC Normal Job.batch 03-check-indices Completed Job completed job-controller logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:31:55 +0000 UTC Normal Pod 04-check-indices-x92fg Binding Scheduled Successfully assigned kuttl-test-ace-dory/04-check-indices-x92fg to ip-10-0-39-223.ec2.internal default-scheduler logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:31:55 +0000 UTC Normal Job.batch 04-check-indices SuccessfulCreate Created pod: 04-check-indices-x92fg job-controller logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:31:56 +0000 UTC Normal Pod 04-check-indices-x92fg AddedInterface Add eth0 [10.128.2.43/23] from ovn-kubernetes multus logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:31:56 +0000 UTC Normal Pod 04-check-indices-x92fg.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:31:56 +0000 UTC Normal Pod 04-check-indices-x92fg.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 478ms (478ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:31:56 +0000 UTC Normal Pod 04-check-indices-x92fg.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:31:56 +0000 UTC Normal Pod 04-check-indices-x92fg.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:31:58 +0000 UTC Normal Job.batch 04-check-indices Completed Job completed job-controller logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:32:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28895492-54wk9 Binding Scheduled Successfully assigned kuttl-test-ace-dory/my-jaeger-es-lookback-28895492-54wk9 to ip-10-0-39-223.ec2.internal default-scheduler logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:32:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28895492-54wk9 AddedInterface Add eth0 [10.128.2.45/23] from ovn-kubernetes multus logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:32:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28895492-54wk9.spec.containers{my-jaeger-es-lookback} Pulled Container image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:44283d1da2e4839c9a23cf2848ddc67d853a6d1fe532e04c642a6ecee08b25cb" already present on machine kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:32:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28895492-54wk9.spec.containers{my-jaeger-es-lookback} Created Created container my-jaeger-es-lookback kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:32:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28895492-54wk9.spec.containers{my-jaeger-es-lookback} Started Started container my-jaeger-es-lookback kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:32:00 +0000 UTC Normal Job.batch my-jaeger-es-lookback-28895492 SuccessfulCreate Created pod: my-jaeger-es-lookback-28895492-54wk9 job-controller logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:32:00 +0000 UTC Normal CronJob.batch my-jaeger-es-lookback SuccessfulCreate Created job my-jaeger-es-lookback-28895492 cronjob-controller logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:32:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28895492-9scns Binding Scheduled Successfully assigned kuttl-test-ace-dory/my-jaeger-es-rollover-28895492-9scns to ip-10-0-39-223.ec2.internal default-scheduler logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:32:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28895492-9scns AddedInterface Add eth0 [10.128.2.44/23] from ovn-kubernetes multus logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:32:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28895492-9scns.spec.containers{my-jaeger-es-rollover} Pulled Container image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:44283d1da2e4839c9a23cf2848ddc67d853a6d1fe532e04c642a6ecee08b25cb" already present on machine kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:32:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28895492-9scns.spec.containers{my-jaeger-es-rollover} Created Created container my-jaeger-es-rollover kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:32:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28895492-9scns.spec.containers{my-jaeger-es-rollover} Started Started container my-jaeger-es-rollover kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:32:00 +0000 UTC Normal Job.batch my-jaeger-es-rollover-28895492 SuccessfulCreate Created pod: my-jaeger-es-rollover-28895492-9scns job-controller logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:32:00 +0000 UTC Normal CronJob.batch my-jaeger-es-rollover SuccessfulCreate Created job my-jaeger-es-rollover-28895492 cronjob-controller logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:32:01 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get cpu utilization: missing request for cpu in container jaeger-collector of Pod my-jaeger-collector-f69797d79-789zx horizontal-pod-autoscaler logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:32:01 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedGetResourceMetric failed to get memory utilization: missing request for memory in container jaeger-collector of Pod my-jaeger-collector-f69797d79-789zx horizontal-pod-autoscaler logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:32:01 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling my-jaeger-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: missing request for cpu in container jaeger-collector of Pod my-jaeger-collector-f69797d79-789zx horizontal-pod-autoscaler logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:32:03 +0000 UTC Normal Job.batch my-jaeger-es-rollover-28895492 Completed Job completed job-controller logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:32:03 +0000 UTC Normal CronJob.batch my-jaeger-es-rollover SawCompletedJob Saw completed job: my-jaeger-es-rollover-28895492, condition: Complete cronjob-controller logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:32:04 +0000 UTC Normal Job.batch my-jaeger-es-lookback-28895492 Completed Job completed job-controller logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:32:04 +0000 UTC Normal CronJob.batch my-jaeger-es-lookback SawCompletedJob Saw completed job: my-jaeger-es-lookback-28895492, condition: Complete cronjob-controller logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:32:08 +0000 UTC Normal Pod 03-report-span-47f5z Binding Scheduled Successfully assigned kuttl-test-ace-dory/03-report-span-47f5z to ip-10-0-39-223.ec2.internal default-scheduler logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:32:08 +0000 UTC Normal Pod 03-report-span-47f5z AddedInterface Add eth0 [10.128.2.46/23] from ovn-kubernetes multus logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:32:08 +0000 UTC Normal Pod 03-report-span-47f5z.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:32:08 +0000 UTC Normal Job.batch 03-report-span SuccessfulCreate Created pod: 03-report-span-47f5z job-controller logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:32:10 +0000 UTC Normal Pod 03-report-span-47f5z.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 1.397s (1.397s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:32:10 +0000 UTC Normal Pod 03-report-span-47f5z.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:32:10 +0000 UTC Normal Pod 03-report-span-47f5z.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:32:32 +0000 UTC Normal Job.batch 03-report-span Completed Job completed job-controller logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:33:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28895493-hhx77 Binding Scheduled Successfully assigned kuttl-test-ace-dory/my-jaeger-es-lookback-28895493-hhx77 to ip-10-0-39-223.ec2.internal default-scheduler logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:33:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28895493-hhx77 AddedInterface Add eth0 [10.128.2.48/23] from ovn-kubernetes multus logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:33:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28895493-hhx77.spec.containers{my-jaeger-es-lookback} Pulled Container image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:44283d1da2e4839c9a23cf2848ddc67d853a6d1fe532e04c642a6ecee08b25cb" already present on machine kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:33:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28895493-hhx77.spec.containers{my-jaeger-es-lookback} Created Created container my-jaeger-es-lookback kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:33:00 +0000 UTC Normal Pod my-jaeger-es-lookback-28895493-hhx77.spec.containers{my-jaeger-es-lookback} Started Started container my-jaeger-es-lookback kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:33:00 +0000 UTC Normal Job.batch my-jaeger-es-lookback-28895493 SuccessfulCreate Created pod: my-jaeger-es-lookback-28895493-hhx77 job-controller logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:33:00 +0000 UTC Normal CronJob.batch my-jaeger-es-lookback SuccessfulCreate Created job my-jaeger-es-lookback-28895493 cronjob-controller logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:33:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28895493-h4sn2 Binding Scheduled Successfully assigned kuttl-test-ace-dory/my-jaeger-es-rollover-28895493-h4sn2 to ip-10-0-39-223.ec2.internal default-scheduler logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:33:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28895493-h4sn2 AddedInterface Add eth0 [10.128.2.47/23] from ovn-kubernetes multus logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:33:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28895493-h4sn2.spec.containers{my-jaeger-es-rollover} Pulled Container image "registry.redhat.io/rhosdt/jaeger-es-rollover-rhel8@sha256:44283d1da2e4839c9a23cf2848ddc67d853a6d1fe532e04c642a6ecee08b25cb" already present on machine kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:33:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28895493-h4sn2.spec.containers{my-jaeger-es-rollover} Created Created container my-jaeger-es-rollover kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:33:00 +0000 UTC Normal Pod my-jaeger-es-rollover-28895493-h4sn2.spec.containers{my-jaeger-es-rollover} Started Started container my-jaeger-es-rollover kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:33:00 +0000 UTC Normal Job.batch my-jaeger-es-rollover-28895493 SuccessfulCreate Created pod: my-jaeger-es-rollover-28895493-h4sn2 job-controller logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:33:00 +0000 UTC Normal CronJob.batch my-jaeger-es-rollover SuccessfulCreate Created job my-jaeger-es-rollover-28895493 cronjob-controller logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:33:04 +0000 UTC Normal Job.batch my-jaeger-es-lookback-28895493 Completed Job completed job-controller logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:33:04 +0000 UTC Normal CronJob.batch my-jaeger-es-lookback SawCompletedJob Saw completed job: my-jaeger-es-lookback-28895493, condition: Complete cronjob-controller logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:33:04 +0000 UTC Normal Job.batch my-jaeger-es-rollover-28895493 Completed Job completed job-controller logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:33:04 +0000 UTC Normal CronJob.batch my-jaeger-es-rollover SawCompletedJob Saw completed job: my-jaeger-es-rollover-28895493, condition: Complete cronjob-controller logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:33:14 +0000 UTC Normal Pod 05-check-indices-h75kv Binding Scheduled Successfully assigned kuttl-test-ace-dory/05-check-indices-h75kv to ip-10-0-39-223.ec2.internal default-scheduler logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:33:14 +0000 UTC Normal Pod 05-check-indices-h75kv AddedInterface Add eth0 [10.128.2.49/23] from ovn-kubernetes multus logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:33:14 +0000 UTC Normal Pod 05-check-indices-h75kv.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:33:14 +0000 UTC Normal Job.batch 05-check-indices SuccessfulCreate Created pod: 05-check-indices-h75kv job-controller logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:33:15 +0000 UTC Normal Pod 05-check-indices-h75kv.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 734ms (734ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:33:15 +0000 UTC Normal Pod 05-check-indices-h75kv.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:33:15 +0000 UTC Normal Pod 05-check-indices-h75kv.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:33:18 +0000 UTC Normal Job.batch 05-check-indices Completed Job completed job-controller logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:33:18 +0000 UTC Normal Pod 06-check-indices-5knls Binding Scheduled Successfully assigned kuttl-test-ace-dory/06-check-indices-5knls to ip-10-0-39-223.ec2.internal default-scheduler logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:33:18 +0000 UTC Normal Pod 06-check-indices-5knls AddedInterface Add eth0 [10.128.2.50/23] from ovn-kubernetes multus logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:33:18 +0000 UTC Normal Pod 06-check-indices-5knls.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:33:18 +0000 UTC Normal Job.batch 06-check-indices SuccessfulCreate Created pod: 06-check-indices-5knls job-controller logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:33:20 +0000 UTC Normal Pod 06-check-indices-5knls.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 1.291s (1.291s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:33:20 +0000 UTC Normal Pod 06-check-indices-5knls.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:33:20 +0000 UTC Normal Pod 06-check-indices-5knls.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:33:23 | es-rollover-autoprov | 2024-12-09 07:33:23 +0000 UTC Normal Job.batch 06-check-indices Completed Job completed job-controller logger.go:42: 07:33:23 | es-rollover-autoprov | Deleting namespace: kuttl-test-ace-dory === CONT kuttl/harness/es-increasing-replicas logger.go:42: 07:33:31 | es-increasing-replicas | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:33:31 | es-increasing-replicas | Ignoring check-es-nodes.sh as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:33:31 | es-increasing-replicas | Ignoring openshift-check-es-nodes.yaml.template as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:33:31 | es-increasing-replicas | Creating namespace: kuttl-test-faithful-stallion logger.go:42: 07:33:31 | es-increasing-replicas/1-install | starting test step 1-install logger.go:42: 07:33:31 | es-increasing-replicas/1-install | Jaeger:kuttl-test-faithful-stallion/simple-prod created logger.go:42: 07:34:08 | es-increasing-replicas/1-install | test step completed 1-install logger.go:42: 07:34:08 | es-increasing-replicas/2-install | starting test step 2-install logger.go:42: 07:34:08 | es-increasing-replicas/2-install | Jaeger:kuttl-test-faithful-stallion/simple-prod updated logger.go:42: 07:34:15 | es-increasing-replicas/2-install | test step completed 2-install logger.go:42: 07:34:15 | es-increasing-replicas/3-smoke-test | starting test step 3-smoke-test logger.go:42: 07:34:15 | es-increasing-replicas/3-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE simple-prod /dev/null] logger.go:42: 07:34:17 | es-increasing-replicas/3-smoke-test | Warning: resource jaegers/simple-prod is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:34:23 | es-increasing-replicas/3-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:34:24 | es-increasing-replicas/3-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:34:24 | es-increasing-replicas/3-smoke-test | job.batch/report-span created logger.go:42: 07:34:24 | es-increasing-replicas/3-smoke-test | job.batch/check-span created logger.go:42: 07:34:36 | es-increasing-replicas/3-smoke-test | test step completed 3-smoke-test logger.go:42: 07:34:36 | es-increasing-replicas/4-install | starting test step 4-install logger.go:42: 07:34:36 | es-increasing-replicas/4-install | Jaeger:kuttl-test-faithful-stallion/simple-prod updated logger.go:42: 07:34:36 | es-increasing-replicas/4-install | test step completed 4-install logger.go:42: 07:34:36 | es-increasing-replicas/5-check-es-nodes | starting test step 5-check-es-nodes logger.go:42: 07:34:36 | es-increasing-replicas/5-check-es-nodes | running command: [sh -c ./check-es-nodes.sh $NAMESPACE] logger.go:42: 07:34:36 | es-increasing-replicas/5-check-es-nodes | Checking if the number of ES instances is the expected logger.go:42: 07:34:36 | es-increasing-replicas/5-check-es-nodes | false logger.go:42: 07:34:36 | es-increasing-replicas/5-check-es-nodes | Error: no matches found logger.go:42: 07:34:41 | es-increasing-replicas/5-check-es-nodes | Checking if the number of ES instances is the expected logger.go:42: 07:34:41 | es-increasing-replicas/5-check-es-nodes | true logger.go:42: 07:34:42 | es-increasing-replicas/5-check-es-nodes | test step completed 5-check-es-nodes logger.go:42: 07:34:42 | es-increasing-replicas | es-increasing-replicas events from ns kuttl-test-faithful-stallion: logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:33:37 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestfaithfulstallionsimpleprod-1-5f5f8c4cf9 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestfaithfulstallionsimpleprod-1-5fpnskq replicaset-controller logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:33:37 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfaithfulstallionsimpleprod-1-5fpnskq Binding Scheduled Successfully assigned kuttl-test-faithful-stallion/elasticsearch-cdm-kuttltestfaithfulstallionsimpleprod-1-5fpnskq to ip-10-0-95-250.ec2.internal default-scheduler logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:33:37 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestfaithfulstallionsimpleprod-1-5fpnskq FailedMount MountVolume.SetUp failed for volume "elasticsearch-metrics" : secret "elasticsearch-metrics" not found kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:33:37 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestfaithfulstallionsimpleprod-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestfaithfulstallionsimpleprod-1-5f5f8c4cf9 to 1 deployment-controller logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:33:38 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfaithfulstallionsimpleprod-1-5fpnskq AddedInterface Add eth0 [10.131.0.35/23] from ovn-kubernetes multus logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:33:38 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfaithfulstallionsimpleprod-1-5fpnskq.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:8e4fbea4983cd58352349ca291383169b286bc166fad95a87807552ca43335e6" already present on machine kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:33:38 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfaithfulstallionsimpleprod-1-5fpnskq.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:33:38 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfaithfulstallionsimpleprod-1-5fpnskq.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:33:38 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfaithfulstallionsimpleprod-1-5fpnskq.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:d68824b0b2c84db8e33edf9ab344eb684c4a7ebd7ef162bbc309043adcb28e6b" already present on machine kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:33:38 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfaithfulstallionsimpleprod-1-5fpnskq.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:33:38 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfaithfulstallionsimpleprod-1-5fpnskq.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:33:48 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestfaithfulstallionsimpleprod-1-5fpnskq.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:33:53 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestfaithfulstallionsimpleprod-1-5fpnskq.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:04 +0000 UTC Normal Pod simple-prod-collector-fc8c6659b-9j9qd Binding Scheduled Successfully assigned kuttl-test-faithful-stallion/simple-prod-collector-fc8c6659b-9j9qd to ip-10-0-79-227.ec2.internal default-scheduler logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:04 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-fc8c6659b SuccessfulCreate Created pod: simple-prod-collector-fc8c6659b-9j9qd replicaset-controller logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:04 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-fc8c6659b to 1 deployment-controller logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:04 +0000 UTC Normal Pod simple-prod-query-cc84fbd54-d5zkf Binding Scheduled Successfully assigned kuttl-test-faithful-stallion/simple-prod-query-cc84fbd54-d5zkf to ip-10-0-79-227.ec2.internal default-scheduler logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:04 +0000 UTC Normal ReplicaSet.apps simple-prod-query-cc84fbd54 SuccessfulCreate Created pod: simple-prod-query-cc84fbd54-d5zkf replicaset-controller logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:04 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-cc84fbd54 to 1 deployment-controller logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:05 +0000 UTC Normal Pod simple-prod-collector-fc8c6659b-9j9qd AddedInterface Add eth0 [10.129.2.64/23] from ovn-kubernetes multus logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:05 +0000 UTC Normal Pod simple-prod-collector-fc8c6659b-9j9qd.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:44686d560aa27cae8ff8693f88c4cb6e2edf1737010ec1f80709cb42250b729d" already present on machine kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:05 +0000 UTC Normal Pod simple-prod-collector-fc8c6659b-9j9qd.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:05 +0000 UTC Normal Pod simple-prod-collector-fc8c6659b-9j9qd.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:05 +0000 UTC Normal Pod simple-prod-query-cc84fbd54-d5zkf AddedInterface Add eth0 [10.129.2.65/23] from ovn-kubernetes multus logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:05 +0000 UTC Normal Pod simple-prod-query-cc84fbd54-d5zkf.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:24a83160aa930e1b72c2a2442a33b28af2b06c24b058e09afcd0d495a8066d6d" already present on machine kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:05 +0000 UTC Normal Pod simple-prod-query-cc84fbd54-d5zkf.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:05 +0000 UTC Normal Pod simple-prod-query-cc84fbd54-d5zkf.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:05 +0000 UTC Normal Pod simple-prod-query-cc84fbd54-d5zkf.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:05 +0000 UTC Normal Pod simple-prod-query-cc84fbd54-d5zkf.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:05 +0000 UTC Normal Pod simple-prod-query-cc84fbd54-d5zkf.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:05 +0000 UTC Normal Pod simple-prod-query-cc84fbd54-d5zkf.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" already present on machine kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:05 +0000 UTC Normal Pod simple-prod-query-cc84fbd54-d5zkf.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:05 +0000 UTC Normal Pod simple-prod-query-cc84fbd54-d5zkf.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:12 +0000 UTC Normal Pod simple-prod-collector-fc8c6659b-8gj8l Binding Scheduled Successfully assigned kuttl-test-faithful-stallion/simple-prod-collector-fc8c6659b-8gj8l to ip-10-0-39-223.ec2.internal default-scheduler logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:12 +0000 UTC Normal Pod simple-prod-collector-fc8c6659b-8gj8l AddedInterface Add eth0 [10.128.2.51/23] from ovn-kubernetes multus logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:12 +0000 UTC Normal Pod simple-prod-collector-fc8c6659b-8gj8l.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:44686d560aa27cae8ff8693f88c4cb6e2edf1737010ec1f80709cb42250b729d" already present on machine kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:12 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-fc8c6659b SuccessfulCreate Created pod: simple-prod-collector-fc8c6659b-8gj8l replicaset-controller logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:12 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-fc8c6659b to 2 from 1 deployment-controller logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:12 +0000 UTC Normal Pod simple-prod-query-cc84fbd54-rsh7d Binding Scheduled Successfully assigned kuttl-test-faithful-stallion/simple-prod-query-cc84fbd54-rsh7d to ip-10-0-39-223.ec2.internal default-scheduler logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:12 +0000 UTC Normal Pod simple-prod-query-cc84fbd54-rsh7d AddedInterface Add eth0 [10.128.2.52/23] from ovn-kubernetes multus logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:12 +0000 UTC Normal ReplicaSet.apps simple-prod-query-cc84fbd54 SuccessfulCreate Created pod: simple-prod-query-cc84fbd54-rsh7d replicaset-controller logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:12 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-cc84fbd54 to 2 from 1 deployment-controller logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:13 +0000 UTC Normal Pod simple-prod-collector-fc8c6659b-8gj8l.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:13 +0000 UTC Normal Pod simple-prod-collector-fc8c6659b-8gj8l.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:13 +0000 UTC Normal Pod simple-prod-query-cc84fbd54-rsh7d.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:24a83160aa930e1b72c2a2442a33b28af2b06c24b058e09afcd0d495a8066d6d" already present on machine kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:13 +0000 UTC Normal Pod simple-prod-query-cc84fbd54-rsh7d.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:13 +0000 UTC Normal Pod simple-prod-query-cc84fbd54-rsh7d.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:13 +0000 UTC Normal Pod simple-prod-query-cc84fbd54-rsh7d.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:13 +0000 UTC Normal Pod simple-prod-query-cc84fbd54-rsh7d.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:13 +0000 UTC Normal Pod simple-prod-query-cc84fbd54-rsh7d.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:13 +0000 UTC Normal Pod simple-prod-query-cc84fbd54-rsh7d.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" already present on machine kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:13 +0000 UTC Normal Pod simple-prod-query-cc84fbd54-rsh7d.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:13 +0000 UTC Normal Pod simple-prod-query-cc84fbd54-rsh7d.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:18 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled down replica set simple-prod-query-cc84fbd54 to 0 from 2 deployment-controller logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:19 +0000 UTC Normal Pod simple-prod-query-cc84fbd54-d5zkf.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:19 +0000 UTC Normal Pod simple-prod-query-cc84fbd54-d5zkf.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:19 +0000 UTC Normal Pod simple-prod-query-cc84fbd54-d5zkf.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:19 +0000 UTC Normal Pod simple-prod-query-cc84fbd54-rsh7d.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:19 +0000 UTC Normal Pod simple-prod-query-cc84fbd54-rsh7d.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:19 +0000 UTC Normal Pod simple-prod-query-cc84fbd54-rsh7d.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:19 +0000 UTC Normal ReplicaSet.apps simple-prod-query-cc84fbd54 SuccessfulDelete Deleted pod: simple-prod-query-cc84fbd54-d5zkf replicaset-controller logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:19 +0000 UTC Normal ReplicaSet.apps simple-prod-query-cc84fbd54 SuccessfulDelete Deleted pod: simple-prod-query-cc84fbd54-rsh7d replicaset-controller logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:20 +0000 UTC Normal Pod simple-prod-query-866c9f7876-qt794 Binding Scheduled Successfully assigned kuttl-test-faithful-stallion/simple-prod-query-866c9f7876-qt794 to ip-10-0-79-227.ec2.internal default-scheduler logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:20 +0000 UTC Normal Pod simple-prod-query-866c9f7876-qt794 AddedInterface Add eth0 [10.129.2.66/23] from ovn-kubernetes multus logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:20 +0000 UTC Normal Pod simple-prod-query-866c9f7876-qt794.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:24a83160aa930e1b72c2a2442a33b28af2b06c24b058e09afcd0d495a8066d6d" already present on machine kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:20 +0000 UTC Normal Pod simple-prod-query-866c9f7876-qt794.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:20 +0000 UTC Normal Pod simple-prod-query-866c9f7876-qt794.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:20 +0000 UTC Normal Pod simple-prod-query-866c9f7876-qt794.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:20 +0000 UTC Normal Pod simple-prod-query-866c9f7876-qt794.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:20 +0000 UTC Normal Pod simple-prod-query-866c9f7876-w6mn6 Binding Scheduled Successfully assigned kuttl-test-faithful-stallion/simple-prod-query-866c9f7876-w6mn6 to ip-10-0-39-223.ec2.internal default-scheduler logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:20 +0000 UTC Normal Pod simple-prod-query-866c9f7876-w6mn6 AddedInterface Add eth0 [10.128.2.53/23] from ovn-kubernetes multus logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:20 +0000 UTC Normal Pod simple-prod-query-866c9f7876-w6mn6.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:24a83160aa930e1b72c2a2442a33b28af2b06c24b058e09afcd0d495a8066d6d" already present on machine kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:20 +0000 UTC Normal Pod simple-prod-query-866c9f7876-w6mn6.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:20 +0000 UTC Normal Pod simple-prod-query-866c9f7876-w6mn6.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:20 +0000 UTC Normal Pod simple-prod-query-866c9f7876-w6mn6.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:20 +0000 UTC Normal ReplicaSet.apps simple-prod-query-866c9f7876 SuccessfulCreate Created pod: simple-prod-query-866c9f7876-qt794 replicaset-controller logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:20 +0000 UTC Normal ReplicaSet.apps simple-prod-query-866c9f7876 SuccessfulCreate Created pod: simple-prod-query-866c9f7876-w6mn6 replicaset-controller logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:20 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-866c9f7876 to 2 deployment-controller logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:21 +0000 UTC Normal Pod simple-prod-query-866c9f7876-qt794.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:21 +0000 UTC Normal Pod simple-prod-query-866c9f7876-qt794.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" already present on machine kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:21 +0000 UTC Normal Pod simple-prod-query-866c9f7876-qt794.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:21 +0000 UTC Normal Pod simple-prod-query-866c9f7876-qt794.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:21 +0000 UTC Normal Pod simple-prod-query-866c9f7876-w6mn6.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:21 +0000 UTC Normal Pod simple-prod-query-866c9f7876-w6mn6.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:21 +0000 UTC Normal Pod simple-prod-query-866c9f7876-w6mn6.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" already present on machine kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:21 +0000 UTC Normal Pod simple-prod-query-866c9f7876-w6mn6.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:21 +0000 UTC Normal Pod simple-prod-query-866c9f7876-w6mn6.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:24 +0000 UTC Normal Pod check-span-8wg25 Binding Scheduled Successfully assigned kuttl-test-faithful-stallion/check-span-8wg25 to ip-10-0-39-223.ec2.internal default-scheduler logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:24 +0000 UTC Normal Pod check-span-8wg25 AddedInterface Add eth0 [10.128.2.54/23] from ovn-kubernetes multus logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:24 +0000 UTC Normal Pod check-span-8wg25.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:24 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-8wg25 job-controller logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:24 +0000 UTC Normal Pod report-span-d5ttc Binding Scheduled Successfully assigned kuttl-test-faithful-stallion/report-span-d5ttc to ip-10-0-79-227.ec2.internal default-scheduler logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:24 +0000 UTC Normal Pod report-span-d5ttc AddedInterface Add eth0 [10.129.2.67/23] from ovn-kubernetes multus logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:24 +0000 UTC Normal Pod report-span-d5ttc.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:24 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-d5ttc job-controller logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:25 +0000 UTC Normal Pod check-span-8wg25.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 445ms (445ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:25 +0000 UTC Normal Pod check-span-8wg25.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:25 +0000 UTC Normal Pod check-span-8wg25.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:25 +0000 UTC Normal Pod report-span-d5ttc.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 498ms (498ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:25 +0000 UTC Normal Pod report-span-d5ttc.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:25 +0000 UTC Normal Pod report-span-d5ttc.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:36 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:38 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestfaithfulstallionsimpleprod-2-7b4dc5449 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestfaithfulstallionsimpleprod-2-7b9mwvb replicaset-controller logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:38 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfaithfulstallionsimpleprod-2-7b9mwvb Binding Scheduled Successfully assigned kuttl-test-faithful-stallion/elasticsearch-cdm-kuttltestfaithfulstallionsimpleprod-2-7b9mwvb to ip-10-0-39-223.ec2.internal default-scheduler logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:38 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfaithfulstallionsimpleprod-2-7b9mwvb AddedInterface Add eth0 [10.128.2.55/23] from ovn-kubernetes multus logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:38 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfaithfulstallionsimpleprod-2-7b9mwvb.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:8e4fbea4983cd58352349ca291383169b286bc166fad95a87807552ca43335e6" already present on machine kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:38 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestfaithfulstallionsimpleprod-2 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestfaithfulstallionsimpleprod-2-7b4dc5449 to 1 deployment-controller logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:38 +0000 UTC Normal Pod simple-prod-collector-fc8c6659b-8gj8l.spec.containers{jaeger-collector} Killing Stopping container jaeger-collector kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:38 +0000 UTC Normal Pod simple-prod-collector-fc8c6659b-9j9qd.spec.containers{jaeger-collector} Killing Stopping container jaeger-collector kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:38 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-fc8c6659b SuccessfulDelete Deleted pod: simple-prod-collector-fc8c6659b-8gj8l replicaset-controller logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:38 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-fc8c6659b SuccessfulDelete Deleted pod: simple-prod-collector-fc8c6659b-9j9qd replicaset-controller logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:38 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled down replica set simple-prod-collector-fc8c6659b to 0 from 2 deployment-controller logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:38 +0000 UTC Normal Pod simple-prod-query-866c9f7876-qt794.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:38 +0000 UTC Normal Pod simple-prod-query-866c9f7876-qt794.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:38 +0000 UTC Normal Pod simple-prod-query-866c9f7876-qt794.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:38 +0000 UTC Normal Pod simple-prod-query-866c9f7876-w6mn6.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:38 +0000 UTC Normal Pod simple-prod-query-866c9f7876-w6mn6.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:38 +0000 UTC Normal Pod simple-prod-query-866c9f7876-w6mn6.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:38 +0000 UTC Normal ReplicaSet.apps simple-prod-query-866c9f7876 SuccessfulDelete Deleted pod: simple-prod-query-866c9f7876-w6mn6 replicaset-controller logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:38 +0000 UTC Normal ReplicaSet.apps simple-prod-query-866c9f7876 SuccessfulDelete Deleted pod: simple-prod-query-866c9f7876-qt794 replicaset-controller logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:38 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled down replica set simple-prod-query-866c9f7876 to 0 from 2 deployment-controller logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:39 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfaithfulstallionsimpleprod-2-7b9mwvb.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:39 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfaithfulstallionsimpleprod-2-7b9mwvb.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:39 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfaithfulstallionsimpleprod-2-7b9mwvb.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:d68824b0b2c84db8e33edf9ab344eb684c4a7ebd7ef162bbc309043adcb28e6b" already present on machine kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:39 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfaithfulstallionsimpleprod-2-7b9mwvb.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:39 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestfaithfulstallionsimpleprod-2-7b9mwvb.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:39 +0000 UTC Normal Pod simple-prod-collector-6d88dc668c-2gkfx Binding Scheduled Successfully assigned kuttl-test-faithful-stallion/simple-prod-collector-6d88dc668c-2gkfx to ip-10-0-79-227.ec2.internal default-scheduler logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:39 +0000 UTC Normal Pod simple-prod-collector-6d88dc668c-c88zq Binding Scheduled Successfully assigned kuttl-test-faithful-stallion/simple-prod-collector-6d88dc668c-c88zq to ip-10-0-39-223.ec2.internal default-scheduler logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:39 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-6d88dc668c SuccessfulCreate Created pod: simple-prod-collector-6d88dc668c-2gkfx replicaset-controller logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:39 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-6d88dc668c SuccessfulCreate Created pod: simple-prod-collector-6d88dc668c-c88zq replicaset-controller logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:39 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-6d88dc668c to 2 deployment-controller logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:39 +0000 UTC Normal Pod simple-prod-query-76d6f9dc79-bdpcn Binding Scheduled Successfully assigned kuttl-test-faithful-stallion/simple-prod-query-76d6f9dc79-bdpcn to ip-10-0-39-223.ec2.internal default-scheduler logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:39 +0000 UTC Normal Pod simple-prod-query-76d6f9dc79-bgdrv Binding Scheduled Successfully assigned kuttl-test-faithful-stallion/simple-prod-query-76d6f9dc79-bgdrv to ip-10-0-79-227.ec2.internal default-scheduler logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:39 +0000 UTC Normal ReplicaSet.apps simple-prod-query-76d6f9dc79 SuccessfulCreate Created pod: simple-prod-query-76d6f9dc79-bgdrv replicaset-controller logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:39 +0000 UTC Normal ReplicaSet.apps simple-prod-query-76d6f9dc79 SuccessfulCreate Created pod: simple-prod-query-76d6f9dc79-bdpcn replicaset-controller logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:39 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-76d6f9dc79 to 2 deployment-controller logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:40 +0000 UTC Normal Pod simple-prod-collector-6d88dc668c-2gkfx AddedInterface Add eth0 [10.129.2.69/23] from ovn-kubernetes multus logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:40 +0000 UTC Normal Pod simple-prod-collector-6d88dc668c-2gkfx.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:44686d560aa27cae8ff8693f88c4cb6e2edf1737010ec1f80709cb42250b729d" already present on machine kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:40 +0000 UTC Normal Pod simple-prod-collector-6d88dc668c-2gkfx.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:40 +0000 UTC Normal Pod simple-prod-collector-6d88dc668c-2gkfx.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:40 +0000 UTC Normal Pod simple-prod-collector-6d88dc668c-c88zq AddedInterface Add eth0 [10.128.2.57/23] from ovn-kubernetes multus logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:40 +0000 UTC Normal Pod simple-prod-collector-6d88dc668c-c88zq.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:44686d560aa27cae8ff8693f88c4cb6e2edf1737010ec1f80709cb42250b729d" already present on machine kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:40 +0000 UTC Normal Pod simple-prod-collector-6d88dc668c-c88zq.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:40 +0000 UTC Normal Pod simple-prod-collector-6d88dc668c-c88zq.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:40 +0000 UTC Normal Pod simple-prod-query-76d6f9dc79-bdpcn AddedInterface Add eth0 [10.128.2.56/23] from ovn-kubernetes multus logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:40 +0000 UTC Normal Pod simple-prod-query-76d6f9dc79-bdpcn.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:24a83160aa930e1b72c2a2442a33b28af2b06c24b058e09afcd0d495a8066d6d" already present on machine kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:40 +0000 UTC Normal Pod simple-prod-query-76d6f9dc79-bdpcn.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:40 +0000 UTC Normal Pod simple-prod-query-76d6f9dc79-bdpcn.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:40 +0000 UTC Normal Pod simple-prod-query-76d6f9dc79-bdpcn.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:40 +0000 UTC Normal Pod simple-prod-query-76d6f9dc79-bdpcn.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:40 +0000 UTC Normal Pod simple-prod-query-76d6f9dc79-bdpcn.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:40 +0000 UTC Normal Pod simple-prod-query-76d6f9dc79-bdpcn.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" already present on machine kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:40 +0000 UTC Normal Pod simple-prod-query-76d6f9dc79-bdpcn.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:40 +0000 UTC Normal Pod simple-prod-query-76d6f9dc79-bdpcn.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:40 +0000 UTC Normal Pod simple-prod-query-76d6f9dc79-bgdrv AddedInterface Add eth0 [10.129.2.68/23] from ovn-kubernetes multus logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:40 +0000 UTC Normal Pod simple-prod-query-76d6f9dc79-bgdrv.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:24a83160aa930e1b72c2a2442a33b28af2b06c24b058e09afcd0d495a8066d6d" already present on machine kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:40 +0000 UTC Normal Pod simple-prod-query-76d6f9dc79-bgdrv.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:40 +0000 UTC Normal Pod simple-prod-query-76d6f9dc79-bgdrv.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:40 +0000 UTC Normal Pod simple-prod-query-76d6f9dc79-bgdrv.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:40 +0000 UTC Normal Pod simple-prod-query-76d6f9dc79-bgdrv.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:40 +0000 UTC Normal Pod simple-prod-query-76d6f9dc79-bgdrv.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:40 +0000 UTC Normal Pod simple-prod-query-76d6f9dc79-bgdrv.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" already present on machine kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:40 +0000 UTC Normal Pod simple-prod-query-76d6f9dc79-bgdrv.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:34:42 | es-increasing-replicas | 2024-12-09 07:34:40 +0000 UTC Normal Pod simple-prod-query-76d6f9dc79-bgdrv.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:34:42 | es-increasing-replicas | Deleting namespace: kuttl-test-faithful-stallion === CONT kuttl/harness/es-index-cleaner-autoprov logger.go:42: 07:35:17 | es-index-cleaner-autoprov | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:35:17 | es-index-cleaner-autoprov | Creating namespace: kuttl-test-master-hyena logger.go:42: 07:35:17 | es-index-cleaner-autoprov/1-install | starting test step 1-install logger.go:42: 07:35:17 | es-index-cleaner-autoprov/1-install | Jaeger:kuttl-test-master-hyena/test-es-index-cleaner-with-prefix created logger.go:42: 07:35:54 | es-index-cleaner-autoprov/1-install | test step completed 1-install logger.go:42: 07:35:54 | es-index-cleaner-autoprov/2-report-spans | starting test step 2-report-spans logger.go:42: 07:35:54 | es-index-cleaner-autoprov/2-report-spans | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE test-es-index-cleaner-with-prefix /dev/null] logger.go:42: 07:35:56 | es-index-cleaner-autoprov/2-report-spans | Warning: resource jaegers/test-es-index-cleaner-with-prefix is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:36:02 | es-index-cleaner-autoprov/2-report-spans | running command: [sh -c DAYS=5 ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JOB_NUMBER=00 JAEGER_COLLECTOR_ENDPOINT=http://test-es-index-cleaner-with-prefix-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://test-es-index-cleaner-with-prefix-query MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/report-spans.yaml.template -o report-span-00-job.yaml] logger.go:42: 07:36:02 | es-index-cleaner-autoprov/2-report-spans | running command: [sh -c kubectl apply -f report-span-00-job.yaml -n $NAMESPACE] logger.go:42: 07:36:03 | es-index-cleaner-autoprov/2-report-spans | job.batch/00-report-span created logger.go:42: 07:36:43 | es-index-cleaner-autoprov/2-report-spans | test step completed 2-report-spans logger.go:42: 07:36:43 | es-index-cleaner-autoprov/3-install | starting test step 3-install logger.go:42: 07:36:43 | es-index-cleaner-autoprov/3-install | Jaeger:kuttl-test-master-hyena/test-es-index-cleaner-with-prefix updated logger.go:42: 07:36:43 | es-index-cleaner-autoprov/3-install | test step completed 3-install logger.go:42: 07:36:43 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | starting test step 4-wait-es-index-cleaner logger.go:42: 07:36:43 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | running command: [sh -c go run ../../../../cmd-utils/wait-cronjob/main.go --cronjob test-es-index-cleaner-with-prefix-es-index-cleaner --namespace $NAMESPACE] logger.go:42: 07:36:44 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | time="2024-12-09T07:36:44Z" level=debug msg="Checking if the test-es-index-cleaner-with-prefix-es-index-cleaner CronJob exists" logger.go:42: 07:36:44 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | time="2024-12-09T07:36:44Z" level=debug msg="No BatchV1beta1/Cronjobs were found" logger.go:42: 07:36:44 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | time="2024-12-09T07:36:44Z" level=warning msg="The BatchV1/Cronjob test-es-index-cleaner-with-prefix-es-index-cleaner was not found" logger.go:42: 07:36:44 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | time="2024-12-09T07:36:44Z" level=debug msg="Found BatchV/Cronjobs:" logger.go:42: 07:36:54 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | time="2024-12-09T07:36:54Z" level=debug msg="No BatchV1beta1/Cronjobs were found" logger.go:42: 07:36:54 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | time="2024-12-09T07:36:54Z" level=info msg="Cronjob test-es-index-cleaner-with-prefix-es-index-cleaner found successfully" logger.go:42: 07:36:54 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | time="2024-12-09T07:36:54Z" level=debug msg="Waiting for the next scheduled job from test-es-index-cleaner-with-prefix-es-index-cleaner cronjob" logger.go:42: 07:36:54 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | time="2024-12-09T07:36:54Z" level=debug msg="Waiting for next job from test-es-index-cleaner-with-prefix-es-index-cleaner to succeed" logger.go:42: 07:37:04 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | time="2024-12-09T07:37:04Z" level=debug msg="Waiting for next job from test-es-index-cleaner-with-prefix-es-index-cleaner to succeed" logger.go:42: 07:37:14 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | time="2024-12-09T07:37:14Z" level=info msg="Job of owner test-es-index-cleaner-with-prefix-es-index-cleaner succeeded after test-es-index-cleaner-with-prefix-es-index-cleaner 20.028268215s" logger.go:42: 07:37:14 | es-index-cleaner-autoprov/4-wait-es-index-cleaner | test step completed 4-wait-es-index-cleaner logger.go:42: 07:37:14 | es-index-cleaner-autoprov/5-install | starting test step 5-install logger.go:42: 07:37:14 | es-index-cleaner-autoprov/5-install | Jaeger:kuttl-test-master-hyena/test-es-index-cleaner-with-prefix updated logger.go:42: 07:37:14 | es-index-cleaner-autoprov/5-install | test step completed 5-install logger.go:42: 07:37:14 | es-index-cleaner-autoprov/6-check-indices | starting test step 6-check-indices logger.go:42: 07:37:15 | es-index-cleaner-autoprov/6-check-indices | Job:kuttl-test-master-hyena/00-check-indices created logger.go:42: 07:37:20 | es-index-cleaner-autoprov/6-check-indices | test step completed 6-check-indices logger.go:42: 07:37:20 | es-index-cleaner-autoprov | es-index-cleaner-autoprov events from ns kuttl-test-master-hyena: logger.go:42: 07:37:20 | es-index-cleaner-autoprov | 2024-12-09 07:35:24 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestmasterhyenatestesindexclean-1-799c65679 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestmasterhyenatestesindexclean-1-7n2n95 replicaset-controller logger.go:42: 07:37:20 | es-index-cleaner-autoprov | 2024-12-09 07:35:24 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmasterhyenatestesindexclean-1-7n2n95 Binding Scheduled Successfully assigned kuttl-test-master-hyena/elasticsearch-cdm-kuttltestmasterhyenatestesindexclean-1-7n2n95 to ip-10-0-95-250.ec2.internal default-scheduler logger.go:42: 07:37:20 | es-index-cleaner-autoprov | 2024-12-09 07:35:24 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmasterhyenatestesindexclean-1-7n2n95 AddedInterface Add eth0 [10.131.0.36/23] from ovn-kubernetes multus logger.go:42: 07:37:20 | es-index-cleaner-autoprov | 2024-12-09 07:35:24 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmasterhyenatestesindexclean-1-7n2n95.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:8e4fbea4983cd58352349ca291383169b286bc166fad95a87807552ca43335e6" already present on machine kubelet logger.go:42: 07:37:20 | es-index-cleaner-autoprov | 2024-12-09 07:35:24 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmasterhyenatestesindexclean-1-7n2n95.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:37:20 | es-index-cleaner-autoprov | 2024-12-09 07:35:24 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmasterhyenatestesindexclean-1-7n2n95.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:37:20 | es-index-cleaner-autoprov | 2024-12-09 07:35:24 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmasterhyenatestesindexclean-1-7n2n95.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:d68824b0b2c84db8e33edf9ab344eb684c4a7ebd7ef162bbc309043adcb28e6b" already present on machine kubelet logger.go:42: 07:37:20 | es-index-cleaner-autoprov | 2024-12-09 07:35:24 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmasterhyenatestesindexclean-1-7n2n95.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:37:20 | es-index-cleaner-autoprov | 2024-12-09 07:35:24 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestmasterhyenatestesindexclean-1-7n2n95.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:37:20 | es-index-cleaner-autoprov | 2024-12-09 07:35:24 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestmasterhyenatestesindexclean-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestmasterhyenatestesindexclean-1-799c65679 to 1 deployment-controller logger.go:42: 07:37:20 | es-index-cleaner-autoprov | 2024-12-09 07:35:34 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestmasterhyenatestesindexclean-1-7n2n95.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 07:37:20 | es-index-cleaner-autoprov | 2024-12-09 07:35:39 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestmasterhyenatestesindexclean-1-7n2n95.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:37:20 | es-index-cleaner-autoprov | 2024-12-09 07:35:50 +0000 UTC Normal Deployment.apps test-es-index-cleaner-with-prefix-query ScalingReplicaSet Scaled up replica set test-es-index-cleaner-with-prefix-query-b7bbfc9cf to 1 deployment-controller logger.go:42: 07:37:20 | es-index-cleaner-autoprov | 2024-12-09 07:35:51 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-collector-6d48795dc4-v4wgk Binding Scheduled Successfully assigned kuttl-test-master-hyena/test-es-index-cleaner-with-prefix-collector-6d48795dc4-v4wgk to ip-10-0-39-223.ec2.internal default-scheduler logger.go:42: 07:37:20 | es-index-cleaner-autoprov | 2024-12-09 07:35:51 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-collector-6d48795dc4-v4wgk AddedInterface Add eth0 [10.128.2.58/23] from ovn-kubernetes multus logger.go:42: 07:37:20 | es-index-cleaner-autoprov | 2024-12-09 07:35:51 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-collector-6d48795dc4-v4wgk.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:44686d560aa27cae8ff8693f88c4cb6e2edf1737010ec1f80709cb42250b729d" already present on machine kubelet logger.go:42: 07:37:20 | es-index-cleaner-autoprov | 2024-12-09 07:35:51 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-collector-6d48795dc4-v4wgk.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:37:20 | es-index-cleaner-autoprov | 2024-12-09 07:35:51 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-collector-6d48795dc4-v4wgk.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:37:20 | es-index-cleaner-autoprov | 2024-12-09 07:35:51 +0000 UTC Normal ReplicaSet.apps test-es-index-cleaner-with-prefix-collector-6d48795dc4 SuccessfulCreate Created pod: test-es-index-cleaner-with-prefix-collector-6d48795dc4-v4wgk replicaset-controller logger.go:42: 07:37:20 | es-index-cleaner-autoprov | 2024-12-09 07:35:51 +0000 UTC Normal Deployment.apps test-es-index-cleaner-with-prefix-collector ScalingReplicaSet Scaled up replica set test-es-index-cleaner-with-prefix-collector-6d48795dc4 to 1 deployment-controller logger.go:42: 07:37:20 | es-index-cleaner-autoprov | 2024-12-09 07:35:51 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-b7bbfc9cf-5wmvp Binding Scheduled Successfully assigned kuttl-test-master-hyena/test-es-index-cleaner-with-prefix-query-b7bbfc9cf-5wmvp to ip-10-0-79-227.ec2.internal default-scheduler logger.go:42: 07:37:20 | es-index-cleaner-autoprov | 2024-12-09 07:35:51 +0000 UTC Warning Pod test-es-index-cleaner-with-prefix-query-b7bbfc9cf-5wmvp FailedMount MountVolume.SetUp failed for volume "test-es-index-cleaner-with-prefix-ui-oauth-proxy-tls" : secret "test-es-index-cleaner-with-prefix-ui-oauth-proxy-tls" not found kubelet logger.go:42: 07:37:20 | es-index-cleaner-autoprov | 2024-12-09 07:35:51 +0000 UTC Normal ReplicaSet.apps test-es-index-cleaner-with-prefix-query-b7bbfc9cf SuccessfulCreate Created pod: test-es-index-cleaner-with-prefix-query-b7bbfc9cf-5wmvp replicaset-controller logger.go:42: 07:37:20 | es-index-cleaner-autoprov | 2024-12-09 07:35:52 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-b7bbfc9cf-5wmvp AddedInterface Add eth0 [10.129.2.70/23] from ovn-kubernetes multus logger.go:42: 07:37:20 | es-index-cleaner-autoprov | 2024-12-09 07:35:52 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-b7bbfc9cf-5wmvp.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:24a83160aa930e1b72c2a2442a33b28af2b06c24b058e09afcd0d495a8066d6d" already present on machine kubelet logger.go:42: 07:37:20 | es-index-cleaner-autoprov | 2024-12-09 07:35:52 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-b7bbfc9cf-5wmvp.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:37:20 | es-index-cleaner-autoprov | 2024-12-09 07:35:52 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-b7bbfc9cf-5wmvp.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:37:20 | es-index-cleaner-autoprov | 2024-12-09 07:35:52 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-b7bbfc9cf-5wmvp.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:37:20 | es-index-cleaner-autoprov | 2024-12-09 07:35:52 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-b7bbfc9cf-5wmvp.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:37:20 | es-index-cleaner-autoprov | 2024-12-09 07:35:52 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-b7bbfc9cf-5wmvp.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:37:20 | es-index-cleaner-autoprov | 2024-12-09 07:35:52 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-b7bbfc9cf-5wmvp.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" already present on machine kubelet logger.go:42: 07:37:20 | es-index-cleaner-autoprov | 2024-12-09 07:35:52 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-b7bbfc9cf-5wmvp.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:37:20 | es-index-cleaner-autoprov | 2024-12-09 07:35:52 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-b7bbfc9cf-5wmvp.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:37:20 | es-index-cleaner-autoprov | 2024-12-09 07:35:58 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-b7bbfc9cf-5wmvp.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:37:20 | es-index-cleaner-autoprov | 2024-12-09 07:35:58 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-b7bbfc9cf-5wmvp.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:37:20 | es-index-cleaner-autoprov | 2024-12-09 07:35:58 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-b7bbfc9cf-5wmvp.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:37:20 | es-index-cleaner-autoprov | 2024-12-09 07:35:58 +0000 UTC Normal ReplicaSet.apps test-es-index-cleaner-with-prefix-query-b7bbfc9cf SuccessfulDelete Deleted pod: test-es-index-cleaner-with-prefix-query-b7bbfc9cf-5wmvp replicaset-controller logger.go:42: 07:37:20 | es-index-cleaner-autoprov | 2024-12-09 07:35:58 +0000 UTC Normal Deployment.apps test-es-index-cleaner-with-prefix-query ScalingReplicaSet Scaled down replica set test-es-index-cleaner-with-prefix-query-b7bbfc9cf to 0 from 1 deployment-controller logger.go:42: 07:37:20 | es-index-cleaner-autoprov | 2024-12-09 07:36:00 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-79c7944477-h49rt Binding Scheduled Successfully assigned kuttl-test-master-hyena/test-es-index-cleaner-with-prefix-query-79c7944477-h49rt to ip-10-0-79-227.ec2.internal default-scheduler logger.go:42: 07:37:20 | es-index-cleaner-autoprov | 2024-12-09 07:36:00 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-79c7944477-h49rt AddedInterface Add eth0 [10.129.2.71/23] from ovn-kubernetes multus logger.go:42: 07:37:20 | es-index-cleaner-autoprov | 2024-12-09 07:36:00 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-79c7944477-h49rt.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:24a83160aa930e1b72c2a2442a33b28af2b06c24b058e09afcd0d495a8066d6d" already present on machine kubelet logger.go:42: 07:37:20 | es-index-cleaner-autoprov | 2024-12-09 07:36:00 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-79c7944477-h49rt.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:37:20 | es-index-cleaner-autoprov | 2024-12-09 07:36:00 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-79c7944477-h49rt.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:37:20 | es-index-cleaner-autoprov | 2024-12-09 07:36:00 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-79c7944477-h49rt.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:37:20 | es-index-cleaner-autoprov | 2024-12-09 07:36:00 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-79c7944477-h49rt.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:37:20 | es-index-cleaner-autoprov | 2024-12-09 07:36:00 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-79c7944477-h49rt.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:37:20 | es-index-cleaner-autoprov | 2024-12-09 07:36:00 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-79c7944477-h49rt.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" already present on machine kubelet logger.go:42: 07:37:20 | es-index-cleaner-autoprov | 2024-12-09 07:36:00 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-79c7944477-h49rt.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:37:20 | es-index-cleaner-autoprov | 2024-12-09 07:36:00 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-query-79c7944477-h49rt.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:37:20 | es-index-cleaner-autoprov | 2024-12-09 07:36:00 +0000 UTC Normal ReplicaSet.apps test-es-index-cleaner-with-prefix-query-79c7944477 SuccessfulCreate Created pod: test-es-index-cleaner-with-prefix-query-79c7944477-h49rt replicaset-controller logger.go:42: 07:37:20 | es-index-cleaner-autoprov | 2024-12-09 07:36:00 +0000 UTC Normal Deployment.apps test-es-index-cleaner-with-prefix-query ScalingReplicaSet Scaled up replica set test-es-index-cleaner-with-prefix-query-79c7944477 to 1 deployment-controller logger.go:42: 07:37:20 | es-index-cleaner-autoprov | 2024-12-09 07:36:03 +0000 UTC Normal Pod 00-report-span-s8vsp Binding Scheduled Successfully assigned kuttl-test-master-hyena/00-report-span-s8vsp to ip-10-0-39-223.ec2.internal default-scheduler logger.go:42: 07:37:20 | es-index-cleaner-autoprov | 2024-12-09 07:36:03 +0000 UTC Normal Pod 00-report-span-s8vsp AddedInterface Add eth0 [10.128.2.59/23] from ovn-kubernetes multus logger.go:42: 07:37:20 | es-index-cleaner-autoprov | 2024-12-09 07:36:03 +0000 UTC Normal Pod 00-report-span-s8vsp.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:37:20 | es-index-cleaner-autoprov | 2024-12-09 07:36:03 +0000 UTC Normal Job.batch 00-report-span SuccessfulCreate Created pod: 00-report-span-s8vsp job-controller logger.go:42: 07:37:20 | es-index-cleaner-autoprov | 2024-12-09 07:36:05 +0000 UTC Normal Pod 00-report-span-s8vsp.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 1.766s (1.766s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:37:20 | es-index-cleaner-autoprov | 2024-12-09 07:36:05 +0000 UTC Normal Pod 00-report-span-s8vsp.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:37:20 | es-index-cleaner-autoprov | 2024-12-09 07:36:05 +0000 UTC Normal Pod 00-report-span-s8vsp.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:37:20 | es-index-cleaner-autoprov | 2024-12-09 07:36:06 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:37:20 | es-index-cleaner-autoprov | 2024-12-09 07:36:06 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:37:20 | es-index-cleaner-autoprov | 2024-12-09 07:36:06 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:37:20 | es-index-cleaner-autoprov | 2024-12-09 07:36:36 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedGetResourceMetric failed to get cpu utilization: missing request for cpu in container jaeger-collector of Pod test-es-index-cleaner-with-prefix-collector-6d48795dc4-v4wgk horizontal-pod-autoscaler logger.go:42: 07:37:20 | es-index-cleaner-autoprov | 2024-12-09 07:36:36 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedGetResourceMetric failed to get memory utilization: missing request for memory in container jaeger-collector of Pod test-es-index-cleaner-with-prefix-collector-6d48795dc4-v4wgk horizontal-pod-autoscaler logger.go:42: 07:37:20 | es-index-cleaner-autoprov | 2024-12-09 07:36:36 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling test-es-index-cleaner-with-prefix-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: missing request for cpu in container jaeger-collector of Pod test-es-index-cleaner-with-prefix-collector-6d48795dc4-v4wgk horizontal-pod-autoscaler logger.go:42: 07:37:20 | es-index-cleaner-autoprov | 2024-12-09 07:36:43 +0000 UTC Normal Job.batch 00-report-span Completed Job completed job-controller logger.go:42: 07:37:20 | es-index-cleaner-autoprov | 2024-12-09 07:37:00 +0000 UTC Normal Job.batch test-es-index-cleaner-with-prefix-es-index-cleaner-28895497 SuccessfulCreate Created pod: test-es-index-cleaner-with-prefix-es-index-cleaner-2889549v2xnt job-controller logger.go:42: 07:37:20 | es-index-cleaner-autoprov | 2024-12-09 07:37:00 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-es-index-cleaner-2889549v2xnt Binding Scheduled Successfully assigned kuttl-test-master-hyena/test-es-index-cleaner-with-prefix-es-index-cleaner-2889549v2xnt to ip-10-0-39-223.ec2.internal default-scheduler logger.go:42: 07:37:20 | es-index-cleaner-autoprov | 2024-12-09 07:37:00 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-es-index-cleaner-2889549v2xnt AddedInterface Add eth0 [10.128.2.60/23] from ovn-kubernetes multus logger.go:42: 07:37:20 | es-index-cleaner-autoprov | 2024-12-09 07:37:00 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-es-index-cleaner-2889549v2xnt.spec.containers{test-es-index-cleaner-with-prefix-es-index-cleaner} Pulling Pulling image "registry.redhat.io/rhosdt/jaeger-es-index-cleaner-rhel8@sha256:99a73c5dcfefd442e1939772e00b11d9c1049ce3e02db23327a8c3a07ee63b89" kubelet logger.go:42: 07:37:20 | es-index-cleaner-autoprov | 2024-12-09 07:37:00 +0000 UTC Normal CronJob.batch test-es-index-cleaner-with-prefix-es-index-cleaner SuccessfulCreate Created job test-es-index-cleaner-with-prefix-es-index-cleaner-28895497 cronjob-controller logger.go:42: 07:37:20 | es-index-cleaner-autoprov | 2024-12-09 07:37:05 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-es-index-cleaner-2889549v2xnt.spec.containers{test-es-index-cleaner-with-prefix-es-index-cleaner} Pulled Successfully pulled image "registry.redhat.io/rhosdt/jaeger-es-index-cleaner-rhel8@sha256:99a73c5dcfefd442e1939772e00b11d9c1049ce3e02db23327a8c3a07ee63b89" in 4.878s (4.878s including waiting). Image size: 111239293 bytes. kubelet logger.go:42: 07:37:20 | es-index-cleaner-autoprov | 2024-12-09 07:37:05 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-es-index-cleaner-2889549v2xnt.spec.containers{test-es-index-cleaner-with-prefix-es-index-cleaner} Created Created container test-es-index-cleaner-with-prefix-es-index-cleaner kubelet logger.go:42: 07:37:20 | es-index-cleaner-autoprov | 2024-12-09 07:37:05 +0000 UTC Normal Pod test-es-index-cleaner-with-prefix-es-index-cleaner-2889549v2xnt.spec.containers{test-es-index-cleaner-with-prefix-es-index-cleaner} Started Started container test-es-index-cleaner-with-prefix-es-index-cleaner kubelet logger.go:42: 07:37:20 | es-index-cleaner-autoprov | 2024-12-09 07:37:08 +0000 UTC Normal Job.batch test-es-index-cleaner-with-prefix-es-index-cleaner-28895497 Completed Job completed job-controller logger.go:42: 07:37:20 | es-index-cleaner-autoprov | 2024-12-09 07:37:08 +0000 UTC Normal CronJob.batch test-es-index-cleaner-with-prefix-es-index-cleaner SawCompletedJob Saw completed job: test-es-index-cleaner-with-prefix-es-index-cleaner-28895497, condition: Complete cronjob-controller logger.go:42: 07:37:20 | es-index-cleaner-autoprov | 2024-12-09 07:37:15 +0000 UTC Normal Pod 00-check-indices-bncpd Binding Scheduled Successfully assigned kuttl-test-master-hyena/00-check-indices-bncpd to ip-10-0-39-223.ec2.internal default-scheduler logger.go:42: 07:37:20 | es-index-cleaner-autoprov | 2024-12-09 07:37:15 +0000 UTC Normal Pod 00-check-indices-bncpd AddedInterface Add eth0 [10.128.2.61/23] from ovn-kubernetes multus logger.go:42: 07:37:20 | es-index-cleaner-autoprov | 2024-12-09 07:37:15 +0000 UTC Normal Pod 00-check-indices-bncpd.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:37:20 | es-index-cleaner-autoprov | 2024-12-09 07:37:15 +0000 UTC Normal Job.batch 00-check-indices SuccessfulCreate Created pod: 00-check-indices-bncpd job-controller logger.go:42: 07:37:20 | es-index-cleaner-autoprov | 2024-12-09 07:37:16 +0000 UTC Normal Pod 00-check-indices-bncpd.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 1.138s (1.138s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:37:20 | es-index-cleaner-autoprov | 2024-12-09 07:37:16 +0000 UTC Normal Pod 00-check-indices-bncpd.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:37:20 | es-index-cleaner-autoprov | 2024-12-09 07:37:16 +0000 UTC Normal Pod 00-check-indices-bncpd.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:37:20 | es-index-cleaner-autoprov | 2024-12-09 07:37:20 +0000 UTC Normal Job.batch 00-check-indices Completed Job completed job-controller logger.go:42: 07:37:20 | es-index-cleaner-autoprov | Deleting namespace: kuttl-test-master-hyena === CONT kuttl/harness/es-from-aio-to-production logger.go:42: 07:37:26 | es-from-aio-to-production | Ignoring README.md as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:37:26 | es-from-aio-to-production | Creating namespace: kuttl-test-worthy-anchovy logger.go:42: 07:37:26 | es-from-aio-to-production/0-install | starting test step 0-install logger.go:42: 07:37:26 | es-from-aio-to-production/0-install | Jaeger:kuttl-test-worthy-anchovy/my-jaeger created logger.go:42: 07:37:33 | es-from-aio-to-production/0-install | test step completed 0-install logger.go:42: 07:37:33 | es-from-aio-to-production/1-smoke-test | starting test step 1-smoke-test logger.go:42: 07:37:33 | es-from-aio-to-production/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:37:34 | es-from-aio-to-production/1-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:37:40 | es-from-aio-to-production/1-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:37:41 | es-from-aio-to-production/1-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:37:41 | es-from-aio-to-production/1-smoke-test | job.batch/report-span created logger.go:42: 07:37:41 | es-from-aio-to-production/1-smoke-test | job.batch/check-span created logger.go:42: 07:37:53 | es-from-aio-to-production/1-smoke-test | test step completed 1-smoke-test logger.go:42: 07:37:53 | es-from-aio-to-production/3-install | starting test step 3-install logger.go:42: 07:37:53 | es-from-aio-to-production/3-install | Jaeger:kuttl-test-worthy-anchovy/my-jaeger updated logger.go:42: 07:38:27 | es-from-aio-to-production/3-install | test step completed 3-install logger.go:42: 07:38:27 | es-from-aio-to-production/4-smoke-test | starting test step 4-smoke-test logger.go:42: 07:38:27 | es-from-aio-to-production/4-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:38:35 | es-from-aio-to-production/4-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:38:35 | es-from-aio-to-production/4-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:38:36 | es-from-aio-to-production/4-smoke-test | job.batch/report-span unchanged logger.go:42: 07:38:36 | es-from-aio-to-production/4-smoke-test | job.batch/check-span unchanged logger.go:42: 07:38:36 | es-from-aio-to-production/4-smoke-test | test step completed 4-smoke-test logger.go:42: 07:38:36 | es-from-aio-to-production | es-from-aio-to-production events from ns kuttl-test-worthy-anchovy: logger.go:42: 07:38:36 | es-from-aio-to-production | 2024-12-09 07:37:30 +0000 UTC Normal Pod my-jaeger-7bcfffdbf8-sg2wm Binding Scheduled Successfully assigned kuttl-test-worthy-anchovy/my-jaeger-7bcfffdbf8-sg2wm to ip-10-0-79-227.ec2.internal default-scheduler logger.go:42: 07:38:36 | es-from-aio-to-production | 2024-12-09 07:37:30 +0000 UTC Normal ReplicaSet.apps my-jaeger-7bcfffdbf8 SuccessfulCreate Created pod: my-jaeger-7bcfffdbf8-sg2wm replicaset-controller logger.go:42: 07:38:36 | es-from-aio-to-production | 2024-12-09 07:37:30 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-7bcfffdbf8 to 1 deployment-controller logger.go:42: 07:38:36 | es-from-aio-to-production | 2024-12-09 07:37:31 +0000 UTC Normal Pod my-jaeger-7bcfffdbf8-sg2wm AddedInterface Add eth0 [10.129.2.72/23] from ovn-kubernetes multus logger.go:42: 07:38:36 | es-from-aio-to-production | 2024-12-09 07:37:31 +0000 UTC Normal Pod my-jaeger-7bcfffdbf8-sg2wm.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1fa5ed13cc8a023f1e987f6cafe86adc195c373cc2b774539df6d0fd02b780a7" already present on machine kubelet logger.go:42: 07:38:36 | es-from-aio-to-production | 2024-12-09 07:37:31 +0000 UTC Normal Pod my-jaeger-7bcfffdbf8-sg2wm.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:38:36 | es-from-aio-to-production | 2024-12-09 07:37:31 +0000 UTC Normal Pod my-jaeger-7bcfffdbf8-sg2wm.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:38:36 | es-from-aio-to-production | 2024-12-09 07:37:31 +0000 UTC Normal Pod my-jaeger-7bcfffdbf8-sg2wm.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:38:36 | es-from-aio-to-production | 2024-12-09 07:37:31 +0000 UTC Normal Pod my-jaeger-7bcfffdbf8-sg2wm.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:38:36 | es-from-aio-to-production | 2024-12-09 07:37:31 +0000 UTC Normal Pod my-jaeger-7bcfffdbf8-sg2wm.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:38:36 | es-from-aio-to-production | 2024-12-09 07:37:36 +0000 UTC Normal Pod my-jaeger-7bcfffdbf8-sg2wm.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:38:36 | es-from-aio-to-production | 2024-12-09 07:37:36 +0000 UTC Normal Pod my-jaeger-7bcfffdbf8-sg2wm.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:38:36 | es-from-aio-to-production | 2024-12-09 07:37:36 +0000 UTC Normal ReplicaSet.apps my-jaeger-7bcfffdbf8 SuccessfulDelete Deleted pod: my-jaeger-7bcfffdbf8-sg2wm replicaset-controller logger.go:42: 07:38:36 | es-from-aio-to-production | 2024-12-09 07:37:36 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled down replica set my-jaeger-7bcfffdbf8 to 0 from 1 deployment-controller logger.go:42: 07:38:36 | es-from-aio-to-production | 2024-12-09 07:37:37 +0000 UTC Normal Pod my-jaeger-69ff6d7bf7-2tck7 Binding Scheduled Successfully assigned kuttl-test-worthy-anchovy/my-jaeger-69ff6d7bf7-2tck7 to ip-10-0-79-227.ec2.internal default-scheduler logger.go:42: 07:38:36 | es-from-aio-to-production | 2024-12-09 07:37:37 +0000 UTC Normal Pod my-jaeger-69ff6d7bf7-2tck7 AddedInterface Add eth0 [10.129.2.73/23] from ovn-kubernetes multus logger.go:42: 07:38:36 | es-from-aio-to-production | 2024-12-09 07:37:37 +0000 UTC Normal ReplicaSet.apps my-jaeger-69ff6d7bf7 SuccessfulCreate Created pod: my-jaeger-69ff6d7bf7-2tck7 replicaset-controller logger.go:42: 07:38:36 | es-from-aio-to-production | 2024-12-09 07:37:37 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-69ff6d7bf7 to 1 deployment-controller logger.go:42: 07:38:36 | es-from-aio-to-production | 2024-12-09 07:37:38 +0000 UTC Normal Pod my-jaeger-69ff6d7bf7-2tck7.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1fa5ed13cc8a023f1e987f6cafe86adc195c373cc2b774539df6d0fd02b780a7" already present on machine kubelet logger.go:42: 07:38:36 | es-from-aio-to-production | 2024-12-09 07:37:38 +0000 UTC Normal Pod my-jaeger-69ff6d7bf7-2tck7.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:38:36 | es-from-aio-to-production | 2024-12-09 07:37:38 +0000 UTC Normal Pod my-jaeger-69ff6d7bf7-2tck7.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:38:36 | es-from-aio-to-production | 2024-12-09 07:37:38 +0000 UTC Normal Pod my-jaeger-69ff6d7bf7-2tck7.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:38:36 | es-from-aio-to-production | 2024-12-09 07:37:38 +0000 UTC Normal Pod my-jaeger-69ff6d7bf7-2tck7.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:38:36 | es-from-aio-to-production | 2024-12-09 07:37:38 +0000 UTC Normal Pod my-jaeger-69ff6d7bf7-2tck7.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:38:36 | es-from-aio-to-production | 2024-12-09 07:37:41 +0000 UTC Normal Pod check-span-vs7kv Binding Scheduled Successfully assigned kuttl-test-worthy-anchovy/check-span-vs7kv to ip-10-0-39-223.ec2.internal default-scheduler logger.go:42: 07:38:36 | es-from-aio-to-production | 2024-12-09 07:37:41 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-vs7kv job-controller logger.go:42: 07:38:36 | es-from-aio-to-production | 2024-12-09 07:37:41 +0000 UTC Normal Pod report-span-sl95v Binding Scheduled Successfully assigned kuttl-test-worthy-anchovy/report-span-sl95v to ip-10-0-95-250.ec2.internal default-scheduler logger.go:42: 07:38:36 | es-from-aio-to-production | 2024-12-09 07:37:41 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-sl95v job-controller logger.go:42: 07:38:36 | es-from-aio-to-production | 2024-12-09 07:37:42 +0000 UTC Normal Pod check-span-vs7kv AddedInterface Add eth0 [10.128.2.62/23] from ovn-kubernetes multus logger.go:42: 07:38:36 | es-from-aio-to-production | 2024-12-09 07:37:42 +0000 UTC Normal Pod check-span-vs7kv.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:38:36 | es-from-aio-to-production | 2024-12-09 07:37:42 +0000 UTC Normal Pod check-span-vs7kv.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 699ms (699ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:38:36 | es-from-aio-to-production | 2024-12-09 07:37:42 +0000 UTC Normal Pod check-span-vs7kv.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:38:36 | es-from-aio-to-production | 2024-12-09 07:37:42 +0000 UTC Normal Pod check-span-vs7kv.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:38:36 | es-from-aio-to-production | 2024-12-09 07:37:42 +0000 UTC Normal Pod report-span-sl95v AddedInterface Add eth0 [10.131.0.37/23] from ovn-kubernetes multus logger.go:42: 07:38:36 | es-from-aio-to-production | 2024-12-09 07:37:42 +0000 UTC Normal Pod report-span-sl95v.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:38:36 | es-from-aio-to-production | 2024-12-09 07:37:42 +0000 UTC Normal Pod report-span-sl95v.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 687ms (687ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:38:36 | es-from-aio-to-production | 2024-12-09 07:37:42 +0000 UTC Normal Pod report-span-sl95v.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:38:36 | es-from-aio-to-production | 2024-12-09 07:37:42 +0000 UTC Normal Pod report-span-sl95v.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:38:36 | es-from-aio-to-production | 2024-12-09 07:37:52 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:38:36 | es-from-aio-to-production | 2024-12-09 07:37:57 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestworthyanchovymyjaeger-1-679476c959 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestworthyanchovymyjaeger-1-679476cp88js replicaset-controller logger.go:42: 07:38:36 | es-from-aio-to-production | 2024-12-09 07:37:57 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestworthyanchovymyjaeger-1-679476cp88js Binding Scheduled Successfully assigned kuttl-test-worthy-anchovy/elasticsearch-cdm-kuttltestworthyanchovymyjaeger-1-679476cp88js to ip-10-0-95-250.ec2.internal default-scheduler logger.go:42: 07:38:36 | es-from-aio-to-production | 2024-12-09 07:37:57 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestworthyanchovymyjaeger-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestworthyanchovymyjaeger-1-679476c959 to 1 deployment-controller logger.go:42: 07:38:36 | es-from-aio-to-production | 2024-12-09 07:37:58 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestworthyanchovymyjaeger-1-679476cp88js AddedInterface Add eth0 [10.131.0.38/23] from ovn-kubernetes multus logger.go:42: 07:38:36 | es-from-aio-to-production | 2024-12-09 07:37:58 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestworthyanchovymyjaeger-1-679476cp88js.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:8e4fbea4983cd58352349ca291383169b286bc166fad95a87807552ca43335e6" already present on machine kubelet logger.go:42: 07:38:36 | es-from-aio-to-production | 2024-12-09 07:37:58 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestworthyanchovymyjaeger-1-679476cp88js.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:38:36 | es-from-aio-to-production | 2024-12-09 07:37:58 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestworthyanchovymyjaeger-1-679476cp88js.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:38:36 | es-from-aio-to-production | 2024-12-09 07:37:58 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestworthyanchovymyjaeger-1-679476cp88js.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:d68824b0b2c84db8e33edf9ab344eb684c4a7ebd7ef162bbc309043adcb28e6b" already present on machine kubelet logger.go:42: 07:38:36 | es-from-aio-to-production | 2024-12-09 07:37:58 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestworthyanchovymyjaeger-1-679476cp88js.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:38:36 | es-from-aio-to-production | 2024-12-09 07:37:58 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestworthyanchovymyjaeger-1-679476cp88js.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:38:36 | es-from-aio-to-production | 2024-12-09 07:38:13 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestworthyanchovymyjaeger-1-679476cp88js.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:38:36 | es-from-aio-to-production | 2024-12-09 07:38:15 +0000 UTC Normal Job.batch report-span Completed Job completed job-controller logger.go:42: 07:38:36 | es-from-aio-to-production | 2024-12-09 07:38:24 +0000 UTC Normal Pod my-jaeger-69ff6d7bf7-2tck7.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:38:36 | es-from-aio-to-production | 2024-12-09 07:38:24 +0000 UTC Normal Pod my-jaeger-69ff6d7bf7-2tck7.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:38:36 | es-from-aio-to-production | 2024-12-09 07:38:24 +0000 UTC Normal Pod my-jaeger-collector-56b44d4d86-97fcz Binding Scheduled Successfully assigned kuttl-test-worthy-anchovy/my-jaeger-collector-56b44d4d86-97fcz to ip-10-0-79-227.ec2.internal default-scheduler logger.go:42: 07:38:36 | es-from-aio-to-production | 2024-12-09 07:38:24 +0000 UTC Normal ReplicaSet.apps my-jaeger-collector-56b44d4d86 SuccessfulCreate Created pod: my-jaeger-collector-56b44d4d86-97fcz replicaset-controller logger.go:42: 07:38:36 | es-from-aio-to-production | 2024-12-09 07:38:24 +0000 UTC Normal Deployment.apps my-jaeger-collector ScalingReplicaSet Scaled up replica set my-jaeger-collector-56b44d4d86 to 1 deployment-controller logger.go:42: 07:38:36 | es-from-aio-to-production | 2024-12-09 07:38:24 +0000 UTC Normal Pod my-jaeger-query-7974c95878-lw2nz Binding Scheduled Successfully assigned kuttl-test-worthy-anchovy/my-jaeger-query-7974c95878-lw2nz to ip-10-0-39-223.ec2.internal default-scheduler logger.go:42: 07:38:36 | es-from-aio-to-production | 2024-12-09 07:38:24 +0000 UTC Normal ReplicaSet.apps my-jaeger-query-7974c95878 SuccessfulCreate Created pod: my-jaeger-query-7974c95878-lw2nz replicaset-controller logger.go:42: 07:38:36 | es-from-aio-to-production | 2024-12-09 07:38:24 +0000 UTC Normal Deployment.apps my-jaeger-query ScalingReplicaSet Scaled up replica set my-jaeger-query-7974c95878 to 1 deployment-controller logger.go:42: 07:38:36 | es-from-aio-to-production | 2024-12-09 07:38:25 +0000 UTC Normal Pod my-jaeger-collector-56b44d4d86-97fcz AddedInterface Add eth0 [10.129.2.74/23] from ovn-kubernetes multus logger.go:42: 07:38:36 | es-from-aio-to-production | 2024-12-09 07:38:25 +0000 UTC Normal Pod my-jaeger-collector-56b44d4d86-97fcz.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:44686d560aa27cae8ff8693f88c4cb6e2edf1737010ec1f80709cb42250b729d" already present on machine kubelet logger.go:42: 07:38:36 | es-from-aio-to-production | 2024-12-09 07:38:25 +0000 UTC Normal Pod my-jaeger-collector-56b44d4d86-97fcz.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:38:36 | es-from-aio-to-production | 2024-12-09 07:38:25 +0000 UTC Normal Pod my-jaeger-collector-56b44d4d86-97fcz.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:38:36 | es-from-aio-to-production | 2024-12-09 07:38:25 +0000 UTC Normal Pod my-jaeger-query-7974c95878-lw2nz AddedInterface Add eth0 [10.128.2.63/23] from ovn-kubernetes multus logger.go:42: 07:38:36 | es-from-aio-to-production | 2024-12-09 07:38:25 +0000 UTC Normal Pod my-jaeger-query-7974c95878-lw2nz.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:24a83160aa930e1b72c2a2442a33b28af2b06c24b058e09afcd0d495a8066d6d" already present on machine kubelet logger.go:42: 07:38:36 | es-from-aio-to-production | 2024-12-09 07:38:25 +0000 UTC Normal Pod my-jaeger-query-7974c95878-lw2nz.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:38:36 | es-from-aio-to-production | 2024-12-09 07:38:25 +0000 UTC Normal Pod my-jaeger-query-7974c95878-lw2nz.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:38:36 | es-from-aio-to-production | 2024-12-09 07:38:25 +0000 UTC Normal Pod my-jaeger-query-7974c95878-lw2nz.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:38:36 | es-from-aio-to-production | 2024-12-09 07:38:25 +0000 UTC Normal Pod my-jaeger-query-7974c95878-lw2nz.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:38:36 | es-from-aio-to-production | 2024-12-09 07:38:25 +0000 UTC Normal Pod my-jaeger-query-7974c95878-lw2nz.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:38:36 | es-from-aio-to-production | 2024-12-09 07:38:25 +0000 UTC Normal Pod my-jaeger-query-7974c95878-lw2nz.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" already present on machine kubelet logger.go:42: 07:38:36 | es-from-aio-to-production | 2024-12-09 07:38:25 +0000 UTC Normal Pod my-jaeger-query-7974c95878-lw2nz.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:38:36 | es-from-aio-to-production | 2024-12-09 07:38:25 +0000 UTC Normal Pod my-jaeger-query-7974c95878-lw2nz.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:38:36 | es-from-aio-to-production | Deleting namespace: kuttl-test-worthy-anchovy === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (654.64s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (6.02s) --- PASS: kuttl/harness/es-multiinstance (99.18s) --- PASS: kuttl/harness/es-simple-prod (6.23s) --- PASS: kuttl/harness/es-rollover-autoprov (231.24s) --- PASS: kuttl/harness/es-increasing-replicas (106.17s) --- PASS: kuttl/harness/es-index-cleaner-autoprov (129.32s) --- PASS: kuttl/harness/es-from-aio-to-production (76.29s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name elasticsearch --report --output /logs/artifacts/elasticsearch.xml ./artifacts/kuttl-report.xml time="2024-12-09T07:38:43Z" level=debug msg="Setting a new name for the test suites" time="2024-12-09T07:38:43Z" level=debug msg="Removing 'artifacts' TestCase" time="2024-12-09T07:38:43Z" level=debug msg="normalizing test case names" time="2024-12-09T07:38:43Z" level=debug msg="elasticsearch/artifacts -> elasticsearch_artifacts" time="2024-12-09T07:38:43Z" level=debug msg="elasticsearch/es-multiinstance -> elasticsearch_es_multiinstance" time="2024-12-09T07:38:43Z" level=debug msg="elasticsearch/es-simple-prod -> elasticsearch_es_simple_prod" time="2024-12-09T07:38:43Z" level=debug msg="elasticsearch/es-rollover-autoprov -> elasticsearch_es_rollover_autoprov" time="2024-12-09T07:38:43Z" level=debug msg="elasticsearch/es-increasing-replicas -> elasticsearch_es_increasing_replicas" time="2024-12-09T07:38:43Z" level=debug msg="elasticsearch/es-index-cleaner-autoprov -> elasticsearch_es_index_cleaner_autoprov" time="2024-12-09T07:38:43Z" level=debug msg="elasticsearch/es-from-aio-to-production -> elasticsearch_es_from_aio_to_production" +-----------------------------------------+--------+ | NAME | RESULT | +-----------------------------------------+--------+ | elasticsearch_artifacts | passed | | elasticsearch_es_multiinstance | passed | | elasticsearch_es_simple_prod | passed | | elasticsearch_es_rollover_autoprov | passed | | elasticsearch_es_increasing_replicas | passed | | elasticsearch_es_index_cleaner_autoprov | passed | | elasticsearch_es_from_aio_to_production | passed | +-----------------------------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh examples false true + '[' 3 -ne 3 ']' + test_suite_name=examples + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/examples.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-examples make[2]: Entering directory '/tmp/jaeger-tests' >>>> Elasticsearch image not loaded because SKIP_ES_EXTERNAL is true KAFKA_VERSION=3.6.0 \ SKIP_KAFKA=false \ VERTX_IMG=jaegertracing/vertx-create-span:operator-e2e-tests \ ./tests/e2e/examples/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.18.0-0.nightly-2024-12-05-191355 True False 33m Cluster version is 4.18.0-0.nightly-2024-12-05-191355' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.18.0-0.nightly-2024-12-05-191355 True False 33m Cluster version is 4.18.0-0.nightly-2024-12-05-191355' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/examples/render.sh ++ export SUITE_DIR=./tests/e2e/examples ++ SUITE_DIR=./tests/e2e/examples ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/examples ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + start_test examples-agent-with-priority-class + '[' 1 -ne 1 ']' + test_name=examples-agent-with-priority-class + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-agent-with-priority-class' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-agent-with-priority-class\e[0m' Rendering files for test examples-agent-with-priority-class + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build + '[' _build '!=' _build ']' + mkdir -p examples-agent-with-priority-class + cd examples-agent-with-priority-class + example_name=agent-with-priority-class + prepare_daemonset 00 + '[' 1 -ne 1 ']' + test_step=00 + '[' true = true ']' + cat /tmp/jaeger-tests/examples/openshift/hostport-scc-daemonset.yaml + echo --- + cat /tmp/jaeger-tests/examples/openshift/service_account_jaeger-agent-daemonset.yaml + '[' true '!=' true ']' + render_install_example agent-with-priority-class 02 + '[' 2 -ne 2 ']' + example_name=agent-with-priority-class + test_step=02 + install_file=./02-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/agent-with-priority-class.yaml -o ./02-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./02-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./02-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./02-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./02-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./02-install.yaml ++ jaeger_name=agent-as-daemonset ++ '[' -z agent-as-daemonset ']' ++ echo agent-as-daemonset ++ return 0 + JAEGER_NAME=agent-as-daemonset + local jaeger_strategy ++ get_jaeger_strategy ./02-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./02-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./02-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./02-install.yaml ++ strategy=DaemonSet ++ '[' DaemonSet = null ']' ++ echo DaemonSet ++ return 0 + jaeger_strategy=DaemonSet + '[' DaemonSet = DaemonSet ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./02-assert.yaml + render_smoke_test_example agent-with-priority-class 02 + '[' 2 -ne 2 ']' + example_name=agent-with-priority-class + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/agent-with-priority-class.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/agent-with-priority-class.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/agent-with-priority-class.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/agent-with-priority-class.yaml ++ jaeger_name=agent-as-daemonset ++ '[' -z agent-as-daemonset ']' ++ echo agent-as-daemonset ++ return 0 + jaeger_name=agent-as-daemonset + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test agent-as-daemonset true 02 + '[' 3 -ne 3 ']' + jaeger=agent-as-daemonset + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://agent-as-daemonset-query:443 + JAEGER_QUERY_ENDPOINT=https://agent-as-daemonset-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://agent-as-daemonset-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://agent-as-daemonset-collector-headless:14268 + export JAEGER_NAME=agent-as-daemonset + JAEGER_NAME=agent-as-daemonset + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-all-in-one-with-options + '[' 1 -ne 1 ']' + test_name=examples-all-in-one-with-options + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-all-in-one-with-options' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-all-in-one-with-options\e[0m' Rendering files for test examples-all-in-one-with-options + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-agent-with-priority-class + '[' examples-agent-with-priority-class '!=' _build ']' + cd .. + mkdir -p examples-all-in-one-with-options + cd examples-all-in-one-with-options + example_name=all-in-one-with-options + render_install_example all-in-one-with-options 00 + '[' 2 -ne 2 ']' + example_name=all-in-one-with-options + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/all-in-one-with-options.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=my-jaeger ++ '[' -z my-jaeger ']' ++ echo my-jaeger ++ return 0 + JAEGER_NAME=my-jaeger + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=allInOne ++ '[' allInOne = production ']' ++ '[' allInOne = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + /tmp/jaeger-tests/bin/yq e -i '.metadata.name="my-jaeger"' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i 'del(.spec.allInOne.image)' ./00-install.yaml + render_smoke_test_example all-in-one-with-options 01 + '[' 2 -ne 2 ']' + example_name=all-in-one-with-options + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/all-in-one-with-options.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/all-in-one-with-options.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/all-in-one-with-options.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/all-in-one-with-options.yaml ++ jaeger_name=my-jaeger ++ '[' -z my-jaeger ']' ++ echo my-jaeger ++ return 0 + jaeger_name=my-jaeger + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test my-jaeger true 01 + '[' 3 -ne 3 ']' + jaeger=my-jaeger + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 + export JAEGER_NAME=my-jaeger + JAEGER_NAME=my-jaeger + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + '[' true = true ']' + sed -i s~my-jaeger-query:443~my-jaeger-query:443/jaeger~gi ./01-smoke-test.yaml + start_test examples-business-application-injected-sidecar + '[' 1 -ne 1 ']' + test_name=examples-business-application-injected-sidecar + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-business-application-injected-sidecar' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-business-application-injected-sidecar\e[0m' Rendering files for test examples-business-application-injected-sidecar + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-all-in-one-with-options + '[' examples-all-in-one-with-options '!=' _build ']' + cd .. + mkdir -p examples-business-application-injected-sidecar + cd examples-business-application-injected-sidecar + example_name=simplest + cp /tmp/jaeger-tests/examples/business-application-injected-sidecar.yaml ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].image=strenv(VERTX_IMG)' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.httpGet.path="/"' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.httpGet.port=8080' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.initialDelaySeconds=1' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.failureThreshold=3' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.periodSeconds=10' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.successThreshold=1' ./00-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.template.spec.containers[0].livenessProbe.timeoutSeconds=1' ./00-install.yaml + render_install_example simplest 01 + '[' 2 -ne 2 ']' + example_name=simplest + test_step=01 + install_file=./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/simplest.yaml -o ./01-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./01-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./01-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./01-install.yaml ++ jaeger_name=simplest ++ '[' -z simplest ']' ++ echo simplest ++ return 0 + JAEGER_NAME=simplest + local jaeger_strategy ++ get_jaeger_strategy ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./01-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./01-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./01-assert.yaml + render_smoke_test_example simplest 02 + '[' 2 -ne 2 ']' + example_name=simplest + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/simplest.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/simplest.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/simplest.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/simplest.yaml ++ jaeger_name=simplest ++ '[' -z simplest ']' ++ echo simplest ++ return 0 + jaeger_name=simplest + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test simplest true 02 + '[' 3 -ne 3 ']' + jaeger=simplest + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simplest-query:443 + JAEGER_QUERY_ENDPOINT=https://simplest-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 + export JAEGER_NAME=simplest + JAEGER_NAME=simplest + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-collector-with-priority-class + '[' 1 -ne 1 ']' + test_name=examples-collector-with-priority-class + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-collector-with-priority-class' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-collector-with-priority-class\e[0m' Rendering files for test examples-collector-with-priority-class + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-business-application-injected-sidecar + '[' examples-business-application-injected-sidecar '!=' _build ']' + cd .. + mkdir -p examples-collector-with-priority-class + cd examples-collector-with-priority-class + example_name=collector-with-priority-class + render_install_example collector-with-priority-class 00 + '[' 2 -ne 2 ']' + example_name=collector-with-priority-class + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/collector-with-priority-class.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=collector-with-high-priority ++ '[' -z collector-with-high-priority ']' ++ echo collector-with-high-priority ++ return 0 + JAEGER_NAME=collector-with-high-priority + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test_example collector-with-priority-class 01 + '[' 2 -ne 2 ']' + example_name=collector-with-priority-class + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/collector-with-priority-class.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/collector-with-priority-class.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/collector-with-priority-class.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/collector-with-priority-class.yaml ++ jaeger_name=collector-with-high-priority ++ '[' -z collector-with-high-priority ']' ++ echo collector-with-high-priority ++ return 0 + jaeger_name=collector-with-high-priority + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test collector-with-high-priority true 01 + '[' 3 -ne 3 ']' + jaeger=collector-with-high-priority + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://collector-with-high-priority-query:443 + JAEGER_QUERY_ENDPOINT=https://collector-with-high-priority-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://collector-with-high-priority-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://collector-with-high-priority-collector-headless:14268 + export JAEGER_NAME=collector-with-high-priority + JAEGER_NAME=collector-with-high-priority + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-service-types + '[' 1 -ne 1 ']' + test_name=examples-service-types + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-service-types' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-service-types\e[0m' Rendering files for test examples-service-types + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-collector-with-priority-class + '[' examples-collector-with-priority-class '!=' _build ']' + cd .. + mkdir -p examples-service-types + cd examples-service-types + example_name=service-types + render_install_example service-types 00 + '[' 2 -ne 2 ']' + example_name=service-types + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/service-types.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=service-types ++ '[' -z service-types ']' ++ echo service-types ++ return 0 + JAEGER_NAME=service-types + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test_example service-types 01 + '[' 2 -ne 2 ']' + example_name=service-types + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/service-types.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/service-types.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/service-types.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/service-types.yaml ++ jaeger_name=service-types ++ '[' -z service-types ']' ++ echo service-types ++ return 0 + jaeger_name=service-types + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test service-types true 01 + '[' 3 -ne 3 ']' + jaeger=service-types + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://service-types-query:443 + JAEGER_QUERY_ENDPOINT=https://service-types-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://service-types-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://service-types-collector-headless:14268 + export JAEGER_NAME=service-types + JAEGER_NAME=service-types + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-simple-prod + '[' 1 -ne 1 ']' + test_name=examples-simple-prod + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-simple-prod' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-simple-prod\e[0m' Rendering files for test examples-simple-prod + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-service-types + '[' examples-service-types '!=' _build ']' + cd .. + mkdir -p examples-simple-prod + cd examples-simple-prod + example_name=simple-prod + render_install_example simple-prod 01 + '[' 2 -ne 2 ']' + example_name=simple-prod + test_step=01 + install_file=./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/simple-prod.yaml -o ./01-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./01-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./01-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./01-install.yaml ++ jaeger_name=simple-prod ++ '[' -z simple-prod ']' ++ echo simple-prod ++ return 0 + JAEGER_NAME=simple-prod + local jaeger_strategy ++ get_jaeger_strategy ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./01-install.yaml ++ strategy=production ++ '[' production = production ']' ++ echo production ++ return 0 + jaeger_strategy=production + '[' production = DaemonSet ']' + '[' production = allInOne ']' + '[' production = production ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + [[ true = true ]] + [[ true = true ]] + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.options={}' ./01-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.elasticsearch={"nodeCount":1,"resources":{"limits":{"memory":"2Gi"}}}' ./01-install.yaml + render_smoke_test_example simple-prod 02 + '[' 2 -ne 2 ']' + example_name=simple-prod + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/simple-prod.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/simple-prod.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/simple-prod.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/simple-prod.yaml ++ jaeger_name=simple-prod ++ '[' -z simple-prod ']' ++ echo simple-prod ++ return 0 + jaeger_name=simple-prod + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test simple-prod true 02 + '[' 3 -ne 3 ']' + jaeger=simple-prod + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + export JAEGER_NAME=simple-prod + JAEGER_NAME=simple-prod + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-simple-prod-with-volumes + '[' 1 -ne 1 ']' + test_name=examples-simple-prod-with-volumes + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-simple-prod-with-volumes' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-simple-prod-with-volumes\e[0m' Rendering files for test examples-simple-prod-with-volumes + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-simple-prod + '[' examples-simple-prod '!=' _build ']' + cd .. + mkdir -p examples-simple-prod-with-volumes + cd examples-simple-prod-with-volumes + example_name=simple-prod-with-volumes + render_install_example simple-prod-with-volumes 01 + '[' 2 -ne 2 ']' + example_name=simple-prod-with-volumes + test_step=01 + install_file=./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/simple-prod-with-volumes.yaml -o ./01-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./01-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./01-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./01-install.yaml ++ jaeger_name=simple-prod ++ '[' -z simple-prod ']' ++ echo simple-prod ++ return 0 + JAEGER_NAME=simple-prod + local jaeger_strategy ++ get_jaeger_strategy ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./01-install.yaml ++ strategy=production ++ '[' production = production ']' ++ echo production ++ return 0 + jaeger_strategy=production + '[' production = DaemonSet ']' + '[' production = allInOne ']' + '[' production = production ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/production-jaeger-assert.yaml.template -o ./01-assert.yaml + [[ true = true ]] + [[ true = true ]] + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.options={}' ./01-install.yaml + /tmp/jaeger-tests/bin/yq e -i '.spec.storage.elasticsearch={"nodeCount":1,"resources":{"limits":{"memory":"2Gi"}}}' ./01-install.yaml + render_smoke_test_example simple-prod-with-volumes 02 + '[' 2 -ne 2 ']' + example_name=simple-prod-with-volumes + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/simple-prod-with-volumes.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/simple-prod-with-volumes.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/simple-prod-with-volumes.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/simple-prod-with-volumes.yaml ++ jaeger_name=simple-prod ++ '[' -z simple-prod ']' ++ echo simple-prod ++ return 0 + jaeger_name=simple-prod + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test simple-prod true 02 + '[' 3 -ne 3 ']' + jaeger=simple-prod + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 + export JAEGER_NAME=simple-prod + JAEGER_NAME=simple-prod + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + /tmp/jaeger-tests/bin/gomplate -f ./03-check-volume.yaml.template -o 03-check-volume.yaml + start_test examples-simplest + '[' 1 -ne 1 ']' + test_name=examples-simplest + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-simplest' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-simplest\e[0m' Rendering files for test examples-simplest + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-simple-prod-with-volumes + '[' examples-simple-prod-with-volumes '!=' _build ']' + cd .. + mkdir -p examples-simplest + cd examples-simplest + example_name=simplest + render_install_example simplest 00 + '[' 2 -ne 2 ']' + example_name=simplest + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/simplest.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=simplest ++ '[' -z simplest ']' ++ echo simplest ++ return 0 + JAEGER_NAME=simplest + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test_example simplest 01 + '[' 2 -ne 2 ']' + example_name=simplest + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/simplest.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/simplest.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/simplest.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/simplest.yaml ++ jaeger_name=simplest ++ '[' -z simplest ']' ++ echo simplest ++ return 0 + jaeger_name=simplest + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test simplest true 01 + '[' 3 -ne 3 ']' + jaeger=simplest + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://simplest-query:443 + JAEGER_QUERY_ENDPOINT=https://simplest-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 + export JAEGER_NAME=simplest + JAEGER_NAME=simplest + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-with-badger + '[' 1 -ne 1 ']' + test_name=examples-with-badger + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-with-badger' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-with-badger\e[0m' Rendering files for test examples-with-badger + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-simplest + '[' examples-simplest '!=' _build ']' + cd .. + mkdir -p examples-with-badger + cd examples-with-badger + example_name=with-badger + render_install_example with-badger 00 + '[' 2 -ne 2 ']' + example_name=with-badger + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/with-badger.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=with-badger ++ '[' -z with-badger ']' ++ echo with-badger ++ return 0 + JAEGER_NAME=with-badger + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test_example with-badger 01 + '[' 2 -ne 2 ']' + example_name=with-badger + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/with-badger.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/with-badger.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/with-badger.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/with-badger.yaml ++ jaeger_name=with-badger ++ '[' -z with-badger ']' ++ echo with-badger ++ return 0 + jaeger_name=with-badger + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test with-badger true 01 + '[' 3 -ne 3 ']' + jaeger=with-badger + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://with-badger-query:443 + JAEGER_QUERY_ENDPOINT=https://with-badger-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://with-badger-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://with-badger-collector-headless:14268 + export JAEGER_NAME=with-badger + JAEGER_NAME=with-badger + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-with-badger-and-volume + '[' 1 -ne 1 ']' + test_name=examples-with-badger-and-volume + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-with-badger-and-volume' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-with-badger-and-volume\e[0m' Rendering files for test examples-with-badger-and-volume + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-with-badger + '[' examples-with-badger '!=' _build ']' + cd .. + mkdir -p examples-with-badger-and-volume + cd examples-with-badger-and-volume + example_name=with-badger-and-volume + render_install_example with-badger-and-volume 00 + '[' 2 -ne 2 ']' + example_name=with-badger-and-volume + test_step=00 + install_file=./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/with-badger-and-volume.yaml -o ./00-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./00-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./00-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./00-install.yaml ++ jaeger_name=with-badger-and-volume ++ '[' -z with-badger-and-volume ']' ++ echo with-badger-and-volume ++ return 0 + JAEGER_NAME=with-badger-and-volume + local jaeger_strategy ++ get_jaeger_strategy ./00-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./00-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = production ']' ++ '[' null = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./00-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./00-assert.yaml + render_smoke_test_example with-badger-and-volume 01 + '[' 2 -ne 2 ']' + example_name=with-badger-and-volume + test_step=01 + deployment_file=/tmp/jaeger-tests/examples/with-badger-and-volume.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/with-badger-and-volume.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/with-badger-and-volume.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/with-badger-and-volume.yaml ++ jaeger_name=with-badger-and-volume ++ '[' -z with-badger-and-volume ']' ++ echo with-badger-and-volume ++ return 0 + jaeger_name=with-badger-and-volume + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test with-badger-and-volume true 01 + '[' 3 -ne 3 ']' + jaeger=with-badger-and-volume + is_secured=true + test_step=01 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://with-badger-and-volume-query:443 + JAEGER_QUERY_ENDPOINT=https://with-badger-and-volume-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://with-badger-and-volume-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://with-badger-and-volume-collector-headless:14268 + export JAEGER_NAME=with-badger-and-volume + JAEGER_NAME=with-badger-and-volume + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./01-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./01-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-with-cassandra + '[' 1 -ne 1 ']' + test_name=examples-with-cassandra + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-with-cassandra' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-with-cassandra\e[0m' Rendering files for test examples-with-cassandra + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-with-badger-and-volume + '[' examples-with-badger-and-volume '!=' _build ']' + cd .. + mkdir -p examples-with-cassandra + cd examples-with-cassandra + example_name=with-cassandra + render_install_cassandra 00 + '[' 1 -ne 1 ']' + test_step=00 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/cassandra-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/cassandra-assert.yaml.template -o ./00-assert.yaml + render_install_example with-cassandra 01 + '[' 2 -ne 2 ']' + example_name=with-cassandra + test_step=01 + install_file=./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/with-cassandra.yaml -o ./01-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./01-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./01-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./01-install.yaml ++ jaeger_name=with-cassandra ++ '[' -z with-cassandra ']' ++ echo with-cassandra ++ return 0 + JAEGER_NAME=with-cassandra + local jaeger_strategy ++ get_jaeger_strategy ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./01-install.yaml ++ strategy=allInOne ++ '[' allInOne = production ']' ++ '[' allInOne = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./01-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./01-assert.yaml + render_smoke_test_example with-cassandra 02 + '[' 2 -ne 2 ']' + example_name=with-cassandra + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/with-cassandra.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/with-cassandra.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/with-cassandra.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/with-cassandra.yaml ++ jaeger_name=with-cassandra ++ '[' -z with-cassandra ']' ++ echo with-cassandra ++ return 0 + jaeger_name=with-cassandra + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test with-cassandra true 02 + '[' 3 -ne 3 ']' + jaeger=with-cassandra + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://with-cassandra-query:443 + JAEGER_QUERY_ENDPOINT=https://with-cassandra-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://with-cassandra-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://with-cassandra-collector-headless:14268 + export JAEGER_NAME=with-cassandra + JAEGER_NAME=with-cassandra + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-with-sampling + '[' 1 -ne 1 ']' + test_name=examples-with-sampling + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-with-sampling' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-with-sampling\e[0m' Rendering files for test examples-with-sampling + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-with-cassandra + '[' examples-with-cassandra '!=' _build ']' + cd .. + mkdir -p examples-with-sampling + cd examples-with-sampling + export example_name=with-sampling + example_name=with-sampling + render_install_cassandra 00 + '[' 1 -ne 1 ']' + test_step=00 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/cassandra-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/cassandra-assert.yaml.template -o ./00-assert.yaml + render_install_example with-sampling 01 + '[' 2 -ne 2 ']' + example_name=with-sampling + test_step=01 + install_file=./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/with-sampling.yaml -o ./01-install.yaml + sed -i 's~server-urls: http://elasticsearch.default.svc:9200~server-urls: http://elasticsearch:9200~gi' ./01-install.yaml + sed -i s~cassandra.default.svc~cassandra~gi ./01-install.yaml + export JAEGER_NAME ++ get_jaeger_name ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' ./01-install.yaml ++ jaeger_name=with-sampling ++ '[' -z with-sampling ']' ++ echo with-sampling ++ return 0 + JAEGER_NAME=with-sampling + local jaeger_strategy ++ get_jaeger_strategy ./01-install.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=./01-install.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.strategy' ./01-install.yaml ++ strategy=allInOne ++ '[' allInOne = production ']' ++ '[' allInOne = streaming ']' +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").spec.agent.strategy' ./01-install.yaml ++ strategy=null ++ '[' null = null ']' ++ echo allInOne ++ return 0 + jaeger_strategy=allInOne + '[' allInOne = DaemonSet ']' + '[' allInOne = allInOne ']' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./01-assert.yaml + render_smoke_test_example with-sampling 02 + '[' 2 -ne 2 ']' + example_name=with-sampling + test_step=02 + deployment_file=/tmp/jaeger-tests/examples/with-sampling.yaml ++ get_jaeger_name /tmp/jaeger-tests/examples/with-sampling.yaml ++ '[' 1 -ne 1 ']' ++ deployment_file=/tmp/jaeger-tests/examples/with-sampling.yaml +++ /tmp/jaeger-tests/bin/yq e '. | select(.kind == "Jaeger").metadata.name' /tmp/jaeger-tests/examples/with-sampling.yaml ++ jaeger_name=with-sampling ++ '[' -z with-sampling ']' ++ echo with-sampling ++ return 0 + jaeger_name=with-sampling + is_secured=false + '[' true = true ']' + is_secured=true + render_smoke_test with-sampling true 02 + '[' 3 -ne 3 ']' + jaeger=with-sampling + is_secured=true + test_step=02 + '[' true = true ']' + protocol=https:// + query_port=:443 + template=/tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template + export JAEGER_QUERY_ENDPOINT=https://with-sampling-query:443 + JAEGER_QUERY_ENDPOINT=https://with-sampling-query:443 + export JAEGER_COLLECTOR_ENDPOINT=http://with-sampling-collector-headless:14268 + JAEGER_COLLECTOR_ENDPOINT=http://with-sampling-collector-headless:14268 + export JAEGER_NAME=with-sampling + JAEGER_NAME=with-sampling + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/openshift/smoke-test.yaml.template -o ./02-smoke-test.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test-assert.yaml.template -o ./02-assert.yaml + unset JAEGER_NAME + unset JAEGER_QUERY_ENDPOINT + unset JAEGER_COLLECTOR_ENDPOINT + start_test examples-agent-as-daemonset + '[' 1 -ne 1 ']' + test_name=examples-agent-as-daemonset + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-agent-as-daemonset' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-agent-as-daemonset\e[0m' Rendering files for test examples-agent-as-daemonset + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-with-sampling + '[' examples-with-sampling '!=' _build ']' + cd .. + mkdir -p examples-agent-as-daemonset + cd examples-agent-as-daemonset + '[' true = true ']' + prepare_daemonset 00 + '[' 1 -ne 1 ']' + test_step=00 + '[' true = true ']' + cat /tmp/jaeger-tests/examples/openshift/hostport-scc-daemonset.yaml + echo --- + cat /tmp/jaeger-tests/examples/openshift/service_account_jaeger-agent-daemonset.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/openshift/agent-as-daemonset.yaml -o 02-install.yaml + '[' true = true ']' + start_test examples-openshift-with-htpasswd + '[' 1 -ne 1 ']' + test_name=examples-openshift-with-htpasswd + echo =========================================================================== =========================================================================== + info 'Rendering files for test examples-openshift-with-htpasswd' + '[' 1 -ne 1 ']' + echo -e '\e[1;34mRendering files for test examples-openshift-with-htpasswd\e[0m' Rendering files for test examples-openshift-with-htpasswd + echo =========================================================================== =========================================================================== +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-agent-as-daemonset + '[' examples-agent-as-daemonset '!=' _build ']' + cd .. + mkdir -p examples-openshift-with-htpasswd + cd examples-openshift-with-htpasswd + export JAEGER_NAME=with-htpasswd + JAEGER_NAME=with-htpasswd + export JAEGER_USERNAME=awesomeuser + JAEGER_USERNAME=awesomeuser + export JAEGER_PASSWORD=awesomepassword + JAEGER_PASSWORD=awesomepassword + export 'JAEGER_USER_PASSWORD_HASH=awesomeuser:{SHA}uUdqPVUyqNBmERU0Qxj3KFaZnjw=' + JAEGER_USER_PASSWORD_HASH='awesomeuser:{SHA}uUdqPVUyqNBmERU0Qxj3KFaZnjw=' ++ echo 'awesomeuser:{SHA}uUdqPVUyqNBmERU0Qxj3KFaZnjw=' ++ base64 + SECRET=YXdlc29tZXVzZXI6e1NIQX11VWRxUFZVeXFOQm1FUlUwUXhqM0tGYVpuanc9Cg== + /tmp/jaeger-tests/bin/gomplate -f ./00-install.yaml.template -o ./00-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/examples/openshift/with-htpasswd.yaml -o ./01-install.yaml + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/allinone-jaeger-assert.yaml.template -o ./01-assert.yaml + export 'GET_URL_COMMAND=kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE' + GET_URL_COMMAND='kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE' + export 'URL=https://$(kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE)/search' + URL='https://$(kubectl get routes -o=jsonpath='\''{.items[0].status.ingress[0].host}'\'' -n $NAMESPACE)/search' + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/ensure-ingress-host.sh.template -o ./ensure-ingress-host.sh + chmod +x ./ensure-ingress-host.sh + INSECURE=true + JAEGER_USERNAME= + JAEGER_PASSWORD= + EXPECTED_CODE=403 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./02-check-unsecured.yaml + JAEGER_USERNAME=wronguser + JAEGER_PASSWORD=wrongpassword + EXPECTED_CODE=403 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./03-check-unauthorized.yaml + EXPECTED_CODE=200 + /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/assert-http-code.yaml.template -o ./04-check-authorized.yaml + skip_test examples-agent-as-daemonset 'This test is flaky in Prow CI' + '[' 2 -ne 2 ']' + test_name=examples-agent-as-daemonset + message='This test is flaky in Prow CI' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build/examples-openshift-with-htpasswd + '[' examples-openshift-with-htpasswd '!=' _build ']' + cd .. + rm -rf examples-agent-as-daemonset + warning 'examples-agent-as-daemonset: This test is flaky in Prow CI' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: examples-agent-as-daemonset: This test is flaky in Prow CI\e[0m' WAR: examples-agent-as-daemonset: This test is flaky in Prow CI + skip_test examples-with-badger-and-volume 'This test is flaky in Prow CI' + '[' 2 -ne 2 ']' + test_name=examples-with-badger-and-volume + message='This test is flaky in Prow CI' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build + '[' _build '!=' _build ']' + rm -rf examples-with-badger-and-volume + warning 'examples-with-badger-and-volume: This test is flaky in Prow CI' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: examples-with-badger-and-volume: This test is flaky in Prow CI\e[0m' WAR: examples-with-badger-and-volume: This test is flaky in Prow CI + skip_test examples-collector-with-priority-class 'This test is flaky in Prow CI' + '[' 2 -ne 2 ']' + test_name=examples-collector-with-priority-class + message='This test is flaky in Prow CI' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/examples/_build + '[' _build '!=' _build ']' + rm -rf examples-collector-with-priority-class + warning 'examples-collector-with-priority-class: This test is flaky in Prow CI' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: examples-collector-with-priority-class: This test is flaky in Prow CI\e[0m' WAR: examples-collector-with-priority-class: This test is flaky in Prow CI make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running examples E2E tests' Running examples E2E tests + cd tests/e2e/examples/_build + set +e + KUBECONFIG=/tmp/kubeconfig-591533165 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-54q9kfc5-9210b.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 12 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === RUN kuttl/harness/examples-agent-with-priority-class === PAUSE kuttl/harness/examples-agent-with-priority-class === RUN kuttl/harness/examples-all-in-one-with-options === PAUSE kuttl/harness/examples-all-in-one-with-options === RUN kuttl/harness/examples-business-application-injected-sidecar === PAUSE kuttl/harness/examples-business-application-injected-sidecar === RUN kuttl/harness/examples-openshift-with-htpasswd === PAUSE kuttl/harness/examples-openshift-with-htpasswd === RUN kuttl/harness/examples-service-types === PAUSE kuttl/harness/examples-service-types === RUN kuttl/harness/examples-simple-prod === PAUSE kuttl/harness/examples-simple-prod === RUN kuttl/harness/examples-simple-prod-with-volumes === PAUSE kuttl/harness/examples-simple-prod-with-volumes === RUN kuttl/harness/examples-simplest === PAUSE kuttl/harness/examples-simplest === RUN kuttl/harness/examples-with-badger === PAUSE kuttl/harness/examples-with-badger === RUN kuttl/harness/examples-with-cassandra === PAUSE kuttl/harness/examples-with-cassandra === RUN kuttl/harness/examples-with-sampling === PAUSE kuttl/harness/examples-with-sampling === CONT kuttl/harness/artifacts logger.go:42: 07:39:14 | artifacts | Creating namespace: kuttl-test-fancy-louse logger.go:42: 07:39:14 | artifacts | artifacts events from ns kuttl-test-fancy-louse: logger.go:42: 07:39:14 | artifacts | Deleting namespace: kuttl-test-fancy-louse === CONT kuttl/harness/examples-simple-prod logger.go:42: 07:39:20 | examples-simple-prod | Creating namespace: kuttl-test-curious-stork logger.go:42: 07:39:20 | examples-simple-prod/1-install | starting test step 1-install logger.go:42: 07:39:20 | examples-simple-prod/1-install | Jaeger:kuttl-test-curious-stork/simple-prod created logger.go:42: 07:39:58 | examples-simple-prod/1-install | test step completed 1-install logger.go:42: 07:39:58 | examples-simple-prod/2-smoke-test | starting test step 2-smoke-test logger.go:42: 07:39:58 | examples-simple-prod/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE simple-prod /dev/null] logger.go:42: 07:40:00 | examples-simple-prod/2-smoke-test | Warning: resource jaegers/simple-prod is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:40:06 | examples-simple-prod/2-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:40:06 | examples-simple-prod/2-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:40:07 | examples-simple-prod/2-smoke-test | job.batch/report-span created logger.go:42: 07:40:07 | examples-simple-prod/2-smoke-test | job.batch/check-span created logger.go:42: 07:40:20 | examples-simple-prod/2-smoke-test | test step completed 2-smoke-test logger.go:42: 07:40:20 | examples-simple-prod | examples-simple-prod events from ns kuttl-test-curious-stork: logger.go:42: 07:40:20 | examples-simple-prod | 2024-12-09 07:39:28 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttltestcuriousstorksimpleprod-1-5fcb746bd5 SuccessfulCreate Created pod: elasticsearch-cdm-kuttltestcuriousstorksimpleprod-1-5fcb74pv8mm replicaset-controller logger.go:42: 07:40:20 | examples-simple-prod | 2024-12-09 07:39:28 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestcuriousstorksimpleprod-1-5fcb74pv8mm Binding Scheduled Successfully assigned kuttl-test-curious-stork/elasticsearch-cdm-kuttltestcuriousstorksimpleprod-1-5fcb74pv8mm to ip-10-0-95-250.ec2.internal default-scheduler logger.go:42: 07:40:20 | examples-simple-prod | 2024-12-09 07:39:28 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttltestcuriousstorksimpleprod-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttltestcuriousstorksimpleprod-1-5fcb746bd5 to 1 deployment-controller logger.go:42: 07:40:20 | examples-simple-prod | 2024-12-09 07:39:29 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestcuriousstorksimpleprod-1-5fcb74pv8mm AddedInterface Add eth0 [10.131.0.39/23] from ovn-kubernetes multus logger.go:42: 07:40:20 | examples-simple-prod | 2024-12-09 07:39:29 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestcuriousstorksimpleprod-1-5fcb74pv8mm.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:8e4fbea4983cd58352349ca291383169b286bc166fad95a87807552ca43335e6" already present on machine kubelet logger.go:42: 07:40:20 | examples-simple-prod | 2024-12-09 07:39:29 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestcuriousstorksimpleprod-1-5fcb74pv8mm.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:40:20 | examples-simple-prod | 2024-12-09 07:39:29 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestcuriousstorksimpleprod-1-5fcb74pv8mm.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:40:20 | examples-simple-prod | 2024-12-09 07:39:29 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestcuriousstorksimpleprod-1-5fcb74pv8mm.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:d68824b0b2c84db8e33edf9ab344eb684c4a7ebd7ef162bbc309043adcb28e6b" already present on machine kubelet logger.go:42: 07:40:20 | examples-simple-prod | 2024-12-09 07:39:29 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestcuriousstorksimpleprod-1-5fcb74pv8mm.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:40:20 | examples-simple-prod | 2024-12-09 07:39:29 +0000 UTC Normal Pod elasticsearch-cdm-kuttltestcuriousstorksimpleprod-1-5fcb74pv8mm.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:40:20 | examples-simple-prod | 2024-12-09 07:39:43 +0000 UTC Warning Pod elasticsearch-cdm-kuttltestcuriousstorksimpleprod-1-5fcb74pv8mm.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:40:20 | examples-simple-prod | 2024-12-09 07:39:55 +0000 UTC Normal Pod simple-prod-collector-678cb75678-fpq7c Binding Scheduled Successfully assigned kuttl-test-curious-stork/simple-prod-collector-678cb75678-fpq7c to ip-10-0-79-227.ec2.internal default-scheduler logger.go:42: 07:40:20 | examples-simple-prod | 2024-12-09 07:39:55 +0000 UTC Normal Pod simple-prod-collector-678cb75678-fpq7c AddedInterface Add eth0 [10.129.2.75/23] from ovn-kubernetes multus logger.go:42: 07:40:20 | examples-simple-prod | 2024-12-09 07:39:55 +0000 UTC Normal Pod simple-prod-collector-678cb75678-fpq7c.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:44686d560aa27cae8ff8693f88c4cb6e2edf1737010ec1f80709cb42250b729d" already present on machine kubelet logger.go:42: 07:40:20 | examples-simple-prod | 2024-12-09 07:39:55 +0000 UTC Normal Pod simple-prod-collector-678cb75678-fpq7c.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:40:20 | examples-simple-prod | 2024-12-09 07:39:55 +0000 UTC Normal Pod simple-prod-collector-678cb75678-fpq7c.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:40:20 | examples-simple-prod | 2024-12-09 07:39:55 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-678cb75678 SuccessfulCreate Created pod: simple-prod-collector-678cb75678-fpq7c replicaset-controller logger.go:42: 07:40:20 | examples-simple-prod | 2024-12-09 07:39:55 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-678cb75678 to 1 deployment-controller logger.go:42: 07:40:20 | examples-simple-prod | 2024-12-09 07:39:55 +0000 UTC Normal Pod simple-prod-query-589d6c7bc8-52ndx Binding Scheduled Successfully assigned kuttl-test-curious-stork/simple-prod-query-589d6c7bc8-52ndx to ip-10-0-79-227.ec2.internal default-scheduler logger.go:42: 07:40:20 | examples-simple-prod | 2024-12-09 07:39:55 +0000 UTC Warning Pod simple-prod-query-589d6c7bc8-52ndx FailedMount MountVolume.SetUp failed for volume "simple-prod-ui-oauth-proxy-tls" : secret "simple-prod-ui-oauth-proxy-tls" not found kubelet logger.go:42: 07:40:20 | examples-simple-prod | 2024-12-09 07:39:55 +0000 UTC Normal ReplicaSet.apps simple-prod-query-589d6c7bc8 SuccessfulCreate Created pod: simple-prod-query-589d6c7bc8-52ndx replicaset-controller logger.go:42: 07:40:20 | examples-simple-prod | 2024-12-09 07:39:55 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-589d6c7bc8 to 1 deployment-controller logger.go:42: 07:40:20 | examples-simple-prod | 2024-12-09 07:39:56 +0000 UTC Normal Pod simple-prod-query-589d6c7bc8-52ndx AddedInterface Add eth0 [10.129.2.76/23] from ovn-kubernetes multus logger.go:42: 07:40:20 | examples-simple-prod | 2024-12-09 07:39:56 +0000 UTC Normal Pod simple-prod-query-589d6c7bc8-52ndx.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:24a83160aa930e1b72c2a2442a33b28af2b06c24b058e09afcd0d495a8066d6d" already present on machine kubelet logger.go:42: 07:40:20 | examples-simple-prod | 2024-12-09 07:39:56 +0000 UTC Normal Pod simple-prod-query-589d6c7bc8-52ndx.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:40:20 | examples-simple-prod | 2024-12-09 07:39:56 +0000 UTC Normal Pod simple-prod-query-589d6c7bc8-52ndx.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:40:20 | examples-simple-prod | 2024-12-09 07:39:56 +0000 UTC Normal Pod simple-prod-query-589d6c7bc8-52ndx.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:40:20 | examples-simple-prod | 2024-12-09 07:39:56 +0000 UTC Normal Pod simple-prod-query-589d6c7bc8-52ndx.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:40:20 | examples-simple-prod | 2024-12-09 07:39:56 +0000 UTC Normal Pod simple-prod-query-589d6c7bc8-52ndx.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:40:20 | examples-simple-prod | 2024-12-09 07:39:56 +0000 UTC Normal Pod simple-prod-query-589d6c7bc8-52ndx.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" already present on machine kubelet logger.go:42: 07:40:20 | examples-simple-prod | 2024-12-09 07:39:56 +0000 UTC Normal Pod simple-prod-query-589d6c7bc8-52ndx.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:40:20 | examples-simple-prod | 2024-12-09 07:39:56 +0000 UTC Normal Pod simple-prod-query-589d6c7bc8-52ndx.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:40:20 | examples-simple-prod | 2024-12-09 07:40:02 +0000 UTC Normal Pod simple-prod-query-589d6c7bc8-52ndx.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:40:20 | examples-simple-prod | 2024-12-09 07:40:02 +0000 UTC Normal Pod simple-prod-query-589d6c7bc8-52ndx.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:40:20 | examples-simple-prod | 2024-12-09 07:40:02 +0000 UTC Normal ReplicaSet.apps simple-prod-query-589d6c7bc8 SuccessfulDelete Deleted pod: simple-prod-query-589d6c7bc8-52ndx replicaset-controller logger.go:42: 07:40:20 | examples-simple-prod | 2024-12-09 07:40:02 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled down replica set simple-prod-query-589d6c7bc8 to 0 from 1 deployment-controller logger.go:42: 07:40:20 | examples-simple-prod | 2024-12-09 07:40:03 +0000 UTC Normal Pod simple-prod-query-589d6c7bc8-52ndx.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:40:20 | examples-simple-prod | 2024-12-09 07:40:03 +0000 UTC Normal Pod simple-prod-query-84f7c59f58-l6pjm Binding Scheduled Successfully assigned kuttl-test-curious-stork/simple-prod-query-84f7c59f58-l6pjm to ip-10-0-79-227.ec2.internal default-scheduler logger.go:42: 07:40:20 | examples-simple-prod | 2024-12-09 07:40:03 +0000 UTC Normal ReplicaSet.apps simple-prod-query-84f7c59f58 SuccessfulCreate Created pod: simple-prod-query-84f7c59f58-l6pjm replicaset-controller logger.go:42: 07:40:20 | examples-simple-prod | 2024-12-09 07:40:03 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-84f7c59f58 to 1 deployment-controller logger.go:42: 07:40:20 | examples-simple-prod | 2024-12-09 07:40:04 +0000 UTC Normal Pod simple-prod-query-84f7c59f58-l6pjm AddedInterface Add eth0 [10.129.2.77/23] from ovn-kubernetes multus logger.go:42: 07:40:20 | examples-simple-prod | 2024-12-09 07:40:04 +0000 UTC Normal Pod simple-prod-query-84f7c59f58-l6pjm.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:24a83160aa930e1b72c2a2442a33b28af2b06c24b058e09afcd0d495a8066d6d" already present on machine kubelet logger.go:42: 07:40:20 | examples-simple-prod | 2024-12-09 07:40:04 +0000 UTC Normal Pod simple-prod-query-84f7c59f58-l6pjm.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:40:20 | examples-simple-prod | 2024-12-09 07:40:04 +0000 UTC Normal Pod simple-prod-query-84f7c59f58-l6pjm.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:40:20 | examples-simple-prod | 2024-12-09 07:40:04 +0000 UTC Normal Pod simple-prod-query-84f7c59f58-l6pjm.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:40:20 | examples-simple-prod | 2024-12-09 07:40:04 +0000 UTC Normal Pod simple-prod-query-84f7c59f58-l6pjm.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:40:20 | examples-simple-prod | 2024-12-09 07:40:04 +0000 UTC Normal Pod simple-prod-query-84f7c59f58-l6pjm.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:40:20 | examples-simple-prod | 2024-12-09 07:40:04 +0000 UTC Normal Pod simple-prod-query-84f7c59f58-l6pjm.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" already present on machine kubelet logger.go:42: 07:40:20 | examples-simple-prod | 2024-12-09 07:40:04 +0000 UTC Normal Pod simple-prod-query-84f7c59f58-l6pjm.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:40:20 | examples-simple-prod | 2024-12-09 07:40:04 +0000 UTC Normal Pod simple-prod-query-84f7c59f58-l6pjm.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:40:20 | examples-simple-prod | 2024-12-09 07:40:07 +0000 UTC Normal Pod check-span-wzqz2 Binding Scheduled Successfully assigned kuttl-test-curious-stork/check-span-wzqz2 to ip-10-0-39-223.ec2.internal default-scheduler logger.go:42: 07:40:20 | examples-simple-prod | 2024-12-09 07:40:07 +0000 UTC Normal Pod check-span-wzqz2 AddedInterface Add eth0 [10.128.2.65/23] from ovn-kubernetes multus logger.go:42: 07:40:20 | examples-simple-prod | 2024-12-09 07:40:07 +0000 UTC Normal Pod check-span-wzqz2.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:40:20 | examples-simple-prod | 2024-12-09 07:40:07 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-wzqz2 job-controller logger.go:42: 07:40:20 | examples-simple-prod | 2024-12-09 07:40:07 +0000 UTC Normal Pod report-span-v7x4q Binding Scheduled Successfully assigned kuttl-test-curious-stork/report-span-v7x4q to ip-10-0-39-223.ec2.internal default-scheduler logger.go:42: 07:40:20 | examples-simple-prod | 2024-12-09 07:40:07 +0000 UTC Normal Pod report-span-v7x4q AddedInterface Add eth0 [10.128.2.64/23] from ovn-kubernetes multus logger.go:42: 07:40:20 | examples-simple-prod | 2024-12-09 07:40:07 +0000 UTC Normal Pod report-span-v7x4q.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:40:20 | examples-simple-prod | 2024-12-09 07:40:07 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-v7x4q job-controller logger.go:42: 07:40:20 | examples-simple-prod | 2024-12-09 07:40:09 +0000 UTC Normal Pod check-span-wzqz2.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 1.961s (1.961s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:40:20 | examples-simple-prod | 2024-12-09 07:40:09 +0000 UTC Normal Pod check-span-wzqz2.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:40:20 | examples-simple-prod | 2024-12-09 07:40:09 +0000 UTC Normal Pod check-span-wzqz2.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:40:20 | examples-simple-prod | 2024-12-09 07:40:09 +0000 UTC Normal Pod report-span-v7x4q.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 1.932s (1.932s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:40:20 | examples-simple-prod | 2024-12-09 07:40:09 +0000 UTC Normal Pod report-span-v7x4q.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:40:20 | examples-simple-prod | 2024-12-09 07:40:09 +0000 UTC Normal Pod report-span-v7x4q.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:40:20 | examples-simple-prod | 2024-12-09 07:40:10 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:40:20 | examples-simple-prod | 2024-12-09 07:40:10 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:40:20 | examples-simple-prod | 2024-12-09 07:40:10 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:40:20 | examples-simple-prod | 2024-12-09 07:40:20 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:40:20 | examples-simple-prod | Deleting namespace: kuttl-test-curious-stork === CONT kuttl/harness/examples-with-sampling logger.go:42: 07:40:33 | examples-with-sampling | Creating namespace: kuttl-test-immense-tahr logger.go:42: 07:40:33 | examples-with-sampling/0-install | starting test step 0-install logger.go:42: 07:40:33 | examples-with-sampling/0-install | running command: [sh -c cd /tmp/jaeger-tests && make cassandra STORAGE_NAMESPACE=$NAMESPACE] logger.go:42: 07:40:33 | examples-with-sampling/0-install | make[2]: Entering directory '/tmp/jaeger-tests' logger.go:42: 07:40:33 | examples-with-sampling/0-install | >>>> Creating namespace kuttl-test-immense-tahr logger.go:42: 07:40:33 | examples-with-sampling/0-install | kubectl create namespace kuttl-test-immense-tahr 2>&1 | grep -v "already exists" || true logger.go:42: 07:40:33 | examples-with-sampling/0-install | kubectl create -f ./tests/cassandra.yml --namespace kuttl-test-immense-tahr 2>&1 | grep -v "already exists" || true logger.go:42: 07:40:33 | examples-with-sampling/0-install | service/cassandra created logger.go:42: 07:40:33 | examples-with-sampling/0-install | statefulset.apps/cassandra created logger.go:42: 07:40:33 | examples-with-sampling/0-install | make[2]: Leaving directory '/tmp/jaeger-tests' logger.go:42: 07:40:42 | examples-with-sampling/0-install | test step completed 0-install logger.go:42: 07:40:42 | examples-with-sampling/1-install | starting test step 1-install logger.go:42: 07:40:42 | examples-with-sampling/1-install | Jaeger:kuttl-test-immense-tahr/with-sampling created logger.go:42: 07:40:50 | examples-with-sampling/1-install | test step completed 1-install logger.go:42: 07:40:50 | examples-with-sampling/2-smoke-test | starting test step 2-smoke-test logger.go:42: 07:40:50 | examples-with-sampling/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE with-sampling /dev/null] logger.go:42: 07:40:51 | examples-with-sampling/2-smoke-test | Warning: resource jaegers/with-sampling is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:40:58 | examples-with-sampling/2-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://with-sampling-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://with-sampling-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:40:58 | examples-with-sampling/2-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:40:58 | examples-with-sampling/2-smoke-test | job.batch/report-span created logger.go:42: 07:40:58 | examples-with-sampling/2-smoke-test | job.batch/check-span created logger.go:42: 07:41:11 | examples-with-sampling/2-smoke-test | test step completed 2-smoke-test logger.go:42: 07:41:11 | examples-with-sampling/3- | starting test step 3- logger.go:42: 07:41:11 | examples-with-sampling/3- | test step completed 3- logger.go:42: 07:41:11 | examples-with-sampling | examples-with-sampling events from ns kuttl-test-immense-tahr: logger.go:42: 07:41:11 | examples-with-sampling | 2024-12-09 07:40:33 +0000 UTC Normal Pod cassandra-0 Binding Scheduled Successfully assigned kuttl-test-immense-tahr/cassandra-0 to ip-10-0-79-227.ec2.internal default-scheduler logger.go:42: 07:41:11 | examples-with-sampling | 2024-12-09 07:40:33 +0000 UTC Normal StatefulSet.apps cassandra SuccessfulCreate create Pod cassandra-0 in StatefulSet cassandra successful statefulset-controller logger.go:42: 07:41:11 | examples-with-sampling | 2024-12-09 07:40:34 +0000 UTC Normal Pod cassandra-0 AddedInterface Add eth0 [10.129.2.78/23] from ovn-kubernetes multus logger.go:42: 07:41:11 | examples-with-sampling | 2024-12-09 07:40:34 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Pulling Pulling image "cassandra:3.11" kubelet logger.go:42: 07:41:11 | examples-with-sampling | 2024-12-09 07:40:37 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Pulled Successfully pulled image "cassandra:3.11" in 3.553s (3.553s including waiting). Image size: 309817322 bytes. kubelet logger.go:42: 07:41:11 | examples-with-sampling | 2024-12-09 07:40:37 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Created Created container cassandra kubelet logger.go:42: 07:41:11 | examples-with-sampling | 2024-12-09 07:40:37 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Started Started container cassandra kubelet logger.go:42: 07:41:11 | examples-with-sampling | 2024-12-09 07:40:38 +0000 UTC Normal Pod cassandra-1 Binding Scheduled Successfully assigned kuttl-test-immense-tahr/cassandra-1 to ip-10-0-39-223.ec2.internal default-scheduler logger.go:42: 07:41:11 | examples-with-sampling | 2024-12-09 07:40:38 +0000 UTC Normal Pod cassandra-1 AddedInterface Add eth0 [10.128.2.66/23] from ovn-kubernetes multus logger.go:42: 07:41:11 | examples-with-sampling | 2024-12-09 07:40:38 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Pulling Pulling image "cassandra:3.11" kubelet logger.go:42: 07:41:11 | examples-with-sampling | 2024-12-09 07:40:38 +0000 UTC Normal StatefulSet.apps cassandra SuccessfulCreate create Pod cassandra-1 in StatefulSet cassandra successful statefulset-controller logger.go:42: 07:41:11 | examples-with-sampling | 2024-12-09 07:40:42 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Pulled Successfully pulled image "cassandra:3.11" in 3.406s (3.406s including waiting). Image size: 309817322 bytes. kubelet logger.go:42: 07:41:11 | examples-with-sampling | 2024-12-09 07:40:42 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Created Created container cassandra kubelet logger.go:42: 07:41:11 | examples-with-sampling | 2024-12-09 07:40:42 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Started Started container cassandra kubelet logger.go:42: 07:41:11 | examples-with-sampling | 2024-12-09 07:40:46 +0000 UTC Normal Pod with-sampling-5657cccd86-nnlj5 Binding Scheduled Successfully assigned kuttl-test-immense-tahr/with-sampling-5657cccd86-nnlj5 to ip-10-0-79-227.ec2.internal default-scheduler logger.go:42: 07:41:11 | examples-with-sampling | 2024-12-09 07:40:46 +0000 UTC Normal ReplicaSet.apps with-sampling-5657cccd86 SuccessfulCreate Created pod: with-sampling-5657cccd86-nnlj5 replicaset-controller logger.go:42: 07:41:11 | examples-with-sampling | 2024-12-09 07:40:46 +0000 UTC Normal Deployment.apps with-sampling ScalingReplicaSet Scaled up replica set with-sampling-5657cccd86 to 1 deployment-controller logger.go:42: 07:41:11 | examples-with-sampling | 2024-12-09 07:40:47 +0000 UTC Normal Pod with-sampling-5657cccd86-nnlj5 AddedInterface Add eth0 [10.129.2.79/23] from ovn-kubernetes multus logger.go:42: 07:41:11 | examples-with-sampling | 2024-12-09 07:40:47 +0000 UTC Normal Pod with-sampling-5657cccd86-nnlj5.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1fa5ed13cc8a023f1e987f6cafe86adc195c373cc2b774539df6d0fd02b780a7" already present on machine kubelet logger.go:42: 07:41:11 | examples-with-sampling | 2024-12-09 07:40:47 +0000 UTC Normal Pod with-sampling-5657cccd86-nnlj5.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:41:11 | examples-with-sampling | 2024-12-09 07:40:47 +0000 UTC Normal Pod with-sampling-5657cccd86-nnlj5.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:41:11 | examples-with-sampling | 2024-12-09 07:40:47 +0000 UTC Normal Pod with-sampling-5657cccd86-nnlj5.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:41:11 | examples-with-sampling | 2024-12-09 07:40:47 +0000 UTC Normal Pod with-sampling-5657cccd86-nnlj5.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:41:11 | examples-with-sampling | 2024-12-09 07:40:47 +0000 UTC Normal Pod with-sampling-5657cccd86-nnlj5.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:41:11 | examples-with-sampling | 2024-12-09 07:40:56 +0000 UTC Normal Pod with-sampling-5657cccd86-nnlj5.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:41:11 | examples-with-sampling | 2024-12-09 07:40:56 +0000 UTC Normal Pod with-sampling-5657cccd86-nnlj5.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:41:11 | examples-with-sampling | 2024-12-09 07:40:56 +0000 UTC Normal ReplicaSet.apps with-sampling-5657cccd86 SuccessfulDelete Deleted pod: with-sampling-5657cccd86-nnlj5 replicaset-controller logger.go:42: 07:41:11 | examples-with-sampling | 2024-12-09 07:40:56 +0000 UTC Normal Deployment.apps with-sampling ScalingReplicaSet Scaled down replica set with-sampling-5657cccd86 to 0 from 1 deployment-controller logger.go:42: 07:41:11 | examples-with-sampling | 2024-12-09 07:40:57 +0000 UTC Normal Pod with-sampling-64db5bb87-vd5cs Binding Scheduled Successfully assigned kuttl-test-immense-tahr/with-sampling-64db5bb87-vd5cs to ip-10-0-79-227.ec2.internal default-scheduler logger.go:42: 07:41:11 | examples-with-sampling | 2024-12-09 07:40:57 +0000 UTC Normal Pod with-sampling-64db5bb87-vd5cs AddedInterface Add eth0 [10.129.2.80/23] from ovn-kubernetes multus logger.go:42: 07:41:11 | examples-with-sampling | 2024-12-09 07:40:57 +0000 UTC Normal Pod with-sampling-64db5bb87-vd5cs.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1fa5ed13cc8a023f1e987f6cafe86adc195c373cc2b774539df6d0fd02b780a7" already present on machine kubelet logger.go:42: 07:41:11 | examples-with-sampling | 2024-12-09 07:40:57 +0000 UTC Normal Pod with-sampling-64db5bb87-vd5cs.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:41:11 | examples-with-sampling | 2024-12-09 07:40:57 +0000 UTC Normal Pod with-sampling-64db5bb87-vd5cs.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:41:11 | examples-with-sampling | 2024-12-09 07:40:57 +0000 UTC Normal Pod with-sampling-64db5bb87-vd5cs.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:41:11 | examples-with-sampling | 2024-12-09 07:40:57 +0000 UTC Normal Pod with-sampling-64db5bb87-vd5cs.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:41:11 | examples-with-sampling | 2024-12-09 07:40:57 +0000 UTC Normal Pod with-sampling-64db5bb87-vd5cs.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:41:11 | examples-with-sampling | 2024-12-09 07:40:57 +0000 UTC Normal ReplicaSet.apps with-sampling-64db5bb87 SuccessfulCreate Created pod: with-sampling-64db5bb87-vd5cs replicaset-controller logger.go:42: 07:41:11 | examples-with-sampling | 2024-12-09 07:40:57 +0000 UTC Normal Deployment.apps with-sampling ScalingReplicaSet Scaled up replica set with-sampling-64db5bb87 to 1 deployment-controller logger.go:42: 07:41:11 | examples-with-sampling | 2024-12-09 07:40:58 +0000 UTC Normal Pod check-span-wtjhl Binding Scheduled Successfully assigned kuttl-test-immense-tahr/check-span-wtjhl to ip-10-0-95-250.ec2.internal default-scheduler logger.go:42: 07:41:11 | examples-with-sampling | 2024-12-09 07:40:58 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-wtjhl job-controller logger.go:42: 07:41:11 | examples-with-sampling | 2024-12-09 07:40:58 +0000 UTC Normal Pod report-span-2999g Binding Scheduled Successfully assigned kuttl-test-immense-tahr/report-span-2999g to ip-10-0-95-250.ec2.internal default-scheduler logger.go:42: 07:41:11 | examples-with-sampling | 2024-12-09 07:40:58 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-2999g job-controller logger.go:42: 07:41:11 | examples-with-sampling | 2024-12-09 07:40:59 +0000 UTC Normal Pod check-span-wtjhl AddedInterface Add eth0 [10.131.0.41/23] from ovn-kubernetes multus logger.go:42: 07:41:11 | examples-with-sampling | 2024-12-09 07:40:59 +0000 UTC Normal Pod check-span-wtjhl.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:41:11 | examples-with-sampling | 2024-12-09 07:40:59 +0000 UTC Normal Pod report-span-2999g AddedInterface Add eth0 [10.131.0.40/23] from ovn-kubernetes multus logger.go:42: 07:41:11 | examples-with-sampling | 2024-12-09 07:40:59 +0000 UTC Normal Pod report-span-2999g.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:41:11 | examples-with-sampling | 2024-12-09 07:41:00 +0000 UTC Normal Pod check-span-wtjhl.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 613ms (613ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:41:11 | examples-with-sampling | 2024-12-09 07:41:00 +0000 UTC Normal Pod check-span-wtjhl.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:41:11 | examples-with-sampling | 2024-12-09 07:41:00 +0000 UTC Normal Pod check-span-wtjhl.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:41:11 | examples-with-sampling | 2024-12-09 07:41:03 +0000 UTC Normal Pod report-span-2999g.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 4.432s (4.432s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:41:11 | examples-with-sampling | 2024-12-09 07:41:03 +0000 UTC Normal Pod report-span-2999g.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:41:11 | examples-with-sampling | 2024-12-09 07:41:03 +0000 UTC Normal Pod report-span-2999g.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:41:11 | examples-with-sampling | 2024-12-09 07:41:10 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:41:11 | examples-with-sampling | Deleting namespace: kuttl-test-immense-tahr === CONT kuttl/harness/examples-with-cassandra logger.go:42: 07:41:29 | examples-with-cassandra | Creating namespace: kuttl-test-united-ox logger.go:42: 07:41:29 | examples-with-cassandra/0-install | starting test step 0-install logger.go:42: 07:41:29 | examples-with-cassandra/0-install | running command: [sh -c cd /tmp/jaeger-tests && make cassandra STORAGE_NAMESPACE=$NAMESPACE] logger.go:42: 07:41:29 | examples-with-cassandra/0-install | make[2]: Entering directory '/tmp/jaeger-tests' logger.go:42: 07:41:29 | examples-with-cassandra/0-install | >>>> Creating namespace kuttl-test-united-ox logger.go:42: 07:41:29 | examples-with-cassandra/0-install | kubectl create namespace kuttl-test-united-ox 2>&1 | grep -v "already exists" || true logger.go:42: 07:41:29 | examples-with-cassandra/0-install | kubectl create -f ./tests/cassandra.yml --namespace kuttl-test-united-ox 2>&1 | grep -v "already exists" || true logger.go:42: 07:41:29 | examples-with-cassandra/0-install | service/cassandra created logger.go:42: 07:41:29 | examples-with-cassandra/0-install | statefulset.apps/cassandra created logger.go:42: 07:41:29 | examples-with-cassandra/0-install | make[2]: Leaving directory '/tmp/jaeger-tests' logger.go:42: 07:41:35 | examples-with-cassandra/0-install | test step completed 0-install logger.go:42: 07:41:35 | examples-with-cassandra/1-install | starting test step 1-install logger.go:42: 07:41:35 | examples-with-cassandra/1-install | Jaeger:kuttl-test-united-ox/with-cassandra created logger.go:42: 07:41:54 | examples-with-cassandra/1-install | test step completed 1-install logger.go:42: 07:41:54 | examples-with-cassandra/2-smoke-test | starting test step 2-smoke-test logger.go:42: 07:41:54 | examples-with-cassandra/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE with-cassandra /dev/null] logger.go:42: 07:41:55 | examples-with-cassandra/2-smoke-test | Warning: resource jaegers/with-cassandra is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:42:01 | examples-with-cassandra/2-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://with-cassandra-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://with-cassandra-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:42:02 | examples-with-cassandra/2-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:42:02 | examples-with-cassandra/2-smoke-test | job.batch/report-span created logger.go:42: 07:42:02 | examples-with-cassandra/2-smoke-test | job.batch/check-span created logger.go:42: 07:42:15 | examples-with-cassandra/2-smoke-test | test step completed 2-smoke-test logger.go:42: 07:42:15 | examples-with-cassandra | examples-with-cassandra events from ns kuttl-test-united-ox: logger.go:42: 07:42:15 | examples-with-cassandra | 2024-12-09 07:41:29 +0000 UTC Normal Pod cassandra-0 Binding Scheduled Successfully assigned kuttl-test-united-ox/cassandra-0 to ip-10-0-95-250.ec2.internal default-scheduler logger.go:42: 07:42:15 | examples-with-cassandra | 2024-12-09 07:41:29 +0000 UTC Normal StatefulSet.apps cassandra SuccessfulCreate create Pod cassandra-0 in StatefulSet cassandra successful statefulset-controller logger.go:42: 07:42:15 | examples-with-cassandra | 2024-12-09 07:41:30 +0000 UTC Normal Pod cassandra-0 AddedInterface Add eth0 [10.131.0.42/23] from ovn-kubernetes multus logger.go:42: 07:42:15 | examples-with-cassandra | 2024-12-09 07:41:30 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Pulling Pulling image "cassandra:3.11" kubelet logger.go:42: 07:42:15 | examples-with-cassandra | 2024-12-09 07:41:34 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Pulled Successfully pulled image "cassandra:3.11" in 3.48s (3.48s including waiting). Image size: 309817322 bytes. kubelet logger.go:42: 07:42:15 | examples-with-cassandra | 2024-12-09 07:41:34 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Created Created container cassandra kubelet logger.go:42: 07:42:15 | examples-with-cassandra | 2024-12-09 07:41:34 +0000 UTC Normal Pod cassandra-0.spec.containers{cassandra} Started Started container cassandra kubelet logger.go:42: 07:42:15 | examples-with-cassandra | 2024-12-09 07:41:34 +0000 UTC Normal Pod cassandra-1 Binding Scheduled Successfully assigned kuttl-test-united-ox/cassandra-1 to ip-10-0-39-223.ec2.internal default-scheduler logger.go:42: 07:42:15 | examples-with-cassandra | 2024-12-09 07:41:34 +0000 UTC Normal StatefulSet.apps cassandra SuccessfulCreate create Pod cassandra-1 in StatefulSet cassandra successful statefulset-controller logger.go:42: 07:42:15 | examples-with-cassandra | 2024-12-09 07:41:35 +0000 UTC Normal Pod cassandra-1 AddedInterface Add eth0 [10.128.2.67/23] from ovn-kubernetes multus logger.go:42: 07:42:15 | examples-with-cassandra | 2024-12-09 07:41:35 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Pulled Container image "cassandra:3.11" already present on machine kubelet logger.go:42: 07:42:15 | examples-with-cassandra | 2024-12-09 07:41:35 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Created Created container cassandra kubelet logger.go:42: 07:42:15 | examples-with-cassandra | 2024-12-09 07:41:35 +0000 UTC Normal Pod cassandra-1.spec.containers{cassandra} Started Started container cassandra kubelet logger.go:42: 07:42:15 | examples-with-cassandra | 2024-12-09 07:41:39 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-lfwzp Binding Scheduled Successfully assigned kuttl-test-united-ox/with-cassandra-cassandra-schema-job-lfwzp to ip-10-0-79-227.ec2.internal default-scheduler logger.go:42: 07:42:15 | examples-with-cassandra | 2024-12-09 07:41:39 +0000 UTC Normal Job.batch with-cassandra-cassandra-schema-job SuccessfulCreate Created pod: with-cassandra-cassandra-schema-job-lfwzp job-controller logger.go:42: 07:42:15 | examples-with-cassandra | 2024-12-09 07:41:40 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-lfwzp AddedInterface Add eth0 [10.129.2.81/23] from ovn-kubernetes multus logger.go:42: 07:42:15 | examples-with-cassandra | 2024-12-09 07:41:40 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-lfwzp.spec.containers{with-cassandra-cassandra-schema-job} Pulling Pulling image "jaegertracing/jaeger-cassandra-schema:1.57.0" kubelet logger.go:42: 07:42:15 | examples-with-cassandra | 2024-12-09 07:41:44 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-lfwzp.spec.containers{with-cassandra-cassandra-schema-job} Pulled Successfully pulled image "jaegertracing/jaeger-cassandra-schema:1.57.0" in 4.206s (4.206s including waiting). Image size: 381891613 bytes. kubelet logger.go:42: 07:42:15 | examples-with-cassandra | 2024-12-09 07:41:44 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-lfwzp.spec.containers{with-cassandra-cassandra-schema-job} Created Created container with-cassandra-cassandra-schema-job kubelet logger.go:42: 07:42:15 | examples-with-cassandra | 2024-12-09 07:41:44 +0000 UTC Normal Pod with-cassandra-cassandra-schema-job-lfwzp.spec.containers{with-cassandra-cassandra-schema-job} Started Started container with-cassandra-cassandra-schema-job kubelet logger.go:42: 07:42:15 | examples-with-cassandra | 2024-12-09 07:41:51 +0000 UTC Normal Pod with-cassandra-677cd69548-wzvbz Binding Scheduled Successfully assigned kuttl-test-united-ox/with-cassandra-677cd69548-wzvbz to ip-10-0-79-227.ec2.internal default-scheduler logger.go:42: 07:42:15 | examples-with-cassandra | 2024-12-09 07:41:51 +0000 UTC Normal ReplicaSet.apps with-cassandra-677cd69548 SuccessfulCreate Created pod: with-cassandra-677cd69548-wzvbz replicaset-controller logger.go:42: 07:42:15 | examples-with-cassandra | 2024-12-09 07:41:51 +0000 UTC Normal Job.batch with-cassandra-cassandra-schema-job Completed Job completed job-controller logger.go:42: 07:42:15 | examples-with-cassandra | 2024-12-09 07:41:51 +0000 UTC Normal Deployment.apps with-cassandra ScalingReplicaSet Scaled up replica set with-cassandra-677cd69548 to 1 deployment-controller logger.go:42: 07:42:15 | examples-with-cassandra | 2024-12-09 07:41:52 +0000 UTC Normal Pod with-cassandra-677cd69548-wzvbz AddedInterface Add eth0 [10.129.2.82/23] from ovn-kubernetes multus logger.go:42: 07:42:15 | examples-with-cassandra | 2024-12-09 07:41:52 +0000 UTC Normal Pod with-cassandra-677cd69548-wzvbz.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1fa5ed13cc8a023f1e987f6cafe86adc195c373cc2b774539df6d0fd02b780a7" already present on machine kubelet logger.go:42: 07:42:15 | examples-with-cassandra | 2024-12-09 07:41:52 +0000 UTC Normal Pod with-cassandra-677cd69548-wzvbz.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:42:15 | examples-with-cassandra | 2024-12-09 07:41:52 +0000 UTC Normal Pod with-cassandra-677cd69548-wzvbz.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:42:15 | examples-with-cassandra | 2024-12-09 07:41:52 +0000 UTC Normal Pod with-cassandra-677cd69548-wzvbz.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:42:15 | examples-with-cassandra | 2024-12-09 07:41:52 +0000 UTC Normal Pod with-cassandra-677cd69548-wzvbz.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:42:15 | examples-with-cassandra | 2024-12-09 07:41:52 +0000 UTC Normal Pod with-cassandra-677cd69548-wzvbz.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:42:15 | examples-with-cassandra | 2024-12-09 07:41:57 +0000 UTC Normal Pod with-cassandra-677cd69548-wzvbz.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:42:15 | examples-with-cassandra | 2024-12-09 07:41:57 +0000 UTC Normal Pod with-cassandra-677cd69548-wzvbz.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:42:15 | examples-with-cassandra | 2024-12-09 07:41:57 +0000 UTC Normal ReplicaSet.apps with-cassandra-677cd69548 SuccessfulDelete Deleted pod: with-cassandra-677cd69548-wzvbz replicaset-controller logger.go:42: 07:42:15 | examples-with-cassandra | 2024-12-09 07:41:57 +0000 UTC Normal Deployment.apps with-cassandra ScalingReplicaSet Scaled down replica set with-cassandra-677cd69548 to 0 from 1 deployment-controller logger.go:42: 07:42:15 | examples-with-cassandra | 2024-12-09 07:41:58 +0000 UTC Normal Pod with-cassandra-6df4f585d8-wt6w5 Binding Scheduled Successfully assigned kuttl-test-united-ox/with-cassandra-6df4f585d8-wt6w5 to ip-10-0-79-227.ec2.internal default-scheduler logger.go:42: 07:42:15 | examples-with-cassandra | 2024-12-09 07:41:58 +0000 UTC Normal Pod with-cassandra-6df4f585d8-wt6w5 AddedInterface Add eth0 [10.129.2.83/23] from ovn-kubernetes multus logger.go:42: 07:42:15 | examples-with-cassandra | 2024-12-09 07:41:58 +0000 UTC Normal Pod with-cassandra-6df4f585d8-wt6w5.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1fa5ed13cc8a023f1e987f6cafe86adc195c373cc2b774539df6d0fd02b780a7" already present on machine kubelet logger.go:42: 07:42:15 | examples-with-cassandra | 2024-12-09 07:41:58 +0000 UTC Normal ReplicaSet.apps with-cassandra-6df4f585d8 SuccessfulCreate Created pod: with-cassandra-6df4f585d8-wt6w5 replicaset-controller logger.go:42: 07:42:15 | examples-with-cassandra | 2024-12-09 07:41:58 +0000 UTC Normal Deployment.apps with-cassandra ScalingReplicaSet Scaled up replica set with-cassandra-6df4f585d8 to 1 deployment-controller logger.go:42: 07:42:15 | examples-with-cassandra | 2024-12-09 07:41:59 +0000 UTC Normal Pod with-cassandra-6df4f585d8-wt6w5.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:42:15 | examples-with-cassandra | 2024-12-09 07:41:59 +0000 UTC Normal Pod with-cassandra-6df4f585d8-wt6w5.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:42:15 | examples-with-cassandra | 2024-12-09 07:41:59 +0000 UTC Normal Pod with-cassandra-6df4f585d8-wt6w5.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:42:15 | examples-with-cassandra | 2024-12-09 07:41:59 +0000 UTC Normal Pod with-cassandra-6df4f585d8-wt6w5.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:42:15 | examples-with-cassandra | 2024-12-09 07:41:59 +0000 UTC Normal Pod with-cassandra-6df4f585d8-wt6w5.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:42:15 | examples-with-cassandra | 2024-12-09 07:42:02 +0000 UTC Normal Pod check-span-mhqfx Binding Scheduled Successfully assigned kuttl-test-united-ox/check-span-mhqfx to ip-10-0-95-250.ec2.internal default-scheduler logger.go:42: 07:42:15 | examples-with-cassandra | 2024-12-09 07:42:02 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-mhqfx job-controller logger.go:42: 07:42:15 | examples-with-cassandra | 2024-12-09 07:42:02 +0000 UTC Normal Pod report-span-r9fh8 Binding Scheduled Successfully assigned kuttl-test-united-ox/report-span-r9fh8 to ip-10-0-95-250.ec2.internal default-scheduler logger.go:42: 07:42:15 | examples-with-cassandra | 2024-12-09 07:42:02 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-r9fh8 job-controller logger.go:42: 07:42:15 | examples-with-cassandra | 2024-12-09 07:42:03 +0000 UTC Normal Pod check-span-mhqfx AddedInterface Add eth0 [10.131.0.44/23] from ovn-kubernetes multus logger.go:42: 07:42:15 | examples-with-cassandra | 2024-12-09 07:42:03 +0000 UTC Normal Pod check-span-mhqfx.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:42:15 | examples-with-cassandra | 2024-12-09 07:42:03 +0000 UTC Normal Pod report-span-r9fh8 AddedInterface Add eth0 [10.131.0.43/23] from ovn-kubernetes multus logger.go:42: 07:42:15 | examples-with-cassandra | 2024-12-09 07:42:03 +0000 UTC Normal Pod report-span-r9fh8.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:42:15 | examples-with-cassandra | 2024-12-09 07:42:04 +0000 UTC Normal Pod check-span-mhqfx.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 1.274s (1.274s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:42:15 | examples-with-cassandra | 2024-12-09 07:42:04 +0000 UTC Normal Pod check-span-mhqfx.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:42:15 | examples-with-cassandra | 2024-12-09 07:42:04 +0000 UTC Normal Pod check-span-mhqfx.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:42:15 | examples-with-cassandra | 2024-12-09 07:42:04 +0000 UTC Normal Pod report-span-r9fh8.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 855ms (855ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:42:15 | examples-with-cassandra | 2024-12-09 07:42:04 +0000 UTC Normal Pod report-span-r9fh8.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:42:15 | examples-with-cassandra | 2024-12-09 07:42:04 +0000 UTC Normal Pod report-span-r9fh8.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:42:15 | examples-with-cassandra | 2024-12-09 07:42:15 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:42:16 | examples-with-cassandra | Deleting namespace: kuttl-test-united-ox === CONT kuttl/harness/examples-with-badger logger.go:42: 07:42:28 | examples-with-badger | Creating namespace: kuttl-test-glowing-manatee logger.go:42: 07:42:28 | examples-with-badger/0-install | starting test step 0-install logger.go:42: 07:42:28 | examples-with-badger/0-install | Jaeger:kuttl-test-glowing-manatee/with-badger created logger.go:42: 07:42:34 | examples-with-badger/0-install | test step completed 0-install logger.go:42: 07:42:34 | examples-with-badger/1-smoke-test | starting test step 1-smoke-test logger.go:42: 07:42:34 | examples-with-badger/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE with-badger /dev/null] logger.go:42: 07:42:35 | examples-with-badger/1-smoke-test | Warning: resource jaegers/with-badger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:42:43 | examples-with-badger/1-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://with-badger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://with-badger-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:42:43 | examples-with-badger/1-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:42:44 | examples-with-badger/1-smoke-test | job.batch/report-span created logger.go:42: 07:42:44 | examples-with-badger/1-smoke-test | job.batch/check-span created logger.go:42: 07:42:56 | examples-with-badger/1-smoke-test | test step completed 1-smoke-test logger.go:42: 07:42:56 | examples-with-badger | examples-with-badger events from ns kuttl-test-glowing-manatee: logger.go:42: 07:42:56 | examples-with-badger | 2024-12-09 07:42:32 +0000 UTC Normal Pod with-badger-56c44bbb47-njp5h Binding Scheduled Successfully assigned kuttl-test-glowing-manatee/with-badger-56c44bbb47-njp5h to ip-10-0-79-227.ec2.internal default-scheduler logger.go:42: 07:42:56 | examples-with-badger | 2024-12-09 07:42:32 +0000 UTC Normal Pod with-badger-56c44bbb47-njp5h AddedInterface Add eth0 [10.129.2.84/23] from ovn-kubernetes multus logger.go:42: 07:42:56 | examples-with-badger | 2024-12-09 07:42:32 +0000 UTC Normal Pod with-badger-56c44bbb47-njp5h.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1fa5ed13cc8a023f1e987f6cafe86adc195c373cc2b774539df6d0fd02b780a7" already present on machine kubelet logger.go:42: 07:42:56 | examples-with-badger | 2024-12-09 07:42:32 +0000 UTC Normal Pod with-badger-56c44bbb47-njp5h.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:42:56 | examples-with-badger | 2024-12-09 07:42:32 +0000 UTC Normal ReplicaSet.apps with-badger-56c44bbb47 SuccessfulCreate Created pod: with-badger-56c44bbb47-njp5h replicaset-controller logger.go:42: 07:42:56 | examples-with-badger | 2024-12-09 07:42:32 +0000 UTC Normal Deployment.apps with-badger ScalingReplicaSet Scaled up replica set with-badger-56c44bbb47 to 1 deployment-controller logger.go:42: 07:42:56 | examples-with-badger | 2024-12-09 07:42:33 +0000 UTC Normal Pod with-badger-56c44bbb47-njp5h.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:42:56 | examples-with-badger | 2024-12-09 07:42:33 +0000 UTC Normal Pod with-badger-56c44bbb47-njp5h.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:42:56 | examples-with-badger | 2024-12-09 07:42:33 +0000 UTC Normal Pod with-badger-56c44bbb47-njp5h.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:42:56 | examples-with-badger | 2024-12-09 07:42:33 +0000 UTC Normal Pod with-badger-56c44bbb47-njp5h.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:42:56 | examples-with-badger | 2024-12-09 07:42:40 +0000 UTC Normal Pod with-badger-56c44bbb47-njp5h.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:42:56 | examples-with-badger | 2024-12-09 07:42:40 +0000 UTC Normal Pod with-badger-56c44bbb47-njp5h.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:42:56 | examples-with-badger | 2024-12-09 07:42:40 +0000 UTC Normal ReplicaSet.apps with-badger-56c44bbb47 SuccessfulDelete Deleted pod: with-badger-56c44bbb47-njp5h replicaset-controller logger.go:42: 07:42:56 | examples-with-badger | 2024-12-09 07:42:40 +0000 UTC Normal Deployment.apps with-badger ScalingReplicaSet Scaled down replica set with-badger-56c44bbb47 to 0 from 1 deployment-controller logger.go:42: 07:42:56 | examples-with-badger | 2024-12-09 07:42:41 +0000 UTC Normal Pod with-badger-5b758cd86f-md26n Binding Scheduled Successfully assigned kuttl-test-glowing-manatee/with-badger-5b758cd86f-md26n to ip-10-0-79-227.ec2.internal default-scheduler logger.go:42: 07:42:56 | examples-with-badger | 2024-12-09 07:42:41 +0000 UTC Normal ReplicaSet.apps with-badger-5b758cd86f SuccessfulCreate Created pod: with-badger-5b758cd86f-md26n replicaset-controller logger.go:42: 07:42:56 | examples-with-badger | 2024-12-09 07:42:41 +0000 UTC Normal Deployment.apps with-badger ScalingReplicaSet Scaled up replica set with-badger-5b758cd86f to 1 deployment-controller logger.go:42: 07:42:56 | examples-with-badger | 2024-12-09 07:42:42 +0000 UTC Normal Pod with-badger-5b758cd86f-md26n AddedInterface Add eth0 [10.129.2.85/23] from ovn-kubernetes multus logger.go:42: 07:42:56 | examples-with-badger | 2024-12-09 07:42:42 +0000 UTC Normal Pod with-badger-5b758cd86f-md26n.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1fa5ed13cc8a023f1e987f6cafe86adc195c373cc2b774539df6d0fd02b780a7" already present on machine kubelet logger.go:42: 07:42:56 | examples-with-badger | 2024-12-09 07:42:42 +0000 UTC Normal Pod with-badger-5b758cd86f-md26n.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:42:56 | examples-with-badger | 2024-12-09 07:42:42 +0000 UTC Normal Pod with-badger-5b758cd86f-md26n.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:42:56 | examples-with-badger | 2024-12-09 07:42:42 +0000 UTC Normal Pod with-badger-5b758cd86f-md26n.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:42:56 | examples-with-badger | 2024-12-09 07:42:42 +0000 UTC Normal Pod with-badger-5b758cd86f-md26n.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:42:56 | examples-with-badger | 2024-12-09 07:42:42 +0000 UTC Normal Pod with-badger-5b758cd86f-md26n.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:42:56 | examples-with-badger | 2024-12-09 07:42:44 +0000 UTC Normal Pod check-span-r54rw Binding Scheduled Successfully assigned kuttl-test-glowing-manatee/check-span-r54rw to ip-10-0-39-223.ec2.internal default-scheduler logger.go:42: 07:42:56 | examples-with-badger | 2024-12-09 07:42:44 +0000 UTC Normal Pod check-span-r54rw AddedInterface Add eth0 [10.128.2.68/23] from ovn-kubernetes multus logger.go:42: 07:42:56 | examples-with-badger | 2024-12-09 07:42:44 +0000 UTC Normal Pod check-span-r54rw.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:42:56 | examples-with-badger | 2024-12-09 07:42:44 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-r54rw job-controller logger.go:42: 07:42:56 | examples-with-badger | 2024-12-09 07:42:44 +0000 UTC Normal Pod report-span-dwzsj Binding Scheduled Successfully assigned kuttl-test-glowing-manatee/report-span-dwzsj to ip-10-0-95-250.ec2.internal default-scheduler logger.go:42: 07:42:56 | examples-with-badger | 2024-12-09 07:42:44 +0000 UTC Normal Pod report-span-dwzsj AddedInterface Add eth0 [10.131.0.45/23] from ovn-kubernetes multus logger.go:42: 07:42:56 | examples-with-badger | 2024-12-09 07:42:44 +0000 UTC Normal Pod report-span-dwzsj.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:42:56 | examples-with-badger | 2024-12-09 07:42:44 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-dwzsj job-controller logger.go:42: 07:42:56 | examples-with-badger | 2024-12-09 07:42:45 +0000 UTC Normal Pod check-span-r54rw.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 709ms (709ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:42:56 | examples-with-badger | 2024-12-09 07:42:45 +0000 UTC Normal Pod check-span-r54rw.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:42:56 | examples-with-badger | 2024-12-09 07:42:45 +0000 UTC Normal Pod check-span-r54rw.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:42:56 | examples-with-badger | 2024-12-09 07:42:45 +0000 UTC Normal Pod report-span-dwzsj.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 1.09s (1.09s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:42:56 | examples-with-badger | 2024-12-09 07:42:45 +0000 UTC Normal Pod report-span-dwzsj.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:42:56 | examples-with-badger | 2024-12-09 07:42:45 +0000 UTC Normal Pod report-span-dwzsj.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:42:56 | examples-with-badger | 2024-12-09 07:42:55 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:42:56 | examples-with-badger | Deleting namespace: kuttl-test-glowing-manatee === CONT kuttl/harness/examples-simplest logger.go:42: 07:43:09 | examples-simplest | Creating namespace: kuttl-test-vital-eel logger.go:42: 07:43:09 | examples-simplest/0-install | starting test step 0-install logger.go:42: 07:43:09 | examples-simplest/0-install | Jaeger:kuttl-test-vital-eel/simplest created logger.go:42: 07:43:15 | examples-simplest/0-install | test step completed 0-install logger.go:42: 07:43:15 | examples-simplest/1-smoke-test | starting test step 1-smoke-test logger.go:42: 07:43:15 | examples-simplest/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE simplest /dev/null] logger.go:42: 07:43:17 | examples-simplest/1-smoke-test | Warning: resource jaegers/simplest is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:43:24 | examples-simplest/1-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://simplest-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:43:25 | examples-simplest/1-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:43:25 | examples-simplest/1-smoke-test | job.batch/report-span created logger.go:42: 07:43:25 | examples-simplest/1-smoke-test | job.batch/check-span created logger.go:42: 07:43:38 | examples-simplest/1-smoke-test | test step completed 1-smoke-test logger.go:42: 07:43:38 | examples-simplest | examples-simplest events from ns kuttl-test-vital-eel: logger.go:42: 07:43:38 | examples-simplest | 2024-12-09 07:43:13 +0000 UTC Normal Pod simplest-59cc6446dc-jkmln Binding Scheduled Successfully assigned kuttl-test-vital-eel/simplest-59cc6446dc-jkmln to ip-10-0-79-227.ec2.internal default-scheduler logger.go:42: 07:43:38 | examples-simplest | 2024-12-09 07:43:13 +0000 UTC Normal ReplicaSet.apps simplest-59cc6446dc SuccessfulCreate Created pod: simplest-59cc6446dc-jkmln replicaset-controller logger.go:42: 07:43:38 | examples-simplest | 2024-12-09 07:43:13 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled up replica set simplest-59cc6446dc to 1 deployment-controller logger.go:42: 07:43:38 | examples-simplest | 2024-12-09 07:43:14 +0000 UTC Normal Pod simplest-59cc6446dc-jkmln AddedInterface Add eth0 [10.129.2.86/23] from ovn-kubernetes multus logger.go:42: 07:43:38 | examples-simplest | 2024-12-09 07:43:14 +0000 UTC Normal Pod simplest-59cc6446dc-jkmln.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1fa5ed13cc8a023f1e987f6cafe86adc195c373cc2b774539df6d0fd02b780a7" already present on machine kubelet logger.go:42: 07:43:38 | examples-simplest | 2024-12-09 07:43:14 +0000 UTC Normal Pod simplest-59cc6446dc-jkmln.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:43:38 | examples-simplest | 2024-12-09 07:43:14 +0000 UTC Normal Pod simplest-59cc6446dc-jkmln.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:43:38 | examples-simplest | 2024-12-09 07:43:14 +0000 UTC Normal Pod simplest-59cc6446dc-jkmln.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:43:38 | examples-simplest | 2024-12-09 07:43:14 +0000 UTC Normal Pod simplest-59cc6446dc-jkmln.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:43:38 | examples-simplest | 2024-12-09 07:43:14 +0000 UTC Normal Pod simplest-59cc6446dc-jkmln.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:43:38 | examples-simplest | 2024-12-09 07:43:21 +0000 UTC Normal Pod simplest-59cc6446dc-jkmln.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:43:38 | examples-simplest | 2024-12-09 07:43:21 +0000 UTC Normal Pod simplest-59cc6446dc-jkmln.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:43:38 | examples-simplest | 2024-12-09 07:43:21 +0000 UTC Normal ReplicaSet.apps simplest-59cc6446dc SuccessfulDelete Deleted pod: simplest-59cc6446dc-jkmln replicaset-controller logger.go:42: 07:43:38 | examples-simplest | 2024-12-09 07:43:21 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled down replica set simplest-59cc6446dc to 0 from 1 deployment-controller logger.go:42: 07:43:38 | examples-simplest | 2024-12-09 07:43:22 +0000 UTC Normal Pod simplest-678bfb8b59-d5xm9 Binding Scheduled Successfully assigned kuttl-test-vital-eel/simplest-678bfb8b59-d5xm9 to ip-10-0-79-227.ec2.internal default-scheduler logger.go:42: 07:43:38 | examples-simplest | 2024-12-09 07:43:22 +0000 UTC Normal ReplicaSet.apps simplest-678bfb8b59 SuccessfulCreate Created pod: simplest-678bfb8b59-d5xm9 replicaset-controller logger.go:42: 07:43:38 | examples-simplest | 2024-12-09 07:43:22 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled up replica set simplest-678bfb8b59 to 1 deployment-controller logger.go:42: 07:43:38 | examples-simplest | 2024-12-09 07:43:23 +0000 UTC Normal Pod simplest-678bfb8b59-d5xm9 AddedInterface Add eth0 [10.129.2.87/23] from ovn-kubernetes multus logger.go:42: 07:43:38 | examples-simplest | 2024-12-09 07:43:23 +0000 UTC Normal Pod simplest-678bfb8b59-d5xm9.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1fa5ed13cc8a023f1e987f6cafe86adc195c373cc2b774539df6d0fd02b780a7" already present on machine kubelet logger.go:42: 07:43:38 | examples-simplest | 2024-12-09 07:43:23 +0000 UTC Normal Pod simplest-678bfb8b59-d5xm9.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:43:38 | examples-simplest | 2024-12-09 07:43:23 +0000 UTC Normal Pod simplest-678bfb8b59-d5xm9.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:43:38 | examples-simplest | 2024-12-09 07:43:23 +0000 UTC Normal Pod simplest-678bfb8b59-d5xm9.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:43:38 | examples-simplest | 2024-12-09 07:43:23 +0000 UTC Normal Pod simplest-678bfb8b59-d5xm9.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:43:38 | examples-simplest | 2024-12-09 07:43:23 +0000 UTC Normal Pod simplest-678bfb8b59-d5xm9.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:43:38 | examples-simplest | 2024-12-09 07:43:25 +0000 UTC Normal Pod check-span-nf9tq Binding Scheduled Successfully assigned kuttl-test-vital-eel/check-span-nf9tq to ip-10-0-39-223.ec2.internal default-scheduler logger.go:42: 07:43:38 | examples-simplest | 2024-12-09 07:43:25 +0000 UTC Normal Pod check-span-nf9tq AddedInterface Add eth0 [10.128.2.69/23] from ovn-kubernetes multus logger.go:42: 07:43:38 | examples-simplest | 2024-12-09 07:43:25 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-nf9tq job-controller logger.go:42: 07:43:38 | examples-simplest | 2024-12-09 07:43:25 +0000 UTC Normal Pod report-span-d6g4c Binding Scheduled Successfully assigned kuttl-test-vital-eel/report-span-d6g4c to ip-10-0-95-250.ec2.internal default-scheduler logger.go:42: 07:43:38 | examples-simplest | 2024-12-09 07:43:25 +0000 UTC Normal Pod report-span-d6g4c AddedInterface Add eth0 [10.131.0.46/23] from ovn-kubernetes multus logger.go:42: 07:43:38 | examples-simplest | 2024-12-09 07:43:25 +0000 UTC Normal Pod report-span-d6g4c.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:43:38 | examples-simplest | 2024-12-09 07:43:25 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-d6g4c job-controller logger.go:42: 07:43:38 | examples-simplest | 2024-12-09 07:43:26 +0000 UTC Normal Pod check-span-nf9tq.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:43:38 | examples-simplest | 2024-12-09 07:43:26 +0000 UTC Normal Pod check-span-nf9tq.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 895ms (895ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:43:38 | examples-simplest | 2024-12-09 07:43:26 +0000 UTC Normal Pod check-span-nf9tq.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:43:38 | examples-simplest | 2024-12-09 07:43:26 +0000 UTC Normal Pod check-span-nf9tq.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:43:38 | examples-simplest | 2024-12-09 07:43:26 +0000 UTC Normal Pod report-span-d6g4c.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 926ms (926ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:43:38 | examples-simplest | 2024-12-09 07:43:26 +0000 UTC Normal Pod report-span-d6g4c.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:43:38 | examples-simplest | 2024-12-09 07:43:26 +0000 UTC Normal Pod report-span-d6g4c.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:43:38 | examples-simplest | 2024-12-09 07:43:37 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:43:38 | examples-simplest | Deleting namespace: kuttl-test-vital-eel === CONT kuttl/harness/examples-simple-prod-with-volumes logger.go:42: 07:43:44 | examples-simple-prod-with-volumes | Ignoring 03-check-volume.yaml.template as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:43:44 | examples-simple-prod-with-volumes | Creating namespace: kuttl-test-stunning-bobcat logger.go:42: 07:43:44 | examples-simple-prod-with-volumes/1-install | starting test step 1-install logger.go:42: 07:43:45 | examples-simple-prod-with-volumes/1-install | Jaeger:kuttl-test-stunning-bobcat/simple-prod created logger.go:42: 07:44:21 | examples-simple-prod-with-volumes/1-install | test step completed 1-install logger.go:42: 07:44:21 | examples-simple-prod-with-volumes/2-smoke-test | starting test step 2-smoke-test logger.go:42: 07:44:21 | examples-simple-prod-with-volumes/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE simple-prod /dev/null] logger.go:42: 07:44:22 | examples-simple-prod-with-volumes/2-smoke-test | Warning: resource jaegers/simple-prod is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:44:29 | examples-simple-prod-with-volumes/2-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://simple-prod-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://simple-prod-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:44:30 | examples-simple-prod-with-volumes/2-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:44:30 | examples-simple-prod-with-volumes/2-smoke-test | job.batch/report-span created logger.go:42: 07:44:30 | examples-simple-prod-with-volumes/2-smoke-test | job.batch/check-span created logger.go:42: 07:44:42 | examples-simple-prod-with-volumes/2-smoke-test | test step completed 2-smoke-test logger.go:42: 07:44:42 | examples-simple-prod-with-volumes/3-check-volume | starting test step 3-check-volume logger.go:42: 07:44:42 | examples-simple-prod-with-volumes/3-check-volume | running command: [sh -c kubectl exec $(kubectl get pods -n $NAMESPACE -l app=jaeger -l app.kubernetes.io/component=collector -o yaml | /tmp/jaeger-tests/bin/yq e '.items[0].metadata.name') -n $NAMESPACE -- ls /usr/share/elasticsearch/data] logger.go:42: 07:44:43 | examples-simple-prod-with-volumes/3-check-volume | test step completed 3-check-volume logger.go:42: 07:44:43 | examples-simple-prod-with-volumes | examples-simple-prod-with-volumes events from ns kuttl-test-stunning-bobcat: logger.go:42: 07:44:43 | examples-simple-prod-with-volumes | 2024-12-09 07:43:51 +0000 UTC Normal ReplicaSet.apps elasticsearch-cdm-kuttlteststunningbobcatsimpleprod-1-76d688b6b5 SuccessfulCreate Created pod: elasticsearch-cdm-kuttlteststunningbobcatsimpleprod-1-76d6dkhdr replicaset-controller logger.go:42: 07:44:43 | examples-simple-prod-with-volumes | 2024-12-09 07:43:51 +0000 UTC Normal Pod elasticsearch-cdm-kuttlteststunningbobcatsimpleprod-1-76d6dkhdr Binding Scheduled Successfully assigned kuttl-test-stunning-bobcat/elasticsearch-cdm-kuttlteststunningbobcatsimpleprod-1-76d6dkhdr to ip-10-0-95-250.ec2.internal default-scheduler logger.go:42: 07:44:43 | examples-simple-prod-with-volumes | 2024-12-09 07:43:51 +0000 UTC Normal Deployment.apps elasticsearch-cdm-kuttlteststunningbobcatsimpleprod-1 ScalingReplicaSet Scaled up replica set elasticsearch-cdm-kuttlteststunningbobcatsimpleprod-1-76d688b6b5 to 1 deployment-controller logger.go:42: 07:44:43 | examples-simple-prod-with-volumes | 2024-12-09 07:43:52 +0000 UTC Normal Pod elasticsearch-cdm-kuttlteststunningbobcatsimpleprod-1-76d6dkhdr AddedInterface Add eth0 [10.131.0.47/23] from ovn-kubernetes multus logger.go:42: 07:44:43 | examples-simple-prod-with-volumes | 2024-12-09 07:43:52 +0000 UTC Normal Pod elasticsearch-cdm-kuttlteststunningbobcatsimpleprod-1-76d6dkhdr.spec.containers{elasticsearch} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch6-rhel9@sha256:8e4fbea4983cd58352349ca291383169b286bc166fad95a87807552ca43335e6" already present on machine kubelet logger.go:42: 07:44:43 | examples-simple-prod-with-volumes | 2024-12-09 07:43:52 +0000 UTC Normal Pod elasticsearch-cdm-kuttlteststunningbobcatsimpleprod-1-76d6dkhdr.spec.containers{elasticsearch} Created Created container elasticsearch kubelet logger.go:42: 07:44:43 | examples-simple-prod-with-volumes | 2024-12-09 07:43:52 +0000 UTC Normal Pod elasticsearch-cdm-kuttlteststunningbobcatsimpleprod-1-76d6dkhdr.spec.containers{elasticsearch} Started Started container elasticsearch kubelet logger.go:42: 07:44:43 | examples-simple-prod-with-volumes | 2024-12-09 07:43:52 +0000 UTC Normal Pod elasticsearch-cdm-kuttlteststunningbobcatsimpleprod-1-76d6dkhdr.spec.containers{proxy} Pulled Container image "registry.redhat.io/openshift-logging/elasticsearch-proxy-rhel9@sha256:d68824b0b2c84db8e33edf9ab344eb684c4a7ebd7ef162bbc309043adcb28e6b" already present on machine kubelet logger.go:42: 07:44:43 | examples-simple-prod-with-volumes | 2024-12-09 07:43:52 +0000 UTC Normal Pod elasticsearch-cdm-kuttlteststunningbobcatsimpleprod-1-76d6dkhdr.spec.containers{proxy} Created Created container proxy kubelet logger.go:42: 07:44:43 | examples-simple-prod-with-volumes | 2024-12-09 07:43:52 +0000 UTC Normal Pod elasticsearch-cdm-kuttlteststunningbobcatsimpleprod-1-76d6dkhdr.spec.containers{proxy} Started Started container proxy kubelet logger.go:42: 07:44:43 | examples-simple-prod-with-volumes | 2024-12-09 07:44:02 +0000 UTC Warning Pod elasticsearch-cdm-kuttlteststunningbobcatsimpleprod-1-76d6dkhdr.spec.containers{elasticsearch} Unhealthy Readiness probe failed: Elasticsearch node is not ready to accept HTTP requests yet [response code: 000] kubelet logger.go:42: 07:44:43 | examples-simple-prod-with-volumes | 2024-12-09 07:44:07 +0000 UTC Warning Pod elasticsearch-cdm-kuttlteststunningbobcatsimpleprod-1-76d6dkhdr.spec.containers{elasticsearch} Unhealthy Readiness probe failed: kubelet logger.go:42: 07:44:43 | examples-simple-prod-with-volumes | 2024-12-09 07:44:18 +0000 UTC Normal Pod simple-prod-collector-c57855f48-2nssh Binding Scheduled Successfully assigned kuttl-test-stunning-bobcat/simple-prod-collector-c57855f48-2nssh to ip-10-0-79-227.ec2.internal default-scheduler logger.go:42: 07:44:43 | examples-simple-prod-with-volumes | 2024-12-09 07:44:18 +0000 UTC Normal ReplicaSet.apps simple-prod-collector-c57855f48 SuccessfulCreate Created pod: simple-prod-collector-c57855f48-2nssh replicaset-controller logger.go:42: 07:44:43 | examples-simple-prod-with-volumes | 2024-12-09 07:44:18 +0000 UTC Normal Deployment.apps simple-prod-collector ScalingReplicaSet Scaled up replica set simple-prod-collector-c57855f48 to 1 deployment-controller logger.go:42: 07:44:43 | examples-simple-prod-with-volumes | 2024-12-09 07:44:18 +0000 UTC Normal Pod simple-prod-query-6877c6d894-jcjx9 Binding Scheduled Successfully assigned kuttl-test-stunning-bobcat/simple-prod-query-6877c6d894-jcjx9 to ip-10-0-79-227.ec2.internal default-scheduler logger.go:42: 07:44:43 | examples-simple-prod-with-volumes | 2024-12-09 07:44:18 +0000 UTC Normal ReplicaSet.apps simple-prod-query-6877c6d894 SuccessfulCreate Created pod: simple-prod-query-6877c6d894-jcjx9 replicaset-controller logger.go:42: 07:44:43 | examples-simple-prod-with-volumes | 2024-12-09 07:44:18 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-6877c6d894 to 1 deployment-controller logger.go:42: 07:44:43 | examples-simple-prod-with-volumes | 2024-12-09 07:44:19 +0000 UTC Normal Pod simple-prod-collector-c57855f48-2nssh AddedInterface Add eth0 [10.129.2.88/23] from ovn-kubernetes multus logger.go:42: 07:44:43 | examples-simple-prod-with-volumes | 2024-12-09 07:44:19 +0000 UTC Normal Pod simple-prod-collector-c57855f48-2nssh.spec.containers{jaeger-collector} Pulled Container image "registry.redhat.io/rhosdt/jaeger-collector-rhel8@sha256:44686d560aa27cae8ff8693f88c4cb6e2edf1737010ec1f80709cb42250b729d" already present on machine kubelet logger.go:42: 07:44:43 | examples-simple-prod-with-volumes | 2024-12-09 07:44:19 +0000 UTC Normal Pod simple-prod-collector-c57855f48-2nssh.spec.containers{jaeger-collector} Created Created container jaeger-collector kubelet logger.go:42: 07:44:43 | examples-simple-prod-with-volumes | 2024-12-09 07:44:19 +0000 UTC Normal Pod simple-prod-collector-c57855f48-2nssh.spec.containers{jaeger-collector} Started Started container jaeger-collector kubelet logger.go:42: 07:44:43 | examples-simple-prod-with-volumes | 2024-12-09 07:44:19 +0000 UTC Normal Pod simple-prod-query-6877c6d894-jcjx9 AddedInterface Add eth0 [10.129.2.89/23] from ovn-kubernetes multus logger.go:42: 07:44:43 | examples-simple-prod-with-volumes | 2024-12-09 07:44:19 +0000 UTC Normal Pod simple-prod-query-6877c6d894-jcjx9.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:24a83160aa930e1b72c2a2442a33b28af2b06c24b058e09afcd0d495a8066d6d" already present on machine kubelet logger.go:42: 07:44:43 | examples-simple-prod-with-volumes | 2024-12-09 07:44:19 +0000 UTC Normal Pod simple-prod-query-6877c6d894-jcjx9.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:44:43 | examples-simple-prod-with-volumes | 2024-12-09 07:44:19 +0000 UTC Normal Pod simple-prod-query-6877c6d894-jcjx9.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:44:43 | examples-simple-prod-with-volumes | 2024-12-09 07:44:19 +0000 UTC Normal Pod simple-prod-query-6877c6d894-jcjx9.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:44:43 | examples-simple-prod-with-volumes | 2024-12-09 07:44:19 +0000 UTC Normal Pod simple-prod-query-6877c6d894-jcjx9.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:44:43 | examples-simple-prod-with-volumes | 2024-12-09 07:44:19 +0000 UTC Normal Pod simple-prod-query-6877c6d894-jcjx9.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:44:43 | examples-simple-prod-with-volumes | 2024-12-09 07:44:19 +0000 UTC Normal Pod simple-prod-query-6877c6d894-jcjx9.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" already present on machine kubelet logger.go:42: 07:44:43 | examples-simple-prod-with-volumes | 2024-12-09 07:44:19 +0000 UTC Normal Pod simple-prod-query-6877c6d894-jcjx9.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:44:43 | examples-simple-prod-with-volumes | 2024-12-09 07:44:19 +0000 UTC Normal Pod simple-prod-query-6877c6d894-jcjx9.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:44:43 | examples-simple-prod-with-volumes | 2024-12-09 07:44:24 +0000 UTC Normal Pod simple-prod-query-6877c6d894-jcjx9.spec.containers{jaeger-query} Killing Stopping container jaeger-query kubelet logger.go:42: 07:44:43 | examples-simple-prod-with-volumes | 2024-12-09 07:44:24 +0000 UTC Normal Pod simple-prod-query-6877c6d894-jcjx9.spec.containers{jaeger-agent} Killing Stopping container jaeger-agent kubelet logger.go:42: 07:44:43 | examples-simple-prod-with-volumes | 2024-12-09 07:44:24 +0000 UTC Normal Pod simple-prod-query-6877c6d894-jcjx9.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:44:43 | examples-simple-prod-with-volumes | 2024-12-09 07:44:24 +0000 UTC Normal ReplicaSet.apps simple-prod-query-6877c6d894 SuccessfulDelete Deleted pod: simple-prod-query-6877c6d894-jcjx9 replicaset-controller logger.go:42: 07:44:43 | examples-simple-prod-with-volumes | 2024-12-09 07:44:24 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled down replica set simple-prod-query-6877c6d894 to 0 from 1 deployment-controller logger.go:42: 07:44:43 | examples-simple-prod-with-volumes | 2024-12-09 07:44:25 +0000 UTC Normal Pod simple-prod-query-66bfd6758d-sh9z2 Binding Scheduled Successfully assigned kuttl-test-stunning-bobcat/simple-prod-query-66bfd6758d-sh9z2 to ip-10-0-79-227.ec2.internal default-scheduler logger.go:42: 07:44:43 | examples-simple-prod-with-volumes | 2024-12-09 07:44:25 +0000 UTC Normal ReplicaSet.apps simple-prod-query-66bfd6758d SuccessfulCreate Created pod: simple-prod-query-66bfd6758d-sh9z2 replicaset-controller logger.go:42: 07:44:43 | examples-simple-prod-with-volumes | 2024-12-09 07:44:25 +0000 UTC Normal Deployment.apps simple-prod-query ScalingReplicaSet Scaled up replica set simple-prod-query-66bfd6758d to 1 deployment-controller logger.go:42: 07:44:43 | examples-simple-prod-with-volumes | 2024-12-09 07:44:26 +0000 UTC Normal Pod simple-prod-query-66bfd6758d-sh9z2 AddedInterface Add eth0 [10.129.2.90/23] from ovn-kubernetes multus logger.go:42: 07:44:43 | examples-simple-prod-with-volumes | 2024-12-09 07:44:26 +0000 UTC Normal Pod simple-prod-query-66bfd6758d-sh9z2.spec.containers{jaeger-query} Pulled Container image "registry.redhat.io/rhosdt/jaeger-query-rhel8@sha256:24a83160aa930e1b72c2a2442a33b28af2b06c24b058e09afcd0d495a8066d6d" already present on machine kubelet logger.go:42: 07:44:43 | examples-simple-prod-with-volumes | 2024-12-09 07:44:26 +0000 UTC Normal Pod simple-prod-query-66bfd6758d-sh9z2.spec.containers{jaeger-query} Created Created container jaeger-query kubelet logger.go:42: 07:44:43 | examples-simple-prod-with-volumes | 2024-12-09 07:44:26 +0000 UTC Normal Pod simple-prod-query-66bfd6758d-sh9z2.spec.containers{jaeger-query} Started Started container jaeger-query kubelet logger.go:42: 07:44:43 | examples-simple-prod-with-volumes | 2024-12-09 07:44:26 +0000 UTC Normal Pod simple-prod-query-66bfd6758d-sh9z2.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:44:43 | examples-simple-prod-with-volumes | 2024-12-09 07:44:26 +0000 UTC Normal Pod simple-prod-query-66bfd6758d-sh9z2.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:44:43 | examples-simple-prod-with-volumes | 2024-12-09 07:44:26 +0000 UTC Normal Pod simple-prod-query-66bfd6758d-sh9z2.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:44:43 | examples-simple-prod-with-volumes | 2024-12-09 07:44:26 +0000 UTC Normal Pod simple-prod-query-66bfd6758d-sh9z2.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" already present on machine kubelet logger.go:42: 07:44:43 | examples-simple-prod-with-volumes | 2024-12-09 07:44:26 +0000 UTC Normal Pod simple-prod-query-66bfd6758d-sh9z2.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:44:43 | examples-simple-prod-with-volumes | 2024-12-09 07:44:26 +0000 UTC Normal Pod simple-prod-query-66bfd6758d-sh9z2.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:44:43 | examples-simple-prod-with-volumes | 2024-12-09 07:44:30 +0000 UTC Normal Pod check-span-ks9kp Binding Scheduled Successfully assigned kuttl-test-stunning-bobcat/check-span-ks9kp to ip-10-0-39-223.ec2.internal default-scheduler logger.go:42: 07:44:43 | examples-simple-prod-with-volumes | 2024-12-09 07:44:30 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-ks9kp job-controller logger.go:42: 07:44:43 | examples-simple-prod-with-volumes | 2024-12-09 07:44:30 +0000 UTC Normal Pod report-span-fljrz Binding Scheduled Successfully assigned kuttl-test-stunning-bobcat/report-span-fljrz to ip-10-0-39-223.ec2.internal default-scheduler logger.go:42: 07:44:43 | examples-simple-prod-with-volumes | 2024-12-09 07:44:30 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-fljrz job-controller logger.go:42: 07:44:43 | examples-simple-prod-with-volumes | 2024-12-09 07:44:31 +0000 UTC Normal Pod check-span-ks9kp AddedInterface Add eth0 [10.128.2.71/23] from ovn-kubernetes multus logger.go:42: 07:44:43 | examples-simple-prod-with-volumes | 2024-12-09 07:44:31 +0000 UTC Normal Pod check-span-ks9kp.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:44:43 | examples-simple-prod-with-volumes | 2024-12-09 07:44:31 +0000 UTC Normal Pod check-span-ks9kp.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 820ms (820ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:44:43 | examples-simple-prod-with-volumes | 2024-12-09 07:44:31 +0000 UTC Normal Pod report-span-fljrz AddedInterface Add eth0 [10.128.2.70/23] from ovn-kubernetes multus logger.go:42: 07:44:43 | examples-simple-prod-with-volumes | 2024-12-09 07:44:31 +0000 UTC Normal Pod report-span-fljrz.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:44:43 | examples-simple-prod-with-volumes | 2024-12-09 07:44:31 +0000 UTC Normal Pod report-span-fljrz.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 813ms (813ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:44:43 | examples-simple-prod-with-volumes | 2024-12-09 07:44:31 +0000 UTC Normal Pod report-span-fljrz.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:44:43 | examples-simple-prod-with-volumes | 2024-12-09 07:44:31 +0000 UTC Normal Pod report-span-fljrz.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:44:43 | examples-simple-prod-with-volumes | 2024-12-09 07:44:32 +0000 UTC Normal Pod check-span-ks9kp.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:44:43 | examples-simple-prod-with-volumes | 2024-12-09 07:44:32 +0000 UTC Normal Pod check-span-ks9kp.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:44:43 | examples-simple-prod-with-volumes | 2024-12-09 07:44:33 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedGetResourceMetric failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:44:43 | examples-simple-prod-with-volumes | 2024-12-09 07:44:33 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedGetResourceMetric failed to get memory utilization: unable to get metrics for resource memory: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:44:43 | examples-simple-prod-with-volumes | 2024-12-09 07:44:33 +0000 UTC Warning HorizontalPodAutoscaler.autoscaling simple-prod-collector FailedComputeMetricsReplicas invalid metrics (2 invalid out of 2), first error is: failed to get cpu resource metric value: failed to get cpu utilization: unable to get metrics for resource cpu: no metrics returned from resource metrics API horizontal-pod-autoscaler logger.go:42: 07:44:43 | examples-simple-prod-with-volumes | 2024-12-09 07:44:42 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:44:43 | examples-simple-prod-with-volumes | Deleting namespace: kuttl-test-stunning-bobcat === CONT kuttl/harness/examples-business-application-injected-sidecar logger.go:42: 07:44:49 | examples-business-application-injected-sidecar | Creating namespace: kuttl-test-expert-pipefish logger.go:42: 07:44:49 | examples-business-application-injected-sidecar/0-install | starting test step 0-install logger.go:42: 07:44:50 | examples-business-application-injected-sidecar/0-install | Deployment:kuttl-test-expert-pipefish/myapp created logger.go:42: 07:44:50 | examples-business-application-injected-sidecar/0-install | test step completed 0-install logger.go:42: 07:44:50 | examples-business-application-injected-sidecar/1-install | starting test step 1-install logger.go:42: 07:44:50 | examples-business-application-injected-sidecar/1-install | Jaeger:kuttl-test-expert-pipefish/simplest created logger.go:42: 07:45:04 | examples-business-application-injected-sidecar/1-install | test step completed 1-install logger.go:42: 07:45:04 | examples-business-application-injected-sidecar/2-smoke-test | starting test step 2-smoke-test logger.go:42: 07:45:04 | examples-business-application-injected-sidecar/2-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE simplest /dev/null] logger.go:42: 07:45:05 | examples-business-application-injected-sidecar/2-smoke-test | Warning: resource jaegers/simplest is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:45:13 | examples-business-application-injected-sidecar/2-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://simplest-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://simplest-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:45:13 | examples-business-application-injected-sidecar/2-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:45:13 | examples-business-application-injected-sidecar/2-smoke-test | job.batch/report-span created logger.go:42: 07:45:13 | examples-business-application-injected-sidecar/2-smoke-test | job.batch/check-span created logger.go:42: 07:45:27 | examples-business-application-injected-sidecar/2-smoke-test | test step completed 2-smoke-test logger.go:42: 07:45:27 | examples-business-application-injected-sidecar | examples-business-application-injected-sidecar events from ns kuttl-test-expert-pipefish: logger.go:42: 07:45:27 | examples-business-application-injected-sidecar | 2024-12-09 07:44:50 +0000 UTC Normal Pod myapp-fc54f78d8-r7qg9 Binding Scheduled Successfully assigned kuttl-test-expert-pipefish/myapp-fc54f78d8-r7qg9 to ip-10-0-79-227.ec2.internal default-scheduler logger.go:42: 07:45:27 | examples-business-application-injected-sidecar | 2024-12-09 07:44:50 +0000 UTC Normal Pod myapp-fc54f78d8-r7qg9 AddedInterface Add eth0 [10.129.2.91/23] from ovn-kubernetes multus logger.go:42: 07:45:27 | examples-business-application-injected-sidecar | 2024-12-09 07:44:50 +0000 UTC Normal Pod myapp-fc54f78d8-r7qg9.spec.containers{myapp} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:45:27 | examples-business-application-injected-sidecar | 2024-12-09 07:44:50 +0000 UTC Normal Pod myapp-fc54f78d8-r7qg9.spec.containers{myapp} Created Created container myapp kubelet logger.go:42: 07:45:27 | examples-business-application-injected-sidecar | 2024-12-09 07:44:50 +0000 UTC Normal Pod myapp-fc54f78d8-r7qg9.spec.containers{myapp} Started Started container myapp kubelet logger.go:42: 07:45:27 | examples-business-application-injected-sidecar | 2024-12-09 07:44:50 +0000 UTC Normal ReplicaSet.apps myapp-fc54f78d8 SuccessfulCreate Created pod: myapp-fc54f78d8-r7qg9 replicaset-controller logger.go:42: 07:45:27 | examples-business-application-injected-sidecar | 2024-12-09 07:44:50 +0000 UTC Normal Deployment.apps myapp ScalingReplicaSet Scaled up replica set myapp-fc54f78d8 to 1 deployment-controller logger.go:42: 07:45:27 | examples-business-application-injected-sidecar | 2024-12-09 07:44:56 +0000 UTC Normal Pod myapp-67589bd466-slt5d Binding Scheduled Successfully assigned kuttl-test-expert-pipefish/myapp-67589bd466-slt5d to ip-10-0-95-250.ec2.internal default-scheduler logger.go:42: 07:45:27 | examples-business-application-injected-sidecar | 2024-12-09 07:44:56 +0000 UTC Warning Pod myapp-67589bd466-slt5d FailedMount MountVolume.SetUp failed for volume "simplest-trusted-ca" : configmap "simplest-trusted-ca" not found kubelet logger.go:42: 07:45:27 | examples-business-application-injected-sidecar | 2024-12-09 07:44:56 +0000 UTC Warning Pod myapp-67589bd466-slt5d FailedMount MountVolume.SetUp failed for volume "simplest-service-ca" : configmap "simplest-service-ca" not found kubelet logger.go:42: 07:45:27 | examples-business-application-injected-sidecar | 2024-12-09 07:44:56 +0000 UTC Normal ReplicaSet.apps myapp-67589bd466 SuccessfulCreate Created pod: myapp-67589bd466-slt5d replicaset-controller logger.go:42: 07:45:27 | examples-business-application-injected-sidecar | 2024-12-09 07:44:56 +0000 UTC Normal Deployment.apps myapp ScalingReplicaSet Scaled up replica set myapp-67589bd466 to 1 deployment-controller logger.go:42: 07:45:27 | examples-business-application-injected-sidecar | 2024-12-09 07:45:01 +0000 UTC Warning Pod myapp-fc54f78d8-r7qg9.spec.containers{myapp} Unhealthy Liveness probe failed: Get "http://10.129.2.91:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:45:27 | examples-business-application-injected-sidecar | 2024-12-09 07:45:01 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled up replica set simplest-75564dfb66 to 1 deployment-controller logger.go:42: 07:45:27 | examples-business-application-injected-sidecar | 2024-12-09 07:45:02 +0000 UTC Normal Pod simplest-75564dfb66-6r6vd Binding Scheduled Successfully assigned kuttl-test-expert-pipefish/simplest-75564dfb66-6r6vd to ip-10-0-79-227.ec2.internal default-scheduler logger.go:42: 07:45:27 | examples-business-application-injected-sidecar | 2024-12-09 07:45:02 +0000 UTC Normal Pod simplest-75564dfb66-6r6vd AddedInterface Add eth0 [10.129.2.93/23] from ovn-kubernetes multus logger.go:42: 07:45:27 | examples-business-application-injected-sidecar | 2024-12-09 07:45:02 +0000 UTC Normal Pod simplest-75564dfb66-6r6vd.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1fa5ed13cc8a023f1e987f6cafe86adc195c373cc2b774539df6d0fd02b780a7" already present on machine kubelet logger.go:42: 07:45:27 | examples-business-application-injected-sidecar | 2024-12-09 07:45:02 +0000 UTC Normal Pod simplest-75564dfb66-6r6vd.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:45:27 | examples-business-application-injected-sidecar | 2024-12-09 07:45:02 +0000 UTC Normal Pod simplest-75564dfb66-6r6vd.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:45:27 | examples-business-application-injected-sidecar | 2024-12-09 07:45:02 +0000 UTC Normal Pod simplest-75564dfb66-6r6vd.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:45:27 | examples-business-application-injected-sidecar | 2024-12-09 07:45:02 +0000 UTC Normal Pod simplest-75564dfb66-6r6vd.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:45:27 | examples-business-application-injected-sidecar | 2024-12-09 07:45:02 +0000 UTC Normal Pod simplest-75564dfb66-6r6vd.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:45:27 | examples-business-application-injected-sidecar | 2024-12-09 07:45:02 +0000 UTC Normal ReplicaSet.apps simplest-75564dfb66 SuccessfulCreate Created pod: simplest-75564dfb66-6r6vd replicaset-controller logger.go:42: 07:45:27 | examples-business-application-injected-sidecar | 2024-12-09 07:45:04 +0000 UTC Normal Pod myapp-67589bd466-slt5d AddedInterface Add eth0 [10.131.0.48/23] from ovn-kubernetes multus logger.go:42: 07:45:27 | examples-business-application-injected-sidecar | 2024-12-09 07:45:04 +0000 UTC Normal Pod myapp-67589bd466-slt5d.spec.containers{myapp} Pulled Container image "jaegertracing/vertx-create-span:operator-e2e-tests" already present on machine kubelet logger.go:42: 07:45:27 | examples-business-application-injected-sidecar | 2024-12-09 07:45:04 +0000 UTC Normal Pod myapp-67589bd466-slt5d.spec.containers{myapp} Created Created container myapp kubelet logger.go:42: 07:45:27 | examples-business-application-injected-sidecar | 2024-12-09 07:45:04 +0000 UTC Normal Pod myapp-67589bd466-slt5d.spec.containers{myapp} Started Started container myapp kubelet logger.go:42: 07:45:27 | examples-business-application-injected-sidecar | 2024-12-09 07:45:04 +0000 UTC Normal Pod myapp-67589bd466-slt5d.spec.containers{jaeger-agent} Pulled Container image "registry.redhat.io/rhosdt/jaeger-agent-rhel8@sha256:986580d136bd26bf3d626578d2b630dd66cd83206fe5bba124f052a1a8d1d0ec" already present on machine kubelet logger.go:42: 07:45:27 | examples-business-application-injected-sidecar | 2024-12-09 07:45:04 +0000 UTC Normal Pod myapp-67589bd466-slt5d.spec.containers{jaeger-agent} Created Created container jaeger-agent kubelet logger.go:42: 07:45:27 | examples-business-application-injected-sidecar | 2024-12-09 07:45:04 +0000 UTC Normal Pod myapp-67589bd466-slt5d.spec.containers{jaeger-agent} Started Started container jaeger-agent kubelet logger.go:42: 07:45:27 | examples-business-application-injected-sidecar | 2024-12-09 07:45:05 +0000 UTC Normal Pod myapp-fc54f78d8-r7qg9.spec.containers{myapp} Killing Stopping container myapp kubelet logger.go:42: 07:45:27 | examples-business-application-injected-sidecar | 2024-12-09 07:45:05 +0000 UTC Normal ReplicaSet.apps myapp-fc54f78d8 SuccessfulDelete Deleted pod: myapp-fc54f78d8-r7qg9 replicaset-controller logger.go:42: 07:45:27 | examples-business-application-injected-sidecar | 2024-12-09 07:45:05 +0000 UTC Normal Deployment.apps myapp ScalingReplicaSet Scaled down replica set myapp-fc54f78d8 to 0 from 1 deployment-controller logger.go:42: 07:45:27 | examples-business-application-injected-sidecar | 2024-12-09 07:45:09 +0000 UTC Normal Pod simplest-75564dfb66-6r6vd.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:45:27 | examples-business-application-injected-sidecar | 2024-12-09 07:45:09 +0000 UTC Normal Pod simplest-75564dfb66-6r6vd.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:45:27 | examples-business-application-injected-sidecar | 2024-12-09 07:45:09 +0000 UTC Normal ReplicaSet.apps simplest-75564dfb66 SuccessfulDelete Deleted pod: simplest-75564dfb66-6r6vd replicaset-controller logger.go:42: 07:45:27 | examples-business-application-injected-sidecar | 2024-12-09 07:45:09 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled down replica set simplest-75564dfb66 to 0 from 1 deployment-controller logger.go:42: 07:45:27 | examples-business-application-injected-sidecar | 2024-12-09 07:45:11 +0000 UTC Normal Pod simplest-66b577d74-fmvmc Binding Scheduled Successfully assigned kuttl-test-expert-pipefish/simplest-66b577d74-fmvmc to ip-10-0-79-227.ec2.internal default-scheduler logger.go:42: 07:45:27 | examples-business-application-injected-sidecar | 2024-12-09 07:45:11 +0000 UTC Normal Pod simplest-66b577d74-fmvmc AddedInterface Add eth0 [10.129.2.94/23] from ovn-kubernetes multus logger.go:42: 07:45:27 | examples-business-application-injected-sidecar | 2024-12-09 07:45:11 +0000 UTC Normal Pod simplest-66b577d74-fmvmc.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1fa5ed13cc8a023f1e987f6cafe86adc195c373cc2b774539df6d0fd02b780a7" already present on machine kubelet logger.go:42: 07:45:27 | examples-business-application-injected-sidecar | 2024-12-09 07:45:11 +0000 UTC Normal Pod simplest-66b577d74-fmvmc.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:45:27 | examples-business-application-injected-sidecar | 2024-12-09 07:45:11 +0000 UTC Normal Pod simplest-66b577d74-fmvmc.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:45:27 | examples-business-application-injected-sidecar | 2024-12-09 07:45:11 +0000 UTC Normal Pod simplest-66b577d74-fmvmc.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:45:27 | examples-business-application-injected-sidecar | 2024-12-09 07:45:11 +0000 UTC Normal Pod simplest-66b577d74-fmvmc.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:45:27 | examples-business-application-injected-sidecar | 2024-12-09 07:45:11 +0000 UTC Normal Pod simplest-66b577d74-fmvmc.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:45:27 | examples-business-application-injected-sidecar | 2024-12-09 07:45:11 +0000 UTC Normal ReplicaSet.apps simplest-66b577d74 SuccessfulCreate Created pod: simplest-66b577d74-fmvmc replicaset-controller logger.go:42: 07:45:27 | examples-business-application-injected-sidecar | 2024-12-09 07:45:11 +0000 UTC Normal Deployment.apps simplest ScalingReplicaSet Scaled up replica set simplest-66b577d74 to 1 deployment-controller logger.go:42: 07:45:27 | examples-business-application-injected-sidecar | 2024-12-09 07:45:13 +0000 UTC Normal Pod check-span-lfcdj Binding Scheduled Successfully assigned kuttl-test-expert-pipefish/check-span-lfcdj to ip-10-0-39-223.ec2.internal default-scheduler logger.go:42: 07:45:27 | examples-business-application-injected-sidecar | 2024-12-09 07:45:13 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-lfcdj job-controller logger.go:42: 07:45:27 | examples-business-application-injected-sidecar | 2024-12-09 07:45:13 +0000 UTC Normal Pod report-span-nsscn Binding Scheduled Successfully assigned kuttl-test-expert-pipefish/report-span-nsscn to ip-10-0-39-223.ec2.internal default-scheduler logger.go:42: 07:45:27 | examples-business-application-injected-sidecar | 2024-12-09 07:45:13 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-nsscn job-controller logger.go:42: 07:45:27 | examples-business-application-injected-sidecar | 2024-12-09 07:45:14 +0000 UTC Normal Pod check-span-lfcdj AddedInterface Add eth0 [10.128.2.73/23] from ovn-kubernetes multus logger.go:42: 07:45:27 | examples-business-application-injected-sidecar | 2024-12-09 07:45:14 +0000 UTC Normal Pod check-span-lfcdj.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:45:27 | examples-business-application-injected-sidecar | 2024-12-09 07:45:14 +0000 UTC Normal Pod report-span-nsscn AddedInterface Add eth0 [10.128.2.72/23] from ovn-kubernetes multus logger.go:42: 07:45:27 | examples-business-application-injected-sidecar | 2024-12-09 07:45:14 +0000 UTC Normal Pod report-span-nsscn.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:45:27 | examples-business-application-injected-sidecar | 2024-12-09 07:45:15 +0000 UTC Normal Pod check-span-lfcdj.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 1.061s (1.061s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:45:27 | examples-business-application-injected-sidecar | 2024-12-09 07:45:15 +0000 UTC Normal Pod check-span-lfcdj.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:45:27 | examples-business-application-injected-sidecar | 2024-12-09 07:45:15 +0000 UTC Normal Pod check-span-lfcdj.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:45:27 | examples-business-application-injected-sidecar | 2024-12-09 07:45:15 +0000 UTC Warning Pod myapp-67589bd466-slt5d.spec.containers{myapp} Unhealthy Liveness probe failed: Get "http://10.131.0.48:8080/": context deadline exceeded (Client.Timeout exceeded while awaiting headers) kubelet logger.go:42: 07:45:27 | examples-business-application-injected-sidecar | 2024-12-09 07:45:16 +0000 UTC Normal Pod report-span-nsscn.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 2.397s (2.397s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:45:27 | examples-business-application-injected-sidecar | 2024-12-09 07:45:16 +0000 UTC Normal Pod report-span-nsscn.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:45:27 | examples-business-application-injected-sidecar | 2024-12-09 07:45:16 +0000 UTC Normal Pod report-span-nsscn.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:45:27 | examples-business-application-injected-sidecar | 2024-12-09 07:45:26 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:45:27 | examples-business-application-injected-sidecar | Deleting namespace: kuttl-test-expert-pipefish === CONT kuttl/harness/examples-service-types logger.go:42: 07:45:39 | examples-service-types | Creating namespace: kuttl-test-alive-marlin logger.go:42: 07:45:39 | examples-service-types/0-install | starting test step 0-install logger.go:42: 07:45:39 | examples-service-types/0-install | Jaeger:kuttl-test-alive-marlin/service-types created logger.go:42: 07:45:45 | examples-service-types/0-install | test step completed 0-install logger.go:42: 07:45:45 | examples-service-types/1-smoke-test | starting test step 1-smoke-test logger.go:42: 07:45:45 | examples-service-types/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE service-types /dev/null] logger.go:42: 07:45:47 | examples-service-types/1-smoke-test | Warning: resource jaegers/service-types is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:45:53 | examples-service-types/1-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://service-types-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://service-types-query:443 MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:45:53 | examples-service-types/1-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:45:54 | examples-service-types/1-smoke-test | job.batch/report-span created logger.go:42: 07:45:54 | examples-service-types/1-smoke-test | job.batch/check-span created logger.go:42: 07:46:06 | examples-service-types/1-smoke-test | test step completed 1-smoke-test logger.go:42: 07:46:06 | examples-service-types/2- | starting test step 2- logger.go:42: 07:46:06 | examples-service-types/2- | test step completed 2- logger.go:42: 07:46:06 | examples-service-types | examples-service-types events from ns kuttl-test-alive-marlin: logger.go:42: 07:46:06 | examples-service-types | 2024-12-09 07:45:43 +0000 UTC Normal Pod service-types-85967d9cb9-gtkcz Binding Scheduled Successfully assigned kuttl-test-alive-marlin/service-types-85967d9cb9-gtkcz to ip-10-0-79-227.ec2.internal default-scheduler logger.go:42: 07:46:06 | examples-service-types | 2024-12-09 07:45:43 +0000 UTC Normal ReplicaSet.apps service-types-85967d9cb9 SuccessfulCreate Created pod: service-types-85967d9cb9-gtkcz replicaset-controller logger.go:42: 07:46:06 | examples-service-types | 2024-12-09 07:45:43 +0000 UTC Normal Service service-types-collector EnsuringLoadBalancer Ensuring load balancer service-controller logger.go:42: 07:46:06 | examples-service-types | 2024-12-09 07:45:43 +0000 UTC Normal Deployment.apps service-types ScalingReplicaSet Scaled up replica set service-types-85967d9cb9 to 1 deployment-controller logger.go:42: 07:46:06 | examples-service-types | 2024-12-09 07:45:44 +0000 UTC Normal Pod service-types-85967d9cb9-gtkcz AddedInterface Add eth0 [10.129.2.95/23] from ovn-kubernetes multus logger.go:42: 07:46:06 | examples-service-types | 2024-12-09 07:45:44 +0000 UTC Normal Pod service-types-85967d9cb9-gtkcz.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1fa5ed13cc8a023f1e987f6cafe86adc195c373cc2b774539df6d0fd02b780a7" already present on machine kubelet logger.go:42: 07:46:06 | examples-service-types | 2024-12-09 07:45:44 +0000 UTC Normal Pod service-types-85967d9cb9-gtkcz.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:46:06 | examples-service-types | 2024-12-09 07:45:44 +0000 UTC Normal Pod service-types-85967d9cb9-gtkcz.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:46:06 | examples-service-types | 2024-12-09 07:45:44 +0000 UTC Normal Pod service-types-85967d9cb9-gtkcz.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:46:06 | examples-service-types | 2024-12-09 07:45:44 +0000 UTC Normal Pod service-types-85967d9cb9-gtkcz.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:46:06 | examples-service-types | 2024-12-09 07:45:44 +0000 UTC Normal Pod service-types-85967d9cb9-gtkcz.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:46:06 | examples-service-types | 2024-12-09 07:45:46 +0000 UTC Normal Service service-types-collector EnsuredLoadBalancer Ensured load balancer service-controller logger.go:42: 07:46:06 | examples-service-types | 2024-12-09 07:45:46 +0000 UTC Normal Service service-types-query EnsuringLoadBalancer Ensuring load balancer service-controller logger.go:42: 07:46:06 | examples-service-types | 2024-12-09 07:45:49 +0000 UTC Normal Pod service-types-85967d9cb9-gtkcz.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:46:06 | examples-service-types | 2024-12-09 07:45:49 +0000 UTC Normal Pod service-types-85967d9cb9-gtkcz.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:46:06 | examples-service-types | 2024-12-09 07:45:49 +0000 UTC Normal ReplicaSet.apps service-types-85967d9cb9 SuccessfulDelete Deleted pod: service-types-85967d9cb9-gtkcz replicaset-controller logger.go:42: 07:46:06 | examples-service-types | 2024-12-09 07:45:49 +0000 UTC Normal Deployment.apps service-types ScalingReplicaSet Scaled down replica set service-types-85967d9cb9 to 0 from 1 deployment-controller logger.go:42: 07:46:06 | examples-service-types | 2024-12-09 07:45:50 +0000 UTC Normal Pod service-types-d5549574c-rzjb5 Binding Scheduled Successfully assigned kuttl-test-alive-marlin/service-types-d5549574c-rzjb5 to ip-10-0-79-227.ec2.internal default-scheduler logger.go:42: 07:46:06 | examples-service-types | 2024-12-09 07:45:50 +0000 UTC Normal Pod service-types-d5549574c-rzjb5 AddedInterface Add eth0 [10.129.2.96/23] from ovn-kubernetes multus logger.go:42: 07:46:06 | examples-service-types | 2024-12-09 07:45:50 +0000 UTC Normal Pod service-types-d5549574c-rzjb5.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1fa5ed13cc8a023f1e987f6cafe86adc195c373cc2b774539df6d0fd02b780a7" already present on machine kubelet logger.go:42: 07:46:06 | examples-service-types | 2024-12-09 07:45:50 +0000 UTC Normal Pod service-types-d5549574c-rzjb5.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:46:06 | examples-service-types | 2024-12-09 07:45:50 +0000 UTC Normal Pod service-types-d5549574c-rzjb5.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:46:06 | examples-service-types | 2024-12-09 07:45:50 +0000 UTC Normal Pod service-types-d5549574c-rzjb5.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:46:06 | examples-service-types | 2024-12-09 07:45:50 +0000 UTC Normal Pod service-types-d5549574c-rzjb5.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:46:06 | examples-service-types | 2024-12-09 07:45:50 +0000 UTC Normal Pod service-types-d5549574c-rzjb5.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:46:06 | examples-service-types | 2024-12-09 07:45:50 +0000 UTC Normal ReplicaSet.apps service-types-d5549574c SuccessfulCreate Created pod: service-types-d5549574c-rzjb5 replicaset-controller logger.go:42: 07:46:06 | examples-service-types | 2024-12-09 07:45:50 +0000 UTC Normal Service service-types-query EnsuredLoadBalancer Ensured load balancer service-controller logger.go:42: 07:46:06 | examples-service-types | 2024-12-09 07:45:50 +0000 UTC Normal Deployment.apps service-types ScalingReplicaSet Scaled up replica set service-types-d5549574c to 1 deployment-controller logger.go:42: 07:46:06 | examples-service-types | 2024-12-09 07:45:54 +0000 UTC Normal Pod check-span-fzzdr Binding Scheduled Successfully assigned kuttl-test-alive-marlin/check-span-fzzdr to ip-10-0-39-223.ec2.internal default-scheduler logger.go:42: 07:46:06 | examples-service-types | 2024-12-09 07:45:54 +0000 UTC Normal Pod check-span-fzzdr AddedInterface Add eth0 [10.128.2.74/23] from ovn-kubernetes multus logger.go:42: 07:46:06 | examples-service-types | 2024-12-09 07:45:54 +0000 UTC Normal Pod check-span-fzzdr.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:46:06 | examples-service-types | 2024-12-09 07:45:54 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-fzzdr job-controller logger.go:42: 07:46:06 | examples-service-types | 2024-12-09 07:45:54 +0000 UTC Normal Pod report-span-dhsll Binding Scheduled Successfully assigned kuttl-test-alive-marlin/report-span-dhsll to ip-10-0-95-250.ec2.internal default-scheduler logger.go:42: 07:46:06 | examples-service-types | 2024-12-09 07:45:54 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-dhsll job-controller logger.go:42: 07:46:06 | examples-service-types | 2024-12-09 07:45:55 +0000 UTC Normal Pod check-span-fzzdr.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 765ms (765ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:46:06 | examples-service-types | 2024-12-09 07:45:55 +0000 UTC Normal Pod check-span-fzzdr.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:46:06 | examples-service-types | 2024-12-09 07:45:55 +0000 UTC Normal Pod check-span-fzzdr.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:46:06 | examples-service-types | 2024-12-09 07:45:55 +0000 UTC Normal Pod report-span-dhsll AddedInterface Add eth0 [10.131.0.49/23] from ovn-kubernetes multus logger.go:42: 07:46:06 | examples-service-types | 2024-12-09 07:45:55 +0000 UTC Normal Pod report-span-dhsll.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:46:06 | examples-service-types | 2024-12-09 07:45:56 +0000 UTC Normal Pod report-span-dhsll.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 1.182s (1.182s including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:46:06 | examples-service-types | 2024-12-09 07:45:56 +0000 UTC Normal Pod report-span-dhsll.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:46:06 | examples-service-types | 2024-12-09 07:45:56 +0000 UTC Normal Pod report-span-dhsll.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:46:06 | examples-service-types | 2024-12-09 07:46:06 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:46:06 | examples-service-types | Deleting namespace: kuttl-test-alive-marlin === CONT kuttl/harness/examples-openshift-with-htpasswd logger.go:42: 07:46:19 | examples-openshift-with-htpasswd | Ignoring 00-install.yaml.template as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:46:19 | examples-openshift-with-htpasswd | Ignoring ensure-ingress-host.sh as it does not match file name regexp: ^(\d+)-(?:[^\.]+)(?:\.yaml)?$ logger.go:42: 07:46:19 | examples-openshift-with-htpasswd | Creating namespace: kuttl-test-enjoyed-cow logger.go:42: 07:46:19 | examples-openshift-with-htpasswd/0-install | starting test step 0-install logger.go:42: 07:46:19 | examples-openshift-with-htpasswd/0-install | Secret:kuttl-test-enjoyed-cow/htpasswd created logger.go:42: 07:46:19 | examples-openshift-with-htpasswd/0-install | test step completed 0-install logger.go:42: 07:46:19 | examples-openshift-with-htpasswd/1-install | starting test step 1-install logger.go:42: 07:46:19 | examples-openshift-with-htpasswd/1-install | Jaeger:kuttl-test-enjoyed-cow/with-htpasswd created logger.go:42: 07:46:25 | examples-openshift-with-htpasswd/1-install | test step completed 1-install logger.go:42: 07:46:25 | examples-openshift-with-htpasswd/2-check-unsecured | starting test step 2-check-unsecured logger.go:42: 07:46:25 | examples-openshift-with-htpasswd/2-check-unsecured | running command: [./ensure-ingress-host.sh] logger.go:42: 07:46:25 | examples-openshift-with-htpasswd/2-check-unsecured | Checking the Ingress host value was populated logger.go:42: 07:46:25 | examples-openshift-with-htpasswd/2-check-unsecured | Try number 0 logger.go:42: 07:46:25 | examples-openshift-with-htpasswd/2-check-unsecured | Hostname is with-htpasswd-kuttl-test-enjoyed-cow.apps.ci-op-54q9kfc5-9210b.cspilp.interop.ccitredhat.com logger.go:42: 07:46:25 | examples-openshift-with-htpasswd/2-check-unsecured | running command: [sh -c INSECURE=true ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 403 true $NAMESPACE with-htpasswd] logger.go:42: 07:46:25 | examples-openshift-with-htpasswd/2-check-unsecured | Checking an expected HTTP response logger.go:42: 07:46:25 | examples-openshift-with-htpasswd/2-check-unsecured | Running in OpenShift logger.go:42: 07:46:25 | examples-openshift-with-htpasswd/2-check-unsecured | Not using any secret logger.go:42: 07:46:25 | examples-openshift-with-htpasswd/2-check-unsecured | Try number 1/30 the https://with-htpasswd-kuttl-test-enjoyed-cow.apps.ci-op-54q9kfc5-9210b.cspilp.interop.ccitredhat.com/search logger.go:42: 07:46:25 | examples-openshift-with-htpasswd/2-check-unsecured | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 07:46:25 | examples-openshift-with-htpasswd/2-check-unsecured | Try number 2/30 the https://with-htpasswd-kuttl-test-enjoyed-cow.apps.ci-op-54q9kfc5-9210b.cspilp.interop.ccitredhat.com/search logger.go:42: 07:46:25 | examples-openshift-with-htpasswd/2-check-unsecured | HTTP response is 503. 403 expected. Waiting 10 s logger.go:42: 07:46:35 | examples-openshift-with-htpasswd/2-check-unsecured | Try number 3/30 the https://with-htpasswd-kuttl-test-enjoyed-cow.apps.ci-op-54q9kfc5-9210b.cspilp.interop.ccitredhat.com/search logger.go:42: 07:46:35 | examples-openshift-with-htpasswd/2-check-unsecured | curl response asserted properly logger.go:42: 07:46:35 | examples-openshift-with-htpasswd/2-check-unsecured | test step completed 2-check-unsecured logger.go:42: 07:46:35 | examples-openshift-with-htpasswd/3-check-unauthorized | starting test step 3-check-unauthorized logger.go:42: 07:46:35 | examples-openshift-with-htpasswd/3-check-unauthorized | running command: [./ensure-ingress-host.sh] logger.go:42: 07:46:35 | examples-openshift-with-htpasswd/3-check-unauthorized | Checking the Ingress host value was populated logger.go:42: 07:46:35 | examples-openshift-with-htpasswd/3-check-unauthorized | Try number 0 logger.go:42: 07:46:36 | examples-openshift-with-htpasswd/3-check-unauthorized | Hostname is with-htpasswd-kuttl-test-enjoyed-cow.apps.ci-op-54q9kfc5-9210b.cspilp.interop.ccitredhat.com logger.go:42: 07:46:36 | examples-openshift-with-htpasswd/3-check-unauthorized | running command: [sh -c JAEGER_USERNAME=wronguser JAEGER_PASSWORD=wrongpassword ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 403 true $NAMESPACE with-htpasswd] logger.go:42: 07:46:36 | examples-openshift-with-htpasswd/3-check-unauthorized | Checking an expected HTTP response logger.go:42: 07:46:36 | examples-openshift-with-htpasswd/3-check-unauthorized | Running in OpenShift logger.go:42: 07:46:36 | examples-openshift-with-htpasswd/3-check-unauthorized | Using Jaeger basic authentication logger.go:42: 07:46:36 | examples-openshift-with-htpasswd/3-check-unauthorized | Try number 1/30 the https://with-htpasswd-kuttl-test-enjoyed-cow.apps.ci-op-54q9kfc5-9210b.cspilp.interop.ccitredhat.com/search logger.go:42: 07:46:36 | examples-openshift-with-htpasswd/3-check-unauthorized | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 07:46:36 | examples-openshift-with-htpasswd/3-check-unauthorized | Try number 2/30 the https://with-htpasswd-kuttl-test-enjoyed-cow.apps.ci-op-54q9kfc5-9210b.cspilp.interop.ccitredhat.com/search logger.go:42: 07:46:36 | examples-openshift-with-htpasswd/3-check-unauthorized | curl response asserted properly logger.go:42: 07:46:36 | examples-openshift-with-htpasswd/3-check-unauthorized | test step completed 3-check-unauthorized logger.go:42: 07:46:36 | examples-openshift-with-htpasswd/4-check-authorized | starting test step 4-check-authorized logger.go:42: 07:46:36 | examples-openshift-with-htpasswd/4-check-authorized | running command: [./ensure-ingress-host.sh] logger.go:42: 07:46:36 | examples-openshift-with-htpasswd/4-check-authorized | Checking the Ingress host value was populated logger.go:42: 07:46:36 | examples-openshift-with-htpasswd/4-check-authorized | Try number 0 logger.go:42: 07:46:36 | examples-openshift-with-htpasswd/4-check-authorized | Hostname is with-htpasswd-kuttl-test-enjoyed-cow.apps.ci-op-54q9kfc5-9210b.cspilp.interop.ccitredhat.com logger.go:42: 07:46:36 | examples-openshift-with-htpasswd/4-check-authorized | running command: [sh -c JAEGER_USERNAME=awesomeuser JAEGER_PASSWORD=awesomepassword ../../../../cmd-utils/assert-jaeger-http-code.sh https://$(kubectl get routes -o=jsonpath='{.items[0].status.ingress[0].host}' -n $NAMESPACE)/search 200 true $NAMESPACE with-htpasswd] logger.go:42: 07:46:36 | examples-openshift-with-htpasswd/4-check-authorized | Checking an expected HTTP response logger.go:42: 07:46:36 | examples-openshift-with-htpasswd/4-check-authorized | Running in OpenShift logger.go:42: 07:46:36 | examples-openshift-with-htpasswd/4-check-authorized | Using Jaeger basic authentication logger.go:42: 07:46:36 | examples-openshift-with-htpasswd/4-check-authorized | Try number 1/30 the https://with-htpasswd-kuttl-test-enjoyed-cow.apps.ci-op-54q9kfc5-9210b.cspilp.interop.ccitredhat.com/search logger.go:42: 07:46:36 | examples-openshift-with-htpasswd/4-check-authorized | Something failed while trying to contact the server. Trying insecure mode logger.go:42: 07:46:36 | examples-openshift-with-htpasswd/4-check-authorized | Try number 2/30 the https://with-htpasswd-kuttl-test-enjoyed-cow.apps.ci-op-54q9kfc5-9210b.cspilp.interop.ccitredhat.com/search logger.go:42: 07:46:36 | examples-openshift-with-htpasswd/4-check-authorized | curl response asserted properly logger.go:42: 07:46:36 | examples-openshift-with-htpasswd/4-check-authorized | test step completed 4-check-authorized logger.go:42: 07:46:36 | examples-openshift-with-htpasswd | examples-openshift-with-htpasswd events from ns kuttl-test-enjoyed-cow: logger.go:42: 07:46:36 | examples-openshift-with-htpasswd | 2024-12-09 07:46:23 +0000 UTC Normal Pod with-htpasswd-858ddbf86d-qzscw Binding Scheduled Successfully assigned kuttl-test-enjoyed-cow/with-htpasswd-858ddbf86d-qzscw to ip-10-0-79-227.ec2.internal default-scheduler logger.go:42: 07:46:36 | examples-openshift-with-htpasswd | 2024-12-09 07:46:23 +0000 UTC Normal Pod with-htpasswd-858ddbf86d-qzscw AddedInterface Add eth0 [10.129.2.97/23] from ovn-kubernetes multus logger.go:42: 07:46:36 | examples-openshift-with-htpasswd | 2024-12-09 07:46:23 +0000 UTC Normal Pod with-htpasswd-858ddbf86d-qzscw.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1fa5ed13cc8a023f1e987f6cafe86adc195c373cc2b774539df6d0fd02b780a7" already present on machine kubelet logger.go:42: 07:46:36 | examples-openshift-with-htpasswd | 2024-12-09 07:46:23 +0000 UTC Normal Pod with-htpasswd-858ddbf86d-qzscw.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:46:36 | examples-openshift-with-htpasswd | 2024-12-09 07:46:23 +0000 UTC Normal Pod with-htpasswd-858ddbf86d-qzscw.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:46:36 | examples-openshift-with-htpasswd | 2024-12-09 07:46:23 +0000 UTC Normal Pod with-htpasswd-858ddbf86d-qzscw.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:46:36 | examples-openshift-with-htpasswd | 2024-12-09 07:46:23 +0000 UTC Normal ReplicaSet.apps with-htpasswd-858ddbf86d SuccessfulCreate Created pod: with-htpasswd-858ddbf86d-qzscw replicaset-controller logger.go:42: 07:46:36 | examples-openshift-with-htpasswd | 2024-12-09 07:46:23 +0000 UTC Normal Deployment.apps with-htpasswd ScalingReplicaSet Scaled up replica set with-htpasswd-858ddbf86d to 1 deployment-controller logger.go:42: 07:46:36 | examples-openshift-with-htpasswd | 2024-12-09 07:46:24 +0000 UTC Normal Pod with-htpasswd-858ddbf86d-qzscw.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:46:36 | examples-openshift-with-htpasswd | 2024-12-09 07:46:24 +0000 UTC Normal Pod with-htpasswd-858ddbf86d-qzscw.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:46:36 | examples-openshift-with-htpasswd | Deleting namespace: kuttl-test-enjoyed-cow === CONT kuttl/harness/examples-all-in-one-with-options logger.go:42: 07:46:42 | examples-all-in-one-with-options | Creating namespace: kuttl-test-square-mantis logger.go:42: 07:46:42 | examples-all-in-one-with-options/0-install | starting test step 0-install logger.go:42: 07:46:42 | examples-all-in-one-with-options/0-install | Jaeger:kuttl-test-square-mantis/my-jaeger created logger.go:42: 07:46:48 | examples-all-in-one-with-options/0-install | test step completed 0-install logger.go:42: 07:46:48 | examples-all-in-one-with-options/1-smoke-test | starting test step 1-smoke-test logger.go:42: 07:46:48 | examples-all-in-one-with-options/1-smoke-test | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE my-jaeger /dev/null] logger.go:42: 07:46:50 | examples-all-in-one-with-options/1-smoke-test | Warning: resource jaegers/my-jaeger is missing the kubectl.kubernetes.io/last-applied-configuration annotation which is required by kubectl apply. kubectl apply should only be used on resources created declaratively by either kubectl create --save-config or kubectl apply. The missing annotation will be patched automatically. logger.go:42: 07:46:58 | examples-all-in-one-with-options/1-smoke-test | running command: [sh -c ASSERT_IMG=quay.io/rhn_support_ikanse/jaeger-asserts:latest JAEGER_COLLECTOR_ENDPOINT=http://my-jaeger-collector-headless:14268 JAEGER_QUERY_ENDPOINT=https://my-jaeger-query:443/jaeger MOUNT_SECRET=e2e-test /tmp/jaeger-tests/bin/gomplate -f /tmp/jaeger-tests/tests/templates/smoke-test.yaml.template -o smoke-test-job.yaml] logger.go:42: 07:46:58 | examples-all-in-one-with-options/1-smoke-test | running command: [sh -c kubectl apply -f smoke-test-job.yaml -n $NAMESPACE] logger.go:42: 07:46:59 | examples-all-in-one-with-options/1-smoke-test | job.batch/report-span created logger.go:42: 07:46:59 | examples-all-in-one-with-options/1-smoke-test | job.batch/check-span created logger.go:42: 07:47:12 | examples-all-in-one-with-options/1-smoke-test | test step completed 1-smoke-test logger.go:42: 07:47:12 | examples-all-in-one-with-options | examples-all-in-one-with-options events from ns kuttl-test-square-mantis: logger.go:42: 07:47:12 | examples-all-in-one-with-options | 2024-12-09 07:46:46 +0000 UTC Normal Pod my-jaeger-779f74bcbd-vdsd5 Binding Scheduled Successfully assigned kuttl-test-square-mantis/my-jaeger-779f74bcbd-vdsd5 to ip-10-0-79-227.ec2.internal default-scheduler logger.go:42: 07:47:12 | examples-all-in-one-with-options | 2024-12-09 07:46:46 +0000 UTC Normal ReplicaSet.apps my-jaeger-779f74bcbd SuccessfulCreate Created pod: my-jaeger-779f74bcbd-vdsd5 replicaset-controller logger.go:42: 07:47:12 | examples-all-in-one-with-options | 2024-12-09 07:46:46 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-779f74bcbd to 1 deployment-controller logger.go:42: 07:47:12 | examples-all-in-one-with-options | 2024-12-09 07:46:47 +0000 UTC Normal Pod my-jaeger-779f74bcbd-vdsd5 AddedInterface Add eth0 [10.129.2.98/23] from ovn-kubernetes multus logger.go:42: 07:47:12 | examples-all-in-one-with-options | 2024-12-09 07:46:47 +0000 UTC Normal Pod my-jaeger-779f74bcbd-vdsd5.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1fa5ed13cc8a023f1e987f6cafe86adc195c373cc2b774539df6d0fd02b780a7" already present on machine kubelet logger.go:42: 07:47:12 | examples-all-in-one-with-options | 2024-12-09 07:46:47 +0000 UTC Normal Pod my-jaeger-779f74bcbd-vdsd5.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:47:12 | examples-all-in-one-with-options | 2024-12-09 07:46:47 +0000 UTC Normal Pod my-jaeger-779f74bcbd-vdsd5.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:47:12 | examples-all-in-one-with-options | 2024-12-09 07:46:47 +0000 UTC Normal Pod my-jaeger-779f74bcbd-vdsd5.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:47:12 | examples-all-in-one-with-options | 2024-12-09 07:46:47 +0000 UTC Normal Pod my-jaeger-779f74bcbd-vdsd5.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:47:12 | examples-all-in-one-with-options | 2024-12-09 07:46:47 +0000 UTC Normal Pod my-jaeger-779f74bcbd-vdsd5.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:47:12 | examples-all-in-one-with-options | 2024-12-09 07:46:55 +0000 UTC Normal Pod my-jaeger-779f74bcbd-vdsd5.spec.containers{jaeger} Killing Stopping container jaeger kubelet logger.go:42: 07:47:12 | examples-all-in-one-with-options | 2024-12-09 07:46:55 +0000 UTC Normal Pod my-jaeger-779f74bcbd-vdsd5.spec.containers{oauth-proxy} Killing Stopping container oauth-proxy kubelet logger.go:42: 07:47:12 | examples-all-in-one-with-options | 2024-12-09 07:46:55 +0000 UTC Normal ReplicaSet.apps my-jaeger-779f74bcbd SuccessfulDelete Deleted pod: my-jaeger-779f74bcbd-vdsd5 replicaset-controller logger.go:42: 07:47:12 | examples-all-in-one-with-options | 2024-12-09 07:46:55 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled down replica set my-jaeger-779f74bcbd to 0 from 1 deployment-controller logger.go:42: 07:47:12 | examples-all-in-one-with-options | 2024-12-09 07:46:56 +0000 UTC Normal Pod my-jaeger-76cc56868f-69xjf Binding Scheduled Successfully assigned kuttl-test-square-mantis/my-jaeger-76cc56868f-69xjf to ip-10-0-79-227.ec2.internal default-scheduler logger.go:42: 07:47:12 | examples-all-in-one-with-options | 2024-12-09 07:46:56 +0000 UTC Normal ReplicaSet.apps my-jaeger-76cc56868f SuccessfulCreate Created pod: my-jaeger-76cc56868f-69xjf replicaset-controller logger.go:42: 07:47:12 | examples-all-in-one-with-options | 2024-12-09 07:46:56 +0000 UTC Normal Deployment.apps my-jaeger ScalingReplicaSet Scaled up replica set my-jaeger-76cc56868f to 1 deployment-controller logger.go:42: 07:47:12 | examples-all-in-one-with-options | 2024-12-09 07:46:57 +0000 UTC Normal Pod my-jaeger-76cc56868f-69xjf AddedInterface Add eth0 [10.129.2.99/23] from ovn-kubernetes multus logger.go:42: 07:47:12 | examples-all-in-one-with-options | 2024-12-09 07:46:57 +0000 UTC Normal Pod my-jaeger-76cc56868f-69xjf.spec.containers{jaeger} Pulled Container image "registry.redhat.io/rhosdt/jaeger-all-in-one-rhel8@sha256:1fa5ed13cc8a023f1e987f6cafe86adc195c373cc2b774539df6d0fd02b780a7" already present on machine kubelet logger.go:42: 07:47:12 | examples-all-in-one-with-options | 2024-12-09 07:46:57 +0000 UTC Normal Pod my-jaeger-76cc56868f-69xjf.spec.containers{jaeger} Created Created container jaeger kubelet logger.go:42: 07:47:12 | examples-all-in-one-with-options | 2024-12-09 07:46:57 +0000 UTC Normal Pod my-jaeger-76cc56868f-69xjf.spec.containers{jaeger} Started Started container jaeger kubelet logger.go:42: 07:47:12 | examples-all-in-one-with-options | 2024-12-09 07:46:57 +0000 UTC Normal Pod my-jaeger-76cc56868f-69xjf.spec.containers{oauth-proxy} Pulled Container image "registry.redhat.io/openshift4/ose-oauth-proxy@sha256:5b2ca060dff5d64979497bd8b7144f8e115763984f2ae7d1ddf6ef416d636508" already present on machine kubelet logger.go:42: 07:47:12 | examples-all-in-one-with-options | 2024-12-09 07:46:57 +0000 UTC Normal Pod my-jaeger-76cc56868f-69xjf.spec.containers{oauth-proxy} Created Created container oauth-proxy kubelet logger.go:42: 07:47:12 | examples-all-in-one-with-options | 2024-12-09 07:46:57 +0000 UTC Normal Pod my-jaeger-76cc56868f-69xjf.spec.containers{oauth-proxy} Started Started container oauth-proxy kubelet logger.go:42: 07:47:12 | examples-all-in-one-with-options | 2024-12-09 07:46:59 +0000 UTC Normal Pod check-span-2pskn Binding Scheduled Successfully assigned kuttl-test-square-mantis/check-span-2pskn to ip-10-0-39-223.ec2.internal default-scheduler logger.go:42: 07:47:12 | examples-all-in-one-with-options | 2024-12-09 07:46:59 +0000 UTC Normal Pod check-span-2pskn AddedInterface Add eth0 [10.128.2.75/23] from ovn-kubernetes multus logger.go:42: 07:47:12 | examples-all-in-one-with-options | 2024-12-09 07:46:59 +0000 UTC Normal Pod check-span-2pskn.spec.containers{asserts-container} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:47:12 | examples-all-in-one-with-options | 2024-12-09 07:46:59 +0000 UTC Normal Job.batch check-span SuccessfulCreate Created pod: check-span-2pskn job-controller logger.go:42: 07:47:12 | examples-all-in-one-with-options | 2024-12-09 07:46:59 +0000 UTC Normal Pod report-span-2z86l Binding Scheduled Successfully assigned kuttl-test-square-mantis/report-span-2z86l to ip-10-0-95-250.ec2.internal default-scheduler logger.go:42: 07:47:12 | examples-all-in-one-with-options | 2024-12-09 07:46:59 +0000 UTC Normal Pod report-span-2z86l AddedInterface Add eth0 [10.131.0.50/23] from ovn-kubernetes multus logger.go:42: 07:47:12 | examples-all-in-one-with-options | 2024-12-09 07:46:59 +0000 UTC Normal Pod report-span-2z86l.spec.containers{report-span} Pulling Pulling image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" kubelet logger.go:42: 07:47:12 | examples-all-in-one-with-options | 2024-12-09 07:46:59 +0000 UTC Normal Job.batch report-span SuccessfulCreate Created pod: report-span-2z86l job-controller logger.go:42: 07:47:12 | examples-all-in-one-with-options | 2024-12-09 07:47:00 +0000 UTC Normal Pod check-span-2pskn.spec.containers{asserts-container} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 987ms (987ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:47:12 | examples-all-in-one-with-options | 2024-12-09 07:47:00 +0000 UTC Normal Pod check-span-2pskn.spec.containers{asserts-container} Created Created container asserts-container kubelet logger.go:42: 07:47:12 | examples-all-in-one-with-options | 2024-12-09 07:47:00 +0000 UTC Normal Pod check-span-2pskn.spec.containers{asserts-container} Started Started container asserts-container kubelet logger.go:42: 07:47:12 | examples-all-in-one-with-options | 2024-12-09 07:47:00 +0000 UTC Normal Pod report-span-2z86l.spec.containers{report-span} Pulled Successfully pulled image "quay.io/rhn_support_ikanse/jaeger-asserts:latest" in 883ms (883ms including waiting). Image size: 60976023 bytes. kubelet logger.go:42: 07:47:12 | examples-all-in-one-with-options | 2024-12-09 07:47:00 +0000 UTC Normal Pod report-span-2z86l.spec.containers{report-span} Created Created container report-span kubelet logger.go:42: 07:47:12 | examples-all-in-one-with-options | 2024-12-09 07:47:00 +0000 UTC Normal Pod report-span-2z86l.spec.containers{report-span} Started Started container report-span kubelet logger.go:42: 07:47:12 | examples-all-in-one-with-options | 2024-12-09 07:47:11 +0000 UTC Normal Job.batch check-span Completed Job completed job-controller logger.go:42: 07:47:12 | examples-all-in-one-with-options | Deleting namespace: kuttl-test-square-mantis === CONT kuttl/harness/examples-agent-with-priority-class logger.go:42: 07:47:24 | examples-agent-with-priority-class | Creating namespace: kuttl-test-flying-anteater logger.go:42: 07:47:24 | examples-agent-with-priority-class/0-install | starting test step 0-install logger.go:42: 07:47:24 | examples-agent-with-priority-class/0-install | SecurityContextConstraints:/daemonset-with-hostport created logger.go:42: 07:47:24 | examples-agent-with-priority-class/0-install | ServiceAccount:kuttl-test-flying-anteater/jaeger-agent-daemonset created logger.go:42: 07:47:24 | examples-agent-with-priority-class/0-install | test step completed 0-install logger.go:42: 07:47:24 | examples-agent-with-priority-class/1-add-policy | starting test step 1-add-policy logger.go:42: 07:47:24 | examples-agent-with-priority-class/1-add-policy | running command: [sh -c oc adm policy --namespace $NAMESPACE add-scc-to-user daemonset-with-hostport -z jaeger-agent-daemonset] logger.go:42: 07:47:24 | examples-agent-with-priority-class/1-add-policy | clusterrole.rbac.authorization.k8s.io/system:openshift:scc:daemonset-with-hostport added: "jaeger-agent-daemonset" logger.go:42: 07:47:24 | examples-agent-with-priority-class/1-add-policy | running command: [sh -c sleep 5] logger.go:42: 07:47:29 | examples-agent-with-priority-class/1-add-policy | test step completed 1-add-policy logger.go:42: 07:47:29 | examples-agent-with-priority-class/2-install | starting test step 2-install logger.go:42: 07:47:29 | examples-agent-with-priority-class/2-install | running command: [sh -c SERVICE_ACCOUNT_NAME=e2e-test ../../../../cmd-utils/get-token.sh $NAMESPACE agent-as-daemonset /dev/null] logger.go:42: 07:47:37 | examples-agent-with-priority-class/2-install | Error from server (NotFound): deployments.apps "agent-as-daemonset" not found logger.go:42: 07:47:37 | examples-agent-with-priority-class/2-install | command failure, skipping 2 additional commands logger.go:42: 07:47:37 | examples-agent-with-priority-class/2-install | PriorityClass:/high-priority created logger.go:42: 07:47:37 | examples-agent-with-priority-class/2-install | Jaeger:kuttl-test-flying-anteater/agent-as-daemonset updated case.go:364: failed in step 2-install case.go:366: exit status 1 logger.go:42: 07:47:37 | examples-agent-with-priority-class | examples-agent-with-priority-class events from ns kuttl-test-flying-anteater: logger.go:42: 07:47:37 | examples-agent-with-priority-class | Deleting namespace: kuttl-test-flying-anteater === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- FAIL: kuttl (509.13s) --- FAIL: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (6.12s) --- PASS: kuttl/harness/examples-simple-prod (72.57s) --- PASS: kuttl/harness/examples-with-sampling (56.06s) --- PASS: kuttl/harness/examples-with-cassandra (59.01s) --- PASS: kuttl/harness/examples-with-badger (41.20s) --- PASS: kuttl/harness/examples-simplest (35.56s) --- PASS: kuttl/harness/examples-simple-prod-with-volumes (65.08s) --- PASS: kuttl/harness/examples-business-application-injected-sidecar (49.49s) --- PASS: kuttl/harness/examples-service-types (39.56s) --- PASS: kuttl/harness/examples-openshift-with-htpasswd (23.67s) --- PASS: kuttl/harness/examples-all-in-one-with-options (41.56s) --- FAIL: kuttl/harness/examples-agent-with-priority-class (19.08s) FAIL + exit_code=1 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name examples --report --output /logs/artifacts/examples.xml ./artifacts/kuttl-report.xml time="2024-12-09T07:47:44Z" level=debug msg="Setting a new name for the test suites" time="2024-12-09T07:47:44Z" level=debug msg="Removing 'artifacts' TestCase" time="2024-12-09T07:47:44Z" level=debug msg="normalizing test case names" time="2024-12-09T07:47:44Z" level=debug msg="examples/artifacts -> examples_artifacts" time="2024-12-09T07:47:44Z" level=debug msg="examples/examples-simple-prod -> examples_examples_simple_prod" time="2024-12-09T07:47:44Z" level=debug msg="examples/examples-with-sampling -> examples_examples_with_sampling" time="2024-12-09T07:47:44Z" level=debug msg="examples/examples-with-cassandra -> examples_examples_with_cassandra" time="2024-12-09T07:47:44Z" level=debug msg="examples/examples-with-badger -> examples_examples_with_badger" time="2024-12-09T07:47:44Z" level=debug msg="examples/examples-simplest -> examples_examples_simplest" time="2024-12-09T07:47:44Z" level=debug msg="examples/examples-simple-prod-with-volumes -> examples_examples_simple_prod_with_volumes" time="2024-12-09T07:47:44Z" level=debug msg="examples/examples-business-application-injected-sidecar -> examples_examples_business_application_injected_sidecar" time="2024-12-09T07:47:44Z" level=debug msg="examples/examples-service-types -> examples_examples_service_types" time="2024-12-09T07:47:44Z" level=debug msg="examples/examples-openshift-with-htpasswd -> examples_examples_openshift_with_htpasswd" time="2024-12-09T07:47:44Z" level=debug msg="examples/examples-all-in-one-with-options -> examples_examples_all_in_one_with_options" time="2024-12-09T07:47:44Z" level=debug msg="examples/examples-agent-with-priority-class -> examples_examples_agent_with_priority_class" +---------------------------------------------------------+--------+ | NAME | RESULT | +---------------------------------------------------------+--------+ | examples_artifacts | passed | | examples_examples_simple_prod | passed | | examples_examples_with_sampling | passed | | examples_examples_with_cassandra | passed | | examples_examples_with_badger | passed | | examples_examples_simplest | passed | | examples_examples_simple_prod_with_volumes | passed | | examples_examples_business_application_injected_sidecar | passed | | examples_examples_service_types | passed | | examples_examples_openshift_with_htpasswd | passed | | examples_examples_all_in_one_with_options | passed | | examples_examples_agent_with_priority_class | failed | +---------------------------------------------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh generate false true + '[' 3 -ne 3 ']' + test_suite_name=generate + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/generate.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-generate make[2]: Entering directory '/tmp/jaeger-tests' test -s /tmp/jaeger-tests/bin/operator-sdk || curl -sLo /tmp/jaeger-tests/bin/operator-sdk https://github.com/operator-framework/operator-sdk/releases/download/v1.32.0/operator-sdk_`go env GOOS`_`go env GOARCH` ./hack/install/install-golangci-lint.sh Installing golangci-lint golangci-lint 1.55.2 is installed already ./hack/install/install-goimports.sh Installing goimports Try 0... go install golang.org/x/tools/cmd/goimports@v0.1.12 >>>> Formatting code... ./.ci/format.sh >>>> Building... ./hack/install/install-dependencies.sh Installing go dependencies Try 0... go mod download GOOS= GOARCH= CGO_ENABLED=0 GO111MODULE=on go build -ldflags "-X "github.com/jaegertracing/jaeger-operator/pkg/version".version="1.62.0" -X "github.com/jaegertracing/jaeger-operator/pkg/version".buildDate=2024-12-09T07:47:46Z -X "github.com/jaegertracing/jaeger-operator/pkg/version".defaultJaeger="1.62.0"" -o "bin/jaeger-operator" main.go JAEGER_VERSION="1.62.0" ./tests/e2e/generate/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.18.0-0.nightly-2024-12-05-191355 True False 42m Cluster version is 4.18.0-0.nightly-2024-12-05-191355' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.18.0-0.nightly-2024-12-05-191355 True False 42m Cluster version is 4.18.0-0.nightly-2024-12-05-191355' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/generate/render.sh ++ export SUITE_DIR=./tests/e2e/generate ++ SUITE_DIR=./tests/e2e/generate ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/generate ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + '[' true = true ']' + skip_test generate 'This test was skipped until https://github.com/jaegertracing/jaeger-operator/issues/2145 is fixed' + '[' 2 -ne 2 ']' + test_name=generate + message='This test was skipped until https://github.com/jaegertracing/jaeger-operator/issues/2145 is fixed' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/generate/_build + '[' _build '!=' _build ']' + rm -rf generate + warning 'generate: This test was skipped until https://github.com/jaegertracing/jaeger-operator/issues/2145 is fixed' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: generate: This test was skipped until https://github.com/jaegertracing/jaeger-operator/issues/2145 is fixed\e[0m' WAR: generate: This test was skipped until https://github.com/jaegertracing/jaeger-operator/issues/2145 is fixed make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running generate E2E tests' Running generate E2E tests + cd tests/e2e/generate/_build + set +e + KUBECONFIG=/tmp/kubeconfig-591533165 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-54q9kfc5-9210b.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 1 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === CONT kuttl/harness/artifacts logger.go:42: 07:48:03 | artifacts | Creating namespace: kuttl-test-epic-mayfly logger.go:42: 07:48:03 | artifacts | artifacts events from ns kuttl-test-epic-mayfly: logger.go:42: 07:48:03 | artifacts | Deleting namespace: kuttl-test-epic-mayfly === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (6.21s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (6.03s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name generate --report --output /logs/artifacts/generate.xml ./artifacts/kuttl-report.xml time="2024-12-09T07:48:09Z" level=debug msg="Setting a new name for the test suites" time="2024-12-09T07:48:09Z" level=debug msg="Removing 'artifacts' TestCase" time="2024-12-09T07:48:09Z" level=debug msg="normalizing test case names" time="2024-12-09T07:48:09Z" level=debug msg="generate/artifacts -> generate_artifacts" +--------------------+--------+ | NAME | RESULT | +--------------------+--------+ | generate_artifacts | passed | +--------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests' make[1]: Entering directory '/tmp/jaeger-tests' KAFKA_OLM=true ./hack/run-e2e-test-suite.sh upgrade false true + '[' 3 -ne 3 ']' + test_suite_name=upgrade + use_kind_cluster=false + jaeger_olm=true + timeout=5m + make prepare-e2e-tests USE_KIND_CLUSTER=false JAEGER_OLM=true make[2]: Entering directory '/tmp/jaeger-tests' make[2]: Nothing to be done for 'prepare-e2e-tests'. make[2]: Leaving directory '/tmp/jaeger-tests' + '[' true = true ']' + echo 'Skipping Jaeger Operator installation because JAEGER_OLM=true' Skipping Jaeger Operator installation because JAEGER_OLM=true + root_dir=/tmp/jaeger-tests/hack/../ + reports_dir=/logs/artifacts + mkdir -p /logs/artifacts + rm -f /logs/artifacts/upgrade.xml + cd /tmp/jaeger-tests/hack/../ + /tmp/jaeger-tests/hack/..//hack/install/install-kuttl.sh Installing kuttl kubectl-kuttl 0.15.0 is installed already + make render-e2e-tests-upgrade make[2]: Entering directory '/tmp/jaeger-tests' make docker JAEGER_VERSION=1.62.1 IMG="quay.io//jaeger-operator:next" make[3]: Entering directory '/tmp/jaeger-tests' [ ! -z "true" ] || docker build --build-arg=GOPROXY= --build-arg=VERSION="1.62.0" --build-arg=JAEGER_VERSION=1.62.1 --build-arg=TARGETARCH= --build-arg VERSION_DATE=2024-12-09T07:48:09Z --build-arg VERSION_PKG="github.com/jaegertracing/jaeger-operator/pkg/version" -t "quay.io//jaeger-operator:next" . make[3]: Leaving directory '/tmp/jaeger-tests' touch build-e2e-upgrade-image SKIP_ES_EXTERNAL=true IMG=quay.io//jaeger-operator:"1.62.0" JAEGER_OPERATOR_VERSION="1.62.0" JAEGER_VERSION="1.62.0" ./tests/e2e/upgrade/render.sh +++ kubectl get clusterversion ++ output='NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.18.0-0.nightly-2024-12-05-191355 True False 42m Cluster version is 4.18.0-0.nightly-2024-12-05-191355' ++ IS_OPENSHIFT=false ++ '[' '!' -z 'NAME VERSION AVAILABLE PROGRESSING SINCE STATUS version 4.18.0-0.nightly-2024-12-05-191355 True False 42m Cluster version is 4.18.0-0.nightly-2024-12-05-191355' ']' ++ warning 'Generating templates for an OpenShift cluster' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;33mWAR: Generating templates for an OpenShift cluster\e[0m' WAR: Generating templates for an OpenShift cluster ++ IS_OPENSHIFT=true ++ export KAFKA_USE_CUSTOM_PODSET ++ '[' -z 3.6.0 ']' ++ version_le 3.6.0 0.25.0 +++ echo 3.6.0 0.25.0 +++ tr ' ' '\n' +++ sort -V +++ head -n 1 ++ test 0.25.0 == 3.6.0 ++ KAFKA_USE_CUSTOM_PODSET=true ++ export IS_OPENSHIFT +++ dirname ./tests/e2e/upgrade/render.sh ++ export SUITE_DIR=./tests/e2e/upgrade ++ SUITE_DIR=./tests/e2e/upgrade ++ /tmp/jaeger-tests/hack/install/install-gomplate.sh Installing Gomplate gomplate 3.10.0 is installed already ++ /tmp/jaeger-tests/hack/install/install-yq.sh Installing yq yq 4.20.2 is installed already ++ /tmp/jaeger-tests/hack/install/install-kustomize.sh Installing kustomize kustomize 4.5.7 is installed already ++ export ELASTICSEARCH_NODECOUNT=1 ++ ELASTICSEARCH_NODECOUNT=1 ++ export ELASTICSEARCH_URL=http://elasticsearch ++ ELASTICSEARCH_URL=http://elasticsearch ++ export ELASTICSEARCH_PORT=:9200 ++ ELASTICSEARCH_PORT=:9200 ++ export CASSANDRA_SERVER=cassandra ++ CASSANDRA_SERVER=cassandra ++ export SERVICE_ACCOUNT_NAME=e2e-test ++ SERVICE_ACCOUNT_NAME=e2e-test ++ PROGRAMS_FOLDER=../../../.. ++ export ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ ASSERT_HTTP_CODE_PROGRAM=../../../../cmd-utils/assert-jaeger-http-code.sh ++ export CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ CHECK_JAEGER_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-version.sh ++ export CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ CHECK_JAEGER_OPERATOR_VERSION_PROGRAM=../../../../cmd-utils/check-jaeger-operator-version.sh ++ export GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ GET_TOKEN_PROGRAM=../../../../cmd-utils/get-token.sh ++ export TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ TEST_UI_CONFIG_PROGRAM=../../../../cmd-utils/uiconfig/main.go ++ export WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ WAIT_CRONJOB_PROGRAM=../../../../cmd-utils/wait-cronjob/main.go ++ export QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ QUERY_PROGRAM=../../../../assert-jobs/query/main.go ++ export REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ REPORTER_PROGRAM=../../../../assert-jobs/reporter/main.go ++ set -e +++ pwd ++ cd /tmp/jaeger-tests/./tests/e2e/upgrade ++ build_dir=_build ++ rm -rf _build ++ mkdir _build ++ find -maxdepth 1 -type d '!' -wholename . '!' -wholename ./_build ++ xargs -I '{}' cp -r '{}' _build ++ cd _build ++ info 'Rendering kuttl-test.yaml' ++ '[' 1 -ne 1 ']' ++ echo -e '\e[1;34mRendering kuttl-test.yaml\e[0m' Rendering kuttl-test.yaml ++ '[' true = true ']' ++ CRD_DIR= ++ export CRD_DIR ++ /tmp/jaeger-tests/bin/gomplate -f ../../../templates/kuttl-test.yaml.template -o ./kuttl-test.yaml ++ mkdir -p artifacts + export JAEGER_NAME + '[' true = true ']' + skip_test upgrade 'Test not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=upgrade + message='Test not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/upgrade/_build + '[' _build '!=' _build ']' + rm -rf upgrade + warning 'upgrade: Test not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: upgrade: Test not supported in OpenShift\e[0m' WAR: upgrade: Test not supported in OpenShift + '[' true = true ']' + skip_test upgrade-from-latest-release 'Test not supported in OpenShift' + '[' 2 -ne 2 ']' + test_name=upgrade-from-latest-release + message='Test not supported in OpenShift' +++ pwd ++ basename /tmp/jaeger-tests/tests/e2e/upgrade/_build + '[' _build '!=' _build ']' + rm -rf upgrade-from-latest-release + warning 'upgrade-from-latest-release: Test not supported in OpenShift' + '[' 1 -ne 1 ']' + echo -e '\e[1;33mWAR: upgrade-from-latest-release: Test not supported in OpenShift\e[0m' WAR: upgrade-from-latest-release: Test not supported in OpenShift make[2]: Leaving directory '/tmp/jaeger-tests' + echo 'Running upgrade E2E tests' Running upgrade E2E tests + cd tests/e2e/upgrade/_build + set +e + KUBECONFIG=/tmp/kubeconfig-591533165 + /tmp/jaeger-tests/hack/..//bin/kubectl-kuttl test --report xml === RUN kuttl harness.go:462: starting setup harness.go:252: running tests using configured kubeconfig. harness.go:275: Successful connection to cluster at: https://api.ci-op-54q9kfc5-9210b.cspilp.interop.ccitredhat.com:6443 harness.go:360: running tests harness.go:73: going to run test suite with timeout of 600 seconds for each step harness.go:372: testsuite: . has 1 tests === RUN kuttl/harness === RUN kuttl/harness/artifacts === PAUSE kuttl/harness/artifacts === CONT kuttl/harness/artifacts logger.go:42: 07:48:10 | artifacts | Creating namespace: kuttl-test-faithful-glowworm logger.go:42: 07:48:10 | artifacts | artifacts events from ns kuttl-test-faithful-glowworm: logger.go:42: 07:48:10 | artifacts | Deleting namespace: kuttl-test-faithful-glowworm === CONT kuttl harness.go:405: run tests finished harness.go:513: cleaning up harness.go:570: removing temp folder: "" --- PASS: kuttl (6.31s) --- PASS: kuttl/harness (0.00s) --- PASS: kuttl/harness/artifacts (6.13s) PASS + exit_code=0 + set -e + go install github.com/RH-QE-Distributed-Tracing/junitcli/cmd/junitcli@v1.0.6 + junitcli --suite-name upgrade --report --output /logs/artifacts/upgrade.xml ./artifacts/kuttl-report.xml time="2024-12-09T07:48:17Z" level=debug msg="Setting a new name for the test suites" time="2024-12-09T07:48:17Z" level=debug msg="Removing 'artifacts' TestCase" time="2024-12-09T07:48:17Z" level=debug msg="normalizing test case names" time="2024-12-09T07:48:17Z" level=debug msg="upgrade/artifacts -> upgrade_artifacts" +-------------------+--------+ | NAME | RESULT | +-------------------+--------+ | upgrade_artifacts | passed | +-------------------+--------+ + '[' '' '!=' true ']' + '[' false == true ']' make[1]: Leaving directory '/tmp/jaeger-tests'