-
Bug
-
Resolution: Done
-
Undefined
-
None
-
None
-
None
-
Quality / Stability / Reliability
-
False
-
-
None
-
Moderate
-
None
-
None
-
OTA 274, OTA 275
-
None
-
None
-
None
Found the issue https://redhat-internal.slack.com/archives/CJ1J9C3V4/p1755122638838849?thread_ts=1755112023.153649&cid=CJ1J9C3V4
The example of CI job:
Example of the output
Result:
curl -s https://gcsweb-ci.apps.ci.l2s4.p1.openshiftapps.com/gcs/test-platform-results/pr-logs/pull/30105/pull-ci-openshift-origin-main-e2e-aws-ovn-upgrade/1955686702345359360/artifacts/e2e-aws-ovn-upgrade/openshift-e2e-test/artifacts/junit/adm-upgrade-status/adm-upgrade-status-2025-08-13%2019:58:23.124650721%20+0000%20UTC%20m\=+1919.964244606__20250813-192642.txt Unable to fetch alerts, ignoring alerts in 'Update Health': failed to get alerts from Thanos: no token is currently in use for this session = Control Plane = Assessment: Progressing Target Version: 4.20.0-0.ci-2025-08-13-182454-test-ci-op-5wilvz46-latest (from 4.20.0-0.ci-2025-08-13-174821-test-ci-op-5wilvz46-initial) Updating: image-registry, monitoring, openshift-controller-manager Completion: 50% (17 operators updated, 3 updating, 14 waiting) Duration: 24m (Est. Time Remaining: 45m) Operator Health: 34 HealthyUpdating Cluster Operators NAME SINCE REASON MESSAGE image-registry 6s DeploymentNotCompleted::NodeCADaemonUnavailable NodeCADaemonProgressing: The daemon set node-ca is deploying node pods Progressing: The deployment has not completed monitoring 4s RollOutInProgress Rolling out the stack. openshift-controller-manager 11s RouteControllerManager_DesiredStateNotYetAchieved::_DesiredStateNotYetAchieved Progressing: deployment/controller-manager: observed generation is 10, desired generation is 11 Progressing: deployment/controller-manager: updated replicas is 1, desired replicas is 3 RouteControllerManagerProgressing: deployment/route-controller-manager: observed generation is 7, desired generation is 8 RouteControllerManagerProgressing: deployment/route-controller-manager: updated replicas is 1, desired replicas is 3Control Plane Nodes NAME ASSESSMENT PHASE VERSION EST MESSAGE ip-10-0-10-232.ec2.internal Outdated Pending 4.20.0-0.ci-2025-08-13-174821-test-ci-op-5wilvz46-initial ? ip-10-0-8-129.ec2.internal Outdated Pending 4.20.0-0.ci-2025-08-13-174821-test-ci-op-5wilvz46-initial ? ip-10-0-88-44.ec2.internal Outdated Pending 4.20.0-0.ci-2025-08-13-174821-test-ci-op-5wilvz46-initial ?= Worker Upgrade =WORKER POOL ASSESSMENT COMPLETION STATUS worker Pending 0% (0/3) 3 Available, 0 Progressing, 0 DrainingWorker Pool Nodes: worker NAME ASSESSMENT PHASE VERSION EST MESSAGE ip-10-0-47-75.ec2.internal Outdated Pending 4.20.0-0.ci-2025-08-13-174821-test-ci-op-5wilvz46-initial ? ip-10-0-57-235.ec2.internal Outdated Pending 4.20.0-0.ci-2025-08-13-174821-test-ci-op-5wilvz46-initial ? ip-10-0-64-121.ec2.internal Outdated Pending 4.20.0-0.ci-2025-08-13-174821-test-ci-op-5wilvz46-initial ?= Update Health = SINCE LEVEL IMPACT MESSAGE 24m12s Info None Update is proceeding well%
Expecting:
There should not be any line breaks that destroy the layout of the table.
Possible cause: The CO status.condistions.message contains line breaks.
curl -s https://gcsweb-ci.apps.ci.l2s4.p1.openshiftapps.com/gcs/test-platform-results/pr-logs/pull/30105/pull-ci-openshift-origin-main-e2e-aws-ovn-upgrade/1955686702345359360/artifacts/e2e-aws-ovn-upgrade/gather-extra/artifacts/inspect/cluster-scoped-resources/config.openshift.io/clusteroperators/image-registry.yaml --- apiVersion: config.openshift.io/v1 kind: ClusterOperator metadata: annotations: capability.openshift.io/name: ImageRegistry include.release.openshift.io/hypershift: "true" include.release.openshift.io/ibm-cloud-managed: "true" include.release.openshift.io/self-managed-high-availability: "true" include.release.openshift.io/single-node-developer: "true" creationTimestamp: "2025-08-13T18:52:02Z" generation: 1 managedFields: - apiVersion: config.openshift.io/v1 fieldsType: FieldsV1 fieldsV1: f:metadata: f:annotations: .: {} f:capability.openshift.io/name: {} f:include.release.openshift.io/hypershift: {} f:include.release.openshift.io/ibm-cloud-managed: {} f:include.release.openshift.io/self-managed-high-availability: {} f:include.release.openshift.io/single-node-developer: {} f:ownerReferences: .: {} k:{"uid":"afc2ae89-a344-4fa6-92a3-294edb109067"}: {} f:spec: {} manager: cluster-version-operator operation: Update time: "2025-08-13T18:52:02Z" - apiVersion: config.openshift.io/v1 fieldsType: FieldsV1 fieldsV1: f:status: .: {} f:extension: {} manager: cluster-version-operator operation: Update subresource: status time: "2025-08-13T18:52:03Z" - apiVersion: config.openshift.io/v1 fieldsType: FieldsV1 fieldsV1: f:status: f:conditions: .: {} k:{"type":"Available"}: .: {} f:lastTransitionTime: {} f:message: {} f:reason: {} f:status: {} f:type: {} k:{"type":"Degraded"}: .: {} f:lastTransitionTime: {} f:reason: {} f:status: {} f:type: {} k:{"type":"Progressing"}: .: {} f:lastTransitionTime: {} f:message: {} f:reason: {} f:status: {} f:type: {} f:relatedObjects: {} f:versions: {} manager: cluster-image-registry-operator operation: Update subresource: status time: "2025-08-13T20:32:38Z" name: image-registry ownerReferences: - apiVersion: config.openshift.io/v1 controller: true kind: ClusterVersion name: version uid: afc2ae89-a344-4fa6-92a3-294edb109067 resourceVersion: "77131" uid: f34b8002-1e36-4acd-a548-37ca46385a70 spec: {} status: conditions: - lastTransitionTime: "2025-08-13T19:07:20Z" message: |- NodeCADaemonAvailable: The daemon set node-ca has available replicas Available: The registry is ready ImagePrunerAvailable: Pruner CronJob has been created reason: Ready status: "True" type: Available - lastTransitionTime: "2025-08-13T20:32:38Z" message: |- NodeCADaemonProgressing: The daemon set node-ca is deployed Progressing: The registry is ready reason: Ready status: "False" type: Progressing - lastTransitionTime: "2025-08-13T19:06:52Z" reason: AsExpected status: "False" type: Degraded extension: null relatedObjects: - group: imageregistry.operator.openshift.io name: cluster resource: configs - group: imageregistry.operator.openshift.io name: cluster resource: imagepruners - group: rbac.authorization.k8s.io name: system:registry resource: clusterroles - group: rbac.authorization.k8s.io name: registry-registry-role resource: clusterrolebindings - group: rbac.authorization.k8s.io name: openshift-image-registry-pruner resource: clusterrolebindings - group: "" name: openshift-image-registry resource: namespaces versions: - name: operator version: 4.20.0-0.ci-2025-08-13-182454-test-ci-op-5wilvz46-latest
One simple fix could be replacing the line breaks in the message.
Attached inspect.zipfrom the above job's artifacts.