-
Bug
-
Resolution: Done-Errata
-
Normal
-
4.15.0
-
Quality / Stability / Reliability
-
False
-
-
1
-
Low
-
No
-
None
-
None
-
OCP VE Sprint 240, OCP VE Sprint 241
-
2
-
In Progress
-
Release Note Not Required
-
None
-
None
-
None
-
None
-
None
Description of problem:
Whenever starting a cluster with multiple nodes and trying to attach multiple devices to them, the Cluster does not become ready. In this case all worker nodes have 2 loop devices with 3GB Block Storage attached. All the VolumeGroupNodeStatus Objects show as ready. apiVersion: lvm.topolvm.io/v1alpha1 kind: LVMCluster metadata: annotations: kubectl.kubernetes.io/last-applied-configuration: | {"apiVersion":"lvm.topolvm.io/v1alpha1","kind":"LVMCluster","metadata":{"annotations":{},"name":"my-lvmcluster","namespace":"openshift-storage"},"spec":{"storage":{"deviceClasses":[{"default":true,"fstype":"xfs","name":"vg1","thinPoolConfig":{"name":"thin-pool-1","overprovisionRatio":10,"sizePercent":90}}]}}} creationTimestamp: "2023-08-16T08:24:37Z" finalizers: - lvmcluster.topolvm.io generation: 1 name: my-lvmcluster namespace: openshift-storage resourceVersion: "46967" uid: fd5c50cd-c2d0-453d-828e-37dae45ddc38 spec: storage: deviceClasses: - default: true fstype: xfs name: vg1 thinPoolConfig: name: thin-pool-1 overprovisionRatio: 10 sizePercent: 90 status: deviceClassStatuses: - name: vg1 nodeStatus: - devices: - /dev/loop0 - /dev/loop1 node: ip-10-0-144-179.us-east-2.compute.internal status: Ready - devices: - /dev/loop0 - /dev/loop1 node: ip-10-0-168-219.us-east-2.compute.internal status: Ready - devices: - /dev/loop0 - /dev/loop1 node: ip-10-0-240-151.us-east-2.compute.internal status: Ready state: Progressing
Version-Release number of selected component (if applicable):
4.13-4.15
How reproducible:
100%
Steps to Reproduce:
1. Create a Multi Node Cluster with 3 worker nodes 2. Attach Loop Storage to them: 3. Create an LVM Cluster as seen above with a generic xfs formatting and a greedy lookup for devices (no deviceSelector) 4. Observer the Cluster does not get into ready even though all components are Ready.
Actual results:
Cluster does not get ready. When injecting a log message on the readiness check one can see in comes from the VG comparison: {"level":"info","ts":"2023-08-16T08:24:40Z","logger":"lvmcluster-controller","msg":"Verifying readiness","Request.Name":"my-lvmcluster","Request.Namespace":"openshift-storage","expectedVGCount":6,"readyVGCount":3} For some reason, the expectedVGCount is 6 while readyVGCount is only 3.
Expected results:
Cluster becomes ready and VGCounts match
Additional info:
- causes
-
OCPBUGS-15744 LVMCluster cluster reports that the state is "progressing" indefinitely even though storage has been successfully provisioned
-
- Closed
-
- links to
-
RHBA-2024:126443 LVMS 4.15 Bug Fix and Enhancement update