Uploaded image for project: 'OpenShift API for Data Protection'
  1. OpenShift API for Data Protection
  2. OADP-1946

Datamover backup over CephFS failed. volsync-src-vsb and pvc in 'pending' status

XMLWordPrintable

    • Icon: Bug Bug
    • Resolution: Done
    • Icon: Major Major
    • OADP 1.2.1
    • OADP 1.2.0
    • data-mover
    • False
    • Hide

      None

      Show
      None
    • False
    • ToDo
    • No
    • Moderate
    • 5
    • 0
    • Very Likely
    • 0
    • None
    • Unset
    • Unknown

      Description of problem:

      Running 3 cycles of Datamover backups using CephFS. 

      The 1st & 2nd cycles were completed, but the 3rd cycle failed.

      volsync-src-vsb pod & volsync-vsb-rcxtq-rep-src-cache pvc in 'pending' status.

      Events:
      [root@f01-h07-000-r640 playbooks]# oc get events -nopenshift-adp --sort-by=.metadata.creationTimestamp | grep pvc-f9cd4159-a393-4492-b0d8-cdad1e8875d3
      35m         Normal    ProvisioningSucceeded           persistentvolumeclaim/snapcontent-a3bece79-b51d-4167-8400-4ad8650459f6-pvc             Successfully provisioned volume pvc-f9cd4159-a393-4492-b0d8-cdad1e8875d3
      34m         Normal    SuccessfulAttachVolume          pod/vsb-rcxtq-pod                                                                      AttachVolume.Attach succeeded for volume "pvc-f9cd4159-a393-4492-b0d8-cdad1e8875d3"
      34m         Warning   FileSystemResizeFailed          pod/vsb-rcxtq-pod                                                                      MountVolume.NodeExpandVolume failed for volume "pvc-f9cd4159-a393-4492-b0d8-cdad1e8875d3" requested read-only file system
      34m         Warning   FileSystemResizeFailed          persistentvolumeclaim/snapcontent-a3bece79-b51d-4167-8400-4ad8650459f6-pvc             MountVolume.NodeExpandVolume failed for volume "pvc-f9cd4159-a393-4492-b0d8-cdad1e8875d3" requested read-only file system

       

      Version-Release number of selected component (if applicable):

      OCP 4.12.9

      ODF 4.12.2
      OADP 1.2.0-69
      Using CephFS

       

      How reproducible:

       

      Steps to Reproduce:
      1. Create NS with 10 pods (6GB PV size , 2G data) - CephFS
      2. Running 3 sequential backups 

      NAME                               STATUS            ERRORS   WARNINGS   CREATED                         EXPIRES   STORAGE LOCATION   SELECTOR
      dm-backup-fs-10pods-multisnaps-1   Completed         0        0          2023-05-17 14:56:36 +0000 UTC   29d       example-velero-1   <none>
      dm-backup-fs-10pods-multisnaps-2   Completed         0        0          2023-05-17 15:13:39 +0000 UTC   29d       example-velero-1   <none>
      dm-backup-fs-10pods-multisnaps-3   PartiallyFailed   1        0          2023-05-17 15:31:44 +0000 UTC   29d       example-velero-1   <none>

      Actual results:

      Backup failed, have PVC & VSB in 'pending' status.

      Expected results:

      Backup completed

      Additional info:

      The test ran twice:
      -Create NS

      • ran 3 backups
      • Delete the NS
      • Create NS
      • ran 3 backup

      In both cycles , the 3rd backup was failed.

      The file "CephFS_Bug_3cycles" include errors from ceph ,velero , vsm

            wnstb Wes Hayutin
            dvaanunu@redhat.com David Vaanunu
            Votes:
            0 Vote for this issue
            Watchers:
            2 Start watching this issue

              Created:
              Updated:
              Resolved: