Uploaded image for project: 'RHEL'
  1. RHEL
  2. RHEL-149795

aborting live volume migration leaves the source disk image larger than original

Linking RHIVOS CVEs to...Migration: Automation ...Sync from "Extern...XMLWordPrintable

    • Icon: Bug Bug
    • Resolution: Unresolved
    • Icon: Major Major
    • None
    • rhel-10.2
    • qemu-kvm / Storage
    • None
    • None
    • rhel-virt-storage
    • None
    • False
    • False
    • Hide

      None

      Show
      None
    • None
    • None
    • None
    • None
    • Unspecified
    • Unspecified
    • Unspecified
    • None

      What were you trying to do that didn't work?

      A live volume migration that gets aborted mid way leaves the source disk larger than it was originally. 

      What is the impact of this issue to you?

      Guest will report

      [   68.973995] vda: detected capacity change from 20971520 to 138586112

      Which may be unexpected to user

      Please provide the package NVR for which the bug is seen:

      How reproducible is this bug?:

      100% with tested storage combination
      (k8s node partition host directory->google cloud netapp volume)

      Steps to reproduce

      1. Live storage migrate a kubevirt VM disk
      2. Cancel/Abort 
      3.  

      Expected results

       

      bash-5.1$ qemu-img info /var/run/kubevirt-private/vmi-disks/dv-disk/disk.img
      image: /var/run/kubevirt-private/vmi-disks/dv-disk/disk.img
      file format: raw
      virtual size: 10 GiB (10737418240 bytes)
      disk size: 1.24 GiB
      Child node '/file':
          filename: /var/run/kubevirt-private/vmi-disks/dv-disk/disk.img
          protocol type: file
          file length: 10 GiB (10737418240 bytes)
          disk size: 1.24 GiB
          Format specific information:
              extent size hint: 1048576

      (This is the source image, originally)

       

      Actual results

      bash-5.1$ qemu-img info /var/run/kubevirt-private/vmi-disks/dv-disk/disk.img 
      image: /var/run/kubevirt-private/vmi-disks/dv-disk/disk.img
      file format: raw
      virtual size: 66.1 GiB (70956089344 bytes)
      disk size: 1.24 GiB
      Child node '/file':
          filename: /var/run/kubevirt-private/vmi-disks/dv-disk/disk.img
          protocol type: file
          file length: 66.1 GiB (70956089344 bytes)
          disk size: 1.24 GiB
          Format specific information:
              extent size hint: 1048576

      Source image post cancel/abort

       

      Additional information:

      The source is a 10Gi disk image on the hosts directory (k8s "hostpath provisioner")
      And it is attempting a migration to gcnv (google cloud netapp volumes) volume which is significantly larger (128Gi).

      Another issue (possibly same underlying cause) - storage migration attempts to transfer significantly larger amount of data compared to original disk size. I was seeing 65~ Gi DataTotal metric but my original volume was a modest 10Gi.
      I assume something is off due to NetApp storage efficiencies, but I am not 100% sure. 
      I guess if someone could clarify if this is expected/not we can take it from there.

      I can also provide the CNV/k8s level reproducer and environment.

              virt-maint virt-maint
              akalenyu Alex Kalenyuk
              virt-maint virt-maint
              Qinghua Cheng Qinghua Cheng
              Votes:
              0 Vote for this issue
              Watchers:
              7 Start watching this issue

                Created:
                Updated: