-
Bug
-
Resolution: Unresolved
-
Major
-
None
-
rhel-10.2
-
None
-
None
-
rhel-virt-storage
-
None
-
False
-
False
-
-
None
-
None
-
None
-
None
-
Unspecified
-
Unspecified
-
Unspecified
-
None
What were you trying to do that didn't work?
A live volume migration that gets aborted mid way leaves the source disk larger than it was originally.
What is the impact of this issue to you?
Guest will report
[ 68.973995] vda: detected capacity change from 20971520 to 138586112
Which may be unexpected to user
Please provide the package NVR for which the bug is seen:
How reproducible is this bug?:
100% with tested storage combination
(k8s node partition host directory->google cloud netapp volume)
Steps to reproduce
- Live storage migrate a kubevirt VM disk
- Cancel/Abort
Expected results
bash-5.1$ qemu-img info /var/run/kubevirt-private/vmi-disks/dv-disk/disk.img
image: /var/run/kubevirt-private/vmi-disks/dv-disk/disk.img
file format: raw
virtual size: 10 GiB (10737418240 bytes)
disk size: 1.24 GiB
Child node '/file':
filename: /var/run/kubevirt-private/vmi-disks/dv-disk/disk.img
protocol type: file
file length: 10 GiB (10737418240 bytes)
disk size: 1.24 GiB
Format specific information:
extent size hint: 1048576
(This is the source image, originally)
Actual results
bash-5.1$ qemu-img info /var/run/kubevirt-private/vmi-disks/dv-disk/disk.img
image: /var/run/kubevirt-private/vmi-disks/dv-disk/disk.img
file format: raw
virtual size: 66.1 GiB (70956089344 bytes)
disk size: 1.24 GiB
Child node '/file':
filename: /var/run/kubevirt-private/vmi-disks/dv-disk/disk.img
protocol type: file
file length: 66.1 GiB (70956089344 bytes)
disk size: 1.24 GiB
Format specific information:
extent size hint: 1048576
Source image post cancel/abort
Additional information:
The source is a 10Gi disk image on the hosts directory (k8s "hostpath provisioner")
And it is attempting a migration to gcnv (google cloud netapp volumes) volume which is significantly larger (128Gi).
Another issue (possibly same underlying cause) - storage migration attempts to transfer significantly larger amount of data compared to original disk size. I was seeing 65~ Gi DataTotal metric but my original volume was a modest 10Gi.
I assume something is off due to NetApp storage efficiencies, but I am not 100% sure.
I guess if someone could clarify if this is expected/not we can take it from there.
I can also provide the CNV/k8s level reproducer and environment.