Uploaded image for project: 'OpenShift Bugs'
  1. OpenShift Bugs
  2. OCPBUGS-58052

VM snapshot fails to restore

XMLWordPrintable

    • Icon: Bug Bug
    • Resolution: Unresolved
    • Icon: Critical Critical
    • None
    • 4.18.z
    • None
    • Quality / Stability / Reliability
    • False
    • Hide

      None

      Show
      None
    • None
    • None
    • None
    • None
    • None
    • None
    • None
    • None
    • None
    • None
    • None
    • None
    • None
    • None

      Description of problem:

          VM snapshot fails to restore

      Version-Release number of selected component (if applicable):

          Openshift Virtualization 4.18.3

      How reproducible:

      Always

      Steps to Reproduce:

          1. Take a snapshot of the affected VM using web interface
          2. Wait a couple days
          3. Poweroff the VM
          4. Try to restore the snapshot
          

      Actual results:

          Snapshot restore never completes

      Expected results:

          Snapshot is restore successfully

      Additional info:

          Not sure if it matters, but the VM had been previously migrated from one storage (HPE Alletra) to another storage (ODF) using Openshift's Migration. HPE storage uses HPE CSI driver. Migration completed successfully as far as we know.
      
      Here is the VM's yaml:
      
      sh-5.1# oc get vm rhv-manager -oyaml -n test-vms
      apiVersion: kubevirt.io/v1
      kind: VirtualMachine
      metadata:
        annotations:
          kubemacpool.io/transaction-timestamp: "2025-06-23T17:47:00.34952844Z"
          kubevirt.io/latest-observed-api-version: v1
          kubevirt.io/storage-observed-api-version: v1
          migrate-storage-class/rhv-manager-mig-fw4pwg: rhv-manager
          vm.kubevirt.io/validations: |
            [
              {
                "name": "minimal-required-memory",
                "path": "jsonpath::.spec.domain.memory.guest",
                "rule": "integer",
                "message": "This VM requires more memory.",
                "min": 1610612736
              }
            ]
        creationTimestamp: "2025-06-20T17:08:45Z"
        finalizers:
        - kubevirt.io/virtualMachineControllerFinalize
        generation: 12
        labels:
          app: rhv-manager
          vm.kubevirt.io/template: rhel8-server-small
          vm.kubevirt.io/template.namespace: openshift
          vm.kubevirt.io/template.revision: "1"
          vm.kubevirt.io/template.version: v0.32.2
        name: rhv-manager
        namespace: test-vms
        resourceVersion: "15788627"
        uid: 367debab-6cef-4bf3-99a6-6bfc465b5e38
      spec:
        dataVolumeTemplates:
        - apiVersion: cdi.kubevirt.io/v1beta1
          kind: DataVolume
          metadata:
            creationTimestamp: null
            name: rhv-manager-mig-fw4pwg
          spec:
            source:
              pvc:
                name: rhv-manager
                namespace: test-vms
            storage:
              resources:
                requests:
                  storage: "322122547200"
              storageClassName: ocs-storagecluster-ceph-rbd-virtualization
        runStrategy: Halted
        template:
          metadata:
            annotations:
              vm.kubevirt.io/flavor: small
              vm.kubevirt.io/os: rhel8
              vm.kubevirt.io/workload: server
            creationTimestamp: null
            labels:
              kubevirt.io/domain: rhv-manager
              kubevirt.io/size: small
              network.kubevirt.io/headlessService: headless
          spec:
            architecture: amd64
            domain:
              cpu:
                cores: 1
                sockets: 2
                threads: 1
              devices:
                disks:
                - bootOrder: 1
                  disk:
                    bus: sata
                  name: rootdisk
                interfaces:
                - bridge: {}
                  macAddress: 02:6c:b5:00:00:01
                  model: virtio
                  name: nic-01
                logSerialConsole: false
                rng: {}
              machine:
                type: pc-q35-rhel9.4.0
              memory:
                guest: 16Gi
              resources: {}
            networks:
            - multus:
                networkName: default/vlan4
              name: nic-01
            terminationGracePeriodSeconds: 180
            volumes:
            - dataVolume:
                name: rhv-manager-mig-fw4pwg
              name: rootdisk
        updateVolumesStrategy: Migration
      status:
        conditions:
        - lastProbeTime: "2025-06-23T17:47:26Z"
          lastTransitionTime: "2025-06-23T17:47:26Z"
          message: VMI does not exist
          reason: VMINotExists
          status: "False"
          type: Ready
        - lastProbeTime: null
          lastTransitionTime: null
          message: All of the VMI's DVs are bound and not running
          reason: AllDVsReady
          status: "True"
          type: DataVolumesReady
        - lastProbeTime: null
          lastTransitionTime: null
          status: "True"
          type: LiveMigratable
        - lastProbeTime: null
          lastTransitionTime: null
          status: "True"
          type: StorageLiveMigratable
        desiredGeneration: 12
        observedGeneration: 12
        printableStatus: Stopped
        restoreInProgress: restore-rhv-manager-snapshot-20250620-163822-1750702545908
        runStrategy: Halted
        volumeSnapshotStatuses:
        - enabled: true
          name: rootdisk
        volumeUpdateState:
          volumeMigrationState:
            migratedVolumes:
            - destinationPVCInfo:
                claimName: rhv-manager-mig-fw4pwg
                volumeMode: Block
              sourcePVCInfo:
                claimName: rhv-manager
                volumeMode: Block
              volumeName: rootdisk
      sh-5.1# 
      
      
      And here's the vmrestore object:
      
      
      sh-5.1# oc get vmrestore restore-rhv-manager-snapshot-20250620-163822-1750702545908 -n test-vms -oyaml
      apiVersion: snapshot.kubevirt.io/v1beta1
      kind: VirtualMachineRestore
      metadata:
        creationTimestamp: "2025-06-23T18:15:50Z"
        finalizers:
        - snapshot.kubevirt.io/vmrestore-protection
        generation: 1
        name: restore-rhv-manager-snapshot-20250620-163822-1750702545908
        namespace: test-vms
        ownerReferences:
        - apiVersion: kubevirt.io/v1
          blockOwnerDeletion: false
          kind: VirtualMachine
          name: rhv-manager
          uid: 367debab-6cef-4bf3-99a6-6bfc465b5e38
        resourceVersion: "15788647"
        uid: e909a4ca-4aea-4e2e-b197-5ee43d6b66c7
      spec:
        target:
          apiGroup: kubevirt.io
          kind: VirtualMachine
          name: rhv-manager
        virtualMachineSnapshotName: rhv-manager-snapshot-20250620-163822
      status:
        complete: false
        conditions:
        - lastProbeTime: null
          lastTransitionTime: "2025-06-23T18:15:51Z"
          reason: 'admission webhook "virtualmachine-validator.kubevirt.io" denied the request:
            DataVolumeTemplate entry spec.dataVolumeTemplate[0] must be referenced in the
            VMI template''s ''volumes'' list'
          status: "False"
          type: Progressing
        - lastProbeTime: null
          lastTransitionTime: "2025-06-23T18:15:51Z"
          reason: 'admission webhook "virtualmachine-validator.kubevirt.io" denied the request:
            DataVolumeTemplate entry spec.dataVolumeTemplate[0] must be referenced in the
            VMI template''s ''volumes'' list'
          status: "False"
          type: Ready
        deletedDataVolumes:
        - rhv-manager-mig-fw4pwg
        restores:
        - persistentVolumeClaim: restore-e909a4ca-4aea-4e2e-b197-5ee43d6b66c7-rootdisk
          volumeName: rootdisk
          volumeSnapshotName: vmsnapshot-5da734e6-7629-428e-8540-363ba1676773-volume-rootdisk
      sh-5.1# 
      
      
      So far, customer has not been able to restore the VM.

              kmajcher@redhat.com Krzysztof Majcher
              gpulido@redhat.com German Pulido (Inactive)
              None
              None
              Yu Li Yu Li
              None
              Votes:
              0 Vote for this issue
              Watchers:
              2 Start watching this issue

                Created:
                Updated: