Uploaded image for project: 'OpenShift Bugs'
  1. OpenShift Bugs
  2. OCPBUGS-6759

New master couldn't be created when update cpms on ASH

    • Moderate
    • None
    • 2
    • Sprint 235, Sprint 237, Sprint 236, Sprint 238, Sprint 239, Sprint 240
    • 6
    • False
    • Hide

      None

      Show
      None
    • Hide
      Previously, clusters on Azure Stack Hub could not create new control plane nodes unless the nodes had custom disk sizes because the default disk size could not be validated. With this update, the default disk size has been set to 128 GB and the installation program enforces user-specified disk size values between 128 and 1023 GB. (link:https://issues.redhat.com/browse/OCPBUGS-6759[*OCPBUGS-6759*])
      Show
      Previously, clusters on Azure Stack Hub could not create new control plane nodes unless the nodes had custom disk sizes because the default disk size could not be validated. With this update, the default disk size has been set to 128 GB and the installation program enforces user-specified disk size values between 128 and 1023 GB. (link: https://issues.redhat.com/browse/OCPBUGS-6759 [* OCPBUGS-6759 *])
    • Bug Fix
    • Done

      Description of problem:

      Update cpms vmSize on ASH, got error "The value 1024 of parameter 'osDisk.diskSizeGB' is out of range. The value must be between '1' and '1023', inclusive." Target="osDisk.diskSizeGB"when provisioning new control plane node, change diskSizeGB to 1023, new nodes are provisioned. But for fresh install, the default diskSizeGB is 1024 for master.

      Version-Release number of selected component (if applicable):

      4.13.0-0.nightly-2023-01-27-165107

      How reproducible:

      Always

      Steps to Reproduce:

      1. Update cpms vmSize to Standard_DS3_v2
      2. Check new machine state
      $ oc get machine  
      NAME                                PHASE     TYPE              REGION   ZONE   AGE
      jima28b-r9zht-master-h7g67-1        Running   Standard_DS5_v2   mtcazs          11h
      jima28b-r9zht-master-hhfzl-0        Failed                                      24s
      jima28b-r9zht-master-qtb9j-0        Running   Standard_DS5_v2   mtcazs          11h
      jima28b-r9zht-master-tprc7-2        Running   Standard_DS5_v2   mtcazs          11h
      
      $ oc get machine jima28b-r9zht-master-hhfzl-0 -o yaml
        errorMessage: 'failed to reconcile machine "jima28b-r9zht-master-hhfzl-0": failed
          to create vm jima28b-r9zht-master-hhfzl-0: failure sending request for machine
          jima28b-r9zht-master-hhfzl-0: cannot create vm: compute.VirtualMachinesClient#CreateOrUpdate:
          Failure sending request: StatusCode=400 -- Original Error: Code="InvalidParameter"
          Message="The value 1024 of parameter ''osDisk.diskSizeGB'' is out of range. The
          value must be between ''1'' and ''1023'', inclusive." Target="osDisk.diskSizeGB"'
        errorReason: InvalidConfiguration
        lastUpdated: "2023-01-29T02:35:13Z"
        phase: Failed
        providerStatus:
          conditions:
          - lastTransitionTime: "2023-01-29T02:35:13Z"
            message: 'failed to create vm jima28b-r9zht-master-hhfzl-0: failure sending
              request for machine jima28b-r9zht-master-hhfzl-0: cannot create vm: compute.VirtualMachinesClient#CreateOrUpdate:
              Failure sending request: StatusCode=400 -- Original Error: Code="InvalidParameter"
              Message="The value 1024 of parameter ''osDisk.diskSizeGB'' is out of range.
              The value must be between ''1'' and ''1023'', inclusive." Target="osDisk.diskSizeGB"'
            reason: MachineCreationFailed
            status: "False"
            type: MachineCreated
          metadata: {}
      3. Checke logs
      $ oc logs -f machine-api-controllers-84444d49f-mlldl -c machine-controller
      I0129 02:35:15.047784       1 recorder.go:103] events "msg"="InvalidConfiguration: failed to reconcile machine \"jima28b-r9zht-master-hhfzl-0\": failed to create vm jima28b-r9zht-master-hhfzl-0: failure sending request for machine jima28b-r9zht-master-hhfzl-0: cannot create vm: compute.VirtualMachinesClient#CreateOrUpdate: Failure sending request: StatusCode=400 -- Original Error: Code=\"InvalidParameter\" Message=\"The value 1024 of parameter 'osDisk.diskSizeGB' is out of range. The value must be between '1' and '1023', inclusive.\" Target=\"osDisk.diskSizeGB\"" "object"={"kind":"Machine","namespace":"openshift-machine-api","name":"jima28b-r9zht-master-hhfzl-0","uid":"6cb07114-41a6-40bc-8e83-d9f27931bc8c","apiVersion":"machine.openshift.io/v1beta1","resourceVersion":"451889"} "reason"="FailedCreate" "type"="Warning"
      
       $ oc logs -f control-plane-machine-set-operator-69b756df4f-skv4x E0129 02:35:13.282358       1 controller.go:818]  "msg"="Observed failed replacement control plane machines" "error"="found replacement control plane machines in an error state, the following machines(s) are currently reporting an error: jima28b-r9zht-master-hhfzl-0" "controller"="controlplanemachineset" "failedReplacements"="jima28b-r9zht-master-hhfzl-0" "name"="cluster" "namespace"="openshift-machine-api" "reconcileID"="a988d699-8ddc-4880-9930-0db64ca51653" I0129 02:35:13.282380       1 controller.go:264]  "msg"="Cluster state is degraded. The control plane machine set will not take any action until issues have been resolved." "controller"="controlplanemachineset" "name"="cluster" "namespace"="openshift-machine-api" "reconcileID"="a988d699-8ddc-4880-9930-0db64ca51653" 
      4. Change diskSizeGB to 1023, new machine Provisioned.
                  osDisk:
                    diskSettings: {}
                    diskSizeGB: 1023
      
      $ oc get machine                  
      NAME                                PHASE      TYPE              REGION   ZONE   AGE
      jima28b-r9zht-master-h7g67-1        Running    Standard_DS5_v2   mtcazs          11h
      jima28b-r9zht-master-hhfzl-0        Deleting                                     7m1s
      jima28b-r9zht-master-qtb9j-0        Running    Standard_DS5_v2   mtcazs          12h
      jima28b-r9zht-master-tprc7-2        Running    Standard_DS5_v2   mtcazs          11h
      jima28b-r9zht-worker-mtcazs-p8d79   Running    Standard_DS3_v2   mtcazs          18h
      jima28b-r9zht-worker-mtcazs-x5gvh   Running    Standard_DS3_v2   mtcazs          18h
      jima28b-r9zht-worker-mtcazs-xmdvw   Running    Standard_DS3_v2   mtcazs          18h
      $ oc get machine        
      NAME                                PHASE         TYPE              REGION   ZONE   AGE
      jima28b-r9zht-master-h7g67-1        Running       Standard_DS5_v2   mtcazs          11h
      jima28b-r9zht-master-qtb9j-0        Running       Standard_DS5_v2   mtcazs          12h
      jima28b-r9zht-master-tprc7-2        Running       Standard_DS5_v2   mtcazs          11h
      jima28b-r9zht-master-vqd7r-0        Provisioned   Standard_DS3_v2   mtcazs          16s
      jima28b-r9zht-worker-mtcazs-p8d79   Running       Standard_DS3_v2   mtcazs          18h
      jima28b-r9zht-worker-mtcazs-x5gvh   Running       Standard_DS3_v2   mtcazs          18h
      jima28b-r9zht-worker-mtcazs-xmdvw   Running       Standard_DS3_v2   mtcazs          18h

      Actual results:

      For fresh install, the default diskSizeGB is 1024 for master. But update cpms vmSize, new master was created failed, report error "The value 1024 of parameter ''osDisk.diskSizeGB'' is out of range.  The value must be between ''1'' and ''1023'', inclusive"
      When changing diskSizeGB to 1023, new machine got Provisioned. 
      
      

      Expected results:

      New master could be created when change vmtype, and don't need update diskSizeGB to 1023.

      Additional info:

      Minimum recommendation for control plane nodes is 1024 GB
      https://docs.openshift.com/container-platform/4.12/installing/installing_azure_stack_hub/installing-azure-stack-hub-network-customizations.html#installation-azure-stack-hub-config-yaml_installing-azure-stack-hub-network-customizations

            [OCPBUGS-6759] New master couldn't be created when update cpms on ASH

            Errata Tool added a comment -

            Since the problem described in this issue should be resolved in a recent advisory, it has been closed.

            For information on the advisory (Important: OpenShift Container Platform 4.14.0 bug fix and security update), and where to find the updated files, follow the link below.

            If the solution does not work for you, open a new bug report.
            https://access.redhat.com/errata/RHSA-2023:5006

            Errata Tool added a comment - Since the problem described in this issue should be resolved in a recent advisory, it has been closed. For information on the advisory (Important: OpenShift Container Platform 4.14.0 bug fix and security update), and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. https://access.redhat.com/errata/RHSA-2023:5006

            Since the problem described in this issue should be resolved in a recent advisory, it has been closed.

            For information on the advisory (Important: OpenShift Container Platform 4.14.0 bug fix and security update), and where to find the updated files, follow the link below.

            If the solution does not work for you, open a new bug report.
            https://access.redhat.com/errata/RHSA-2023:5006

            Errata Tool added a comment - Since the problem described in this issue should be resolved in a recent advisory, it has been closed. For information on the advisory (Important: OpenShift Container Platform 4.14.0 bug fix and security update), and where to find the updated files, follow the link below. If the solution does not work for you, open a new bug report. https://access.redhat.com/errata/RHSA-2023:5006

            Jinyun Ma added a comment -

            based on pre-merge testing on comments here, move bug to VERIFIED.

            Jinyun Ma added a comment - based on pre-merge testing on comments here, move bug to VERIFIED.

            Jinyun Ma added a comment -

            pre-merge testing with payload including PR installer#7100 and passed.

            1. Without diskSizeGB setting on both ControlPlane and compute, the default disk size is 128GB.

            2. Set ControlPlane.platform.azure.osDisk.diskSizeGB as 120, compute.platform.azure.osDisk.diskSizeGB as 120, got error from installer:

            ERROR failed to fetch Master Machines: failed to load asset "Install Config": failed to create install config: invalid "install-config.yaml" file: [controlPlane.platform.azure.diskSizeGB: Invalid value: 120: Storage DiskSizeGB must be between 128 and 1023 inclusive for Azure Stack, compute[0].platform.azure.diskSizeGB: Invalid value: 120: Storage DiskSizeGB must be between 128 and 1023 inclusive for Azure Stack]  

            3. Set ControlPlane.platform.azure.osDisk.diskSizeGB as 1024, compute.platform.azure.osDisk.diskSizeGB as 1024, got error from installer:

            ERROR failed to fetch Master Machines: failed to load asset "Install Config": failed to create install config: invalid "install-config.yaml" file: [controlPlane.platform.azure.diskSizeGB: Invalid value: 1024: Storage DiskSizeGB must be between 128 and 1023 inclusive for Azure Stack, compute[0].platform.azure.diskSizeGB: Invalid value: 1024: Storage DiskSizeGB must be between 128 and 1023 inclusive for Azure Stack]

            4. Set both diskSizeGB as 1020, cluster is deployed successfully, operation to delete one master machine have expected result, new master machine is provisioned.

            $ oc get controlplanemachineset -n openshift-machine-api -ojson | jq -r '.items[].spec.template.machines_v1beta1_machine_openshift_io.spec.providerSpec.value.osDisk.diskSizeGB'
            1020
            
            $ oc get machine -n openshift-machine-api
            NAME                                    PHASE     TYPE              REGION   ZONE   AGE
            jima710026a-mhq6n-master-1              Running   Standard_DS4_v2   mtcazs          151m
            jima710026a-mhq6n-master-2              Running   Standard_DS4_v2   mtcazs          151m
            jima710026a-mhq6n-master-zfgbc-0        Running   Standard_DS4_v2   mtcazs          42m
            jima710026a-mhq6n-worker-mtcazs-87xzd   Running   Standard_DS3_v2   mtcazs          140m
            jima710026a-mhq6n-worker-mtcazs-cfzv6   Running   Standard_DS3_v2   mtcazs          140m
            jima710026a-mhq6n-worker-mtcazs-ct7tp   Running   Standard_DS3_v2   mtcazs          140m 

            Jinyun Ma added a comment - pre-merge testing with payload including PR installer#7100 and passed. 1. Without diskSizeGB setting on both ControlPlane and compute, the default disk size is 128GB. 2. Set ControlPlane.platform.azure.osDisk.diskSizeGB as 120, compute.platform.azure.osDisk.diskSizeGB as 120, got error from installer: ERROR failed to fetch Master Machines: failed to load asset "Install Config" : failed to create install config: invalid "install-config.yaml" file: [controlPlane.platform.azure.diskSizeGB: Invalid value: 120: Storage DiskSizeGB must be between 128 and 1023 inclusive for Azure Stack, compute[0].platform.azure.diskSizeGB: Invalid value: 120: Storage DiskSizeGB must be between 128 and 1023 inclusive for Azure Stack]  3. Set ControlPlane.platform.azure.osDisk.diskSizeGB as 1024, compute.platform.azure.osDisk.diskSizeGB as 1024, got error from installer: ERROR failed to fetch Master Machines: failed to load asset "Install Config" : failed to create install config: invalid "install-config.yaml" file: [controlPlane.platform.azure.diskSizeGB: Invalid value: 1024: Storage DiskSizeGB must be between 128 and 1023 inclusive for Azure Stack, compute[0].platform.azure.diskSizeGB: Invalid value: 1024: Storage DiskSizeGB must be between 128 and 1023 inclusive for Azure Stack] 4. Set both diskSizeGB as 1020, cluster is deployed successfully, operation to delete one master machine have expected result, new master machine is provisioned. $ oc get controlplanemachineset -n openshift-machine-api -ojson | jq -r '.items[].spec.template.machines_v1beta1_machine_openshift_io.spec.providerSpec.value.osDisk.diskSizeGB' 1020 $ oc get machine -n openshift-machine-api NAME                                    PHASE     TYPE              REGION   ZONE   AGE jima710026a-mhq6n-master-1              Running   Standard_DS4_v2   mtcazs          151m jima710026a-mhq6n-master-2              Running   Standard_DS4_v2   mtcazs          151m jima710026a-mhq6n-master-zfgbc-0        Running   Standard_DS4_v2   mtcazs          42m jima710026a-mhq6n-worker-mtcazs-87xzd   Running   Standard_DS3_v2   mtcazs          140m jima710026a-mhq6n-worker-mtcazs-cfzv6   Running   Standard_DS3_v2   mtcazs          140m jima710026a-mhq6n-worker-mtcazs-ct7tp   Running   Standard_DS3_v2   mtcazs          140m

            Jinyun Ma added a comment -

            pre-merge testing against payload 4.14.0-0.test-2023-05-05-025106-ci-ln-pxvdnrt-latest including installer PR#7100

            1. Not define controlPlane.platform.azure.osDisk.diskSizeGB in install-config.yaml

            Installation passed, default diskSize is 128GB for control plane machine disk.  `spec.template.machines_v1beta1_machine_openshift_io.spec.providerSpec.value.osDisk.diskSizeGB` in controlplanemachineset is also 128GB, new control plane nodes can be provisioned correctly after `oc delete` one master node

            $ oc get machine -n openshift-machine-api
            NAME                                PHASE     TYPE              REGION   ZONE   AGE
            jima05c-trbgp-master-1              Running   Standard_DS4_v2   mtcazs          130m
            jima05c-trbgp-master-2              Running   Standard_DS4_v2   mtcazs          130m
            jima05c-trbgp-master-hhmpl-0        Running   Standard_DS4_v2   mtcazs          90m
            jima05c-trbgp-worker-mtcazs-66blc   Running   Standard_DS3_v2   mtcazs          120m
            jima05c-trbgp-worker-mtcazs-mgf7v   Running   Standard_DS3_v2   mtcazs          120m
            jima05c-trbgp-worker-mtcazs-t62m5   Running   Standard_DS3_v2   mtcazs          120m
            

             

            2. Set controlPlane.platform.azure.osDisk.diskSizeGB as 1024 in install-config.yaml

            Installation passed, diskSize is 1023GB for control plane machine disk. But `spec.template.machines_v1beta1_machine_openshift_io.spec.providerSpec.value.osDisk.diskSizeGB` in controlplanemachineset is 1024, failed to provision new control plane node after `oc delete` one master node

            $ oc get machine -n openshift-machine-api
            NAME                                PHASE      TYPE              REGION   ZONE   AGE
            jima06a-5z465-master-0              Deleting   Standard_DS4_v2   mtcazs          44m
            jima06a-5z465-master-1              Running    Standard_DS4_v2   mtcazs          44m
            jima06a-5z465-master-2              Running    Standard_DS4_v2   mtcazs          44m
            jima06a-5z465-master-l8crw-0        Failed                                       4m1s
            jima06a-5z465-worker-mtcazs-7rgh4   Running    Standard_DS3_v2   mtcazs          36m
            jima06a-5z465-worker-mtcazs-fzcfx   Running    Standard_DS3_v2   mtcazs          36m
            jima06a-5z465-worker-mtcazs-wth2c   Running    Standard_DS3_v2   mtcazs          36m
            
            $ oc get controlplanemachineset -n openshift-machine-api -ojson | jq .items[].spec.template.machines_v1beta1_machine_openshift_io.spec
            {
              "lifecycleHooks": {},
              "metadata": {},
              "providerSpec": {
                "value": {
                  "apiVersion": "machine.openshift.io/v1beta1",
                  "availabilitySet": "jima06a-5z465-cluster",
                  "credentialsSecret": {
                    "name": "azure-cloud-credentials",
                    "namespace": "openshift-machine-api"
                  },
                  "diagnostics": {},
                  "image": {
                    "offer": "",
                    "publisher": "",
                    "resourceID": "/resourceGroups/jima06a-5z465-rg/providers/Microsoft.Compute/images/jima06a-5z465",
                    "sku": "",
                    "version": ""
                  },
                  "internalLoadBalancer": "jima06a-5z465-internal",
                  "kind": "AzureMachineProviderSpec",
                  "location": "mtcazs",
                  "metadata": {
                    "creationTimestamp": null
                  },
                  "networkResourceGroup": "jima06a-5z465-rg",
                  "osDisk": {
                    "diskSettings": {},
                    "diskSizeGB": 1024,
                    "managedDisk": {
                      "storageAccountType": "Premium_LRS"
                    },
                    "osType": "Linux"
                  },
                  "publicIP": false,
                  "publicLoadBalancer": "jima06a-5z465",
                  "resourceGroup": "jima06a-5z465-rg",
                  "subnet": "jima06a-5z465-master-subnet",
                  "userDataSecret": {
                    "name": "master-user-data"
                  },
                  "vmSize": "Standard_DS4_v2",
                  "vnet": "jima06a-5z465-vnet"
                }
              }
            }
            
            $ oc get machine jima06a-5z465-master-l8crw-0 -n openshift-machine-api -ojson | jq .status.errorMessage
            "failed to reconcile machine \"jima06a-5z465-master-l8crw-0\": failed to create vm jima06a-5z465-master-l8crw-0: failure sending request for machine jima06a-5z465-master-l8crw-0: cannot create vm: compute.VirtualMachinesClient#CreateOrUpdate: Failure sending request: StatusCode=400 -- Original Error: Code=\"InvalidParameter\" Message=\"The value 1024 of parameter 'osDisk.diskSizeGB' is out of range. The value must be between '1' and '1023', inclusive.\" Target=\"osDisk.diskSizeGB\""
            

            rna-afk could you help to check?

            Jinyun Ma added a comment - pre-merge testing against payload 4.14.0-0.test-2023-05-05-025106-ci-ln-pxvdnrt-latest including installer PR#7100 1. Not define controlPlane.platform.azure.osDisk.diskSizeGB in install-config.yaml Installation passed, default diskSize is 128GB for control plane machine disk.  `spec.template.machines_v1beta1_machine_openshift_io.spec.providerSpec.value.osDisk.diskSizeGB` in controlplanemachineset is also 128GB, new control plane nodes can be provisioned correctly after `oc delete` one master node $ oc get machine -n openshift-machine-api NAME                                PHASE     TYPE              REGION   ZONE   AGE jima05c-trbgp-master-1              Running   Standard_DS4_v2   mtcazs          130m jima05c-trbgp-master-2              Running   Standard_DS4_v2   mtcazs          130m jima05c-trbgp-master-hhmpl-0        Running   Standard_DS4_v2   mtcazs          90m jima05c-trbgp-worker-mtcazs-66blc   Running   Standard_DS3_v2   mtcazs          120m jima05c-trbgp-worker-mtcazs-mgf7v   Running   Standard_DS3_v2   mtcazs          120m jima05c-trbgp-worker-mtcazs-t62m5   Running   Standard_DS3_v2   mtcazs          120m   2. Set controlPlane.platform.azure.osDisk.diskSizeGB as 1024 in install-config.yaml Installation passed, diskSize is 1023GB for control plane machine disk. But `spec.template.machines_v1beta1_machine_openshift_io.spec.providerSpec.value.osDisk.diskSizeGB` in controlplanemachineset is 1024, failed to provision new control plane node after `oc delete` one master node $ oc get machine -n openshift-machine-api NAME                                PHASE      TYPE              REGION   ZONE   AGE jima06a-5z465-master-0              Deleting   Standard_DS4_v2   mtcazs          44m jima06a-5z465-master-1              Running    Standard_DS4_v2   mtcazs          44m jima06a-5z465-master-2              Running    Standard_DS4_v2   mtcazs          44m jima06a-5z465-master-l8crw-0        Failed                                       4m1s jima06a-5z465-worker-mtcazs-7rgh4   Running    Standard_DS3_v2   mtcazs          36m jima06a-5z465-worker-mtcazs-fzcfx   Running    Standard_DS3_v2   mtcazs          36m jima06a-5z465-worker-mtcazs-wth2c   Running    Standard_DS3_v2   mtcazs          36m $ oc get controlplanemachineset -n openshift-machine-api -ojson | jq .items[].spec.template.machines_v1beta1_machine_openshift_io.spec {   "lifecycleHooks" : {},   "metadata" : {},   "providerSpec" : {     "value" : {       "apiVersion" : "machine.openshift.io/v1beta1" ,       "availabilitySet" : "jima06a-5z465-cluster" ,       "credentialsSecret" : {         "name" : "azure-cloud-credentials" ,         "namespace" : "openshift-machine-api"       },       "diagnostics" : {},       "image" : {         "offer" : "",         "publisher" : "",         "resourceID" : "/resourceGroups/jima06a-5z465-rg/providers/Microsoft.Compute/images/jima06a-5z465" ,         "sku" : "",         "version" : ""       },       "internalLoadBalancer" : "jima06a-5z465-internal" ,       "kind" : "AzureMachineProviderSpec" ,       "location" : "mtcazs" ,       "metadata" : {         "creationTimestamp" : null       },       "networkResourceGroup" : "jima06a-5z465-rg" ,       "osDisk" : {         "diskSettings" : {},         "diskSizeGB" : 1024,         "managedDisk" : {           "storageAccountType" : "Premium_LRS"         },         "osType" : "Linux"       },       "publicIP" : false ,       "publicLoadBalancer" : "jima06a-5z465" ,       "resourceGroup" : "jima06a-5z465-rg" ,       "subnet" : "jima06a-5z465-master-subnet" ,       "userDataSecret" : {         "name" : "master-user-data"       },       "vmSize" : "Standard_DS4_v2" ,       "vnet" : "jima06a-5z465-vnet"     }   } } $ oc get machine jima06a-5z465-master-l8crw-0 -n openshift-machine-api -ojson | jq .status.errorMessage "failed to reconcile machine \" jima06a-5z465-master-l8crw-0\ ": failed to create vm jima06a-5z465-master-l8crw-0: failure sending request for machine jima06a-5z465-master-l8crw-0: cannot create vm: compute.VirtualMachinesClient#CreateOrUpdate: Failure sending request: StatusCode=400 -- Original Error: Code=\" InvalidParameter\ " Message=\" The value 1024 of parameter 'osDisk.diskSizeGB' is out of range. The value must be between '1' and '1023' , inclusive.\ " Target=\" osDisk.diskSizeGB\"" rna-afk could you help to check?

            Patrick Dillon added a comment - We should also check: https://github.com/openshift/installer/blob/master/data/data/azurestack/cluster/master/master.tf#L57 Seems like it should be max–not min.

            Patrick Dillon added a comment - - edited

            Let's work on this after FF, in sprint 233.

            Patrick Dillon added a comment - - edited Let's work on this after FF, in sprint 233.

            We can reduce the default disk size for ASH to be lower (I think 128 is default). ASH does not get the same disk performance improvements that Azure gets when increasing disk size. 

             

             

            Patrick Dillon added a comment - We can reduce the default disk size for ASH to be lower (I think 128 is default). ASH does not get the same disk performance improvements that Azure gets when increasing disk size.     

            Some initial discussion on Slack https://redhat-internal.slack.com/archives/C01V1DP387R/p1675362267448989. This will probably require further team discussions.

            Rafael Fonseca dos Santos added a comment - Some initial discussion on Slack https://redhat-internal.slack.com/archives/C01V1DP387R/p1675362267448989. This will probably require further team discussions.

            Jinyun Ma added a comment -

            I did some check after creating manifests file,

            In master machine file and master cpms file, diskSizeGB is 1024 in both files.

            [root@preserve-jima openshift]# cat 99_openshift-cluster-api_master-machines-0.yaml  | grep "diskSizeGB" -B2
                  osDisk:
                    diskSettings: {}
                    diskSizeGB: 1024
            [root@preserve-jima openshift]# cat 99_openshift-machine-api_master-control-plane-machine-set.yaml | grep "diskSizeGB" -B2
                        osDisk:
                          diskSettings: {}
                          diskSizeGB: 1024
            

            continue to install cluster, diskSizeGB in terraform tfvars is still 1024, but check disk size again after master node is provisioned, it changed to 1023.

            [root@preserve-jima openshift]# az vm show -n jimaashtest-5nkls-master-0 -g jimaashtest-5nkls-rg --query "storageProfile.osDisk.diskSizeGb"
            1023 

            Relevant code is https://github.com/openshift/installer/blob/master/data/data/azurestack/cluster/master/master.tf#L57, not sure why limit the max diskszie on azurestack cloud to 1023 while the default disk size is 1024?

            rna-afk , could you help to check?

            Jinyun Ma added a comment - I did some check after creating manifests file, In master machine file and master cpms file, diskSizeGB is 1024 in both files. [root@preserve-jima openshift]# cat 99_openshift-cluster-api_master-machines-0.yaml  | grep "diskSizeGB" -B2       osDisk:         diskSettings: {}         diskSizeGB: 1024 [root@preserve-jima openshift]# cat 99_openshift-machine-api_master-control-plane-machine-set.yaml | grep "diskSizeGB" -B2             osDisk:               diskSettings: {}               diskSizeGB: 1024 continue to install cluster, diskSizeGB in terraform tfvars is still 1024, but check disk size again after master node is provisioned, it changed to 1023. [root@preserve-jima openshift]# az vm show -n jimaashtest-5nkls-master-0 -g jimaashtest-5nkls-rg --query "storageProfile.osDisk.diskSizeGb" 1023 Relevant code is https://github.com/openshift/installer/blob/master/data/data/azurestack/cluster/master/master.tf#L57, not sure why limit the max diskszie on azurestack cloud to 1023 while the default disk size is 1024? rna-afk , could you help to check?

              rna-afk Aditya Narayanaswamy
              rhn-support-zhsun Zhaohua Sun
              Jinyun Ma Jinyun Ma
              Votes:
              0 Vote for this issue
              Watchers:
              8 Start watching this issue

                Created:
                Updated:
                Resolved: