-
Feature Request
-
Resolution: Unresolved
-
Undefined
-
None
-
None
-
None
-
False
-
-
False
- Outline the proposed title of this feature request.
Support for the ODF operator to cleanup ceph bluestore metadata from OSD disks before deployment.
What we propose is a new parameter for the StorageCluster where the user can specify that they want the OSD disks to be cleaned before the deployment. We think it's okay if this requires extra validation like adding an annotation to the StorageCluster understanding that this is a dangerous operation and whatnot.
- What is the nature and problem description of the request?
When the Ceph cluster is being created, before the OSDs get initialized, during the "prepare OSD" phase, if the disks to be used by the OSDs have previous metadata from an old install, the prepare task will fail and the ceph cluster won't be created. This is a expected behavior and requires users to cleanup the ceph metadata from disks before deploying the cluster.
In the past, the ceph metadata was only written into the first sectors of the disk, starting in ODF 4.18, this metadata is written to several parts of the disk. This impacts re-deployments where the disk cleaning relies on existing methods provided by the baremetal operator or the local storage operator (both tools will run wipefs). In previous versions the wipefs was enough since ceph metadata was stored at the beginning of the disk, in ODF 4.18, wipefs is not enough and installation will fail due to existing metadata. This was discussed with the ODF team here https://issues.redhat.com/browse/DFBUGS-1655.
- Why does the customer need this? (List the business requirements here)
Telco customers mainly rely on the disk cleaning methods from the baremetal operator (which will run a wipefs on all disks during the pre-provisioning phase). This was enough to clean ceph metadata in the past, this is no longer the case.
In most telco environments, clusters are not re-deployed in an ordered way: Uninstall -> Reinstall. Instead, clusters are powered off and re-deployed without removing anything from the running cluster.
- Are there any Documentation Requirements for this request?
https://issues.redhat.com/browse/DFBUGS-1655
- Is the request coming from a specific customer or subset of customers (on prem only, cloud only, etc). Please don't mention any customer related sensitive information.
This is something impacting telco customers running on-prem baremetal deployments.
- What is the urgency of the request?
A workaround exists and it's being validated by the ODF team (https://issues.redhat.com/browse/DFBUGS-1655). But the sooner we get this supported in the operator without a workaround, the better.
- Please provide contact information, in case of follow up questions.
mavazque@redhat.com,feferran@redhat.com
- is related to
-
DFBUGS-1655 [GSS]OSD disks are not being properly cleaned in ODF 4.18
-
- Closed
-
- relates to
-
OCPBUGS-57425 forceWipeDevicesAndDestroyAllData does not remove Ceph cluster information
-
- ASSIGNED
-