Create an LVM software RAID (raid1)¶
The EFI system partition
partflags: ['esp'] must be
a physical partition in the main partition table of the disk, not under
LVM or mdadm software RAID.
During configuration of your custom bare metal host profile,
you can create an LVM-based software RAID device
raid1 by adding
type: raid1 to the
logicalVolume spec in
For the LVM RAID parameters description, refer to Container Cloud API: BareMetalHostProfile spec.
For a bare metal host profile configuration, refer to Create a custom bare metal host profile.
logicalVolume spec of the
raid1 type requires
at least two devices (partitions) in
volumeGroup where you
build a logical volume. For an LVM of the
one device is enough.
You can use flexible
size units throughout bare metal host profiles.
For example, you can now use either
sizeGiB: 0.1 or
when specifying a device size.
Mirantis recommends using only one parameter name type and units throughout
the configuration files. If both
size are used,
sizeGiB is ignored during deployment and the suffix is adjusted
accordingly. For example,
1.5Gi will be serialized as
1536Mi. The size
without units is counted in bytes. For example,
size: 120 means 120 bytes.
raid1 requires additional space to store the
metadata on a volume group, roughly 4 MB for each partition.
Therefore, you cannot create a logical volume of exactly the same
size as the partitions it works on.
For example, if you have two partitions of 10 GiB, the corresponding
raid1 logical volume size will be less than 10 GiB. For that
reason, you can either set
size: 0 to use all available
space on the volume group, or set a smaller size than the partition
size. For example, use
size: 9.9Gi instead of
size: 10Gi for the logical volume.
The following example illustrates an extract of
/ on the LVM
... devices: - device: workBy: "by_id,by_wwn,by_path,by_name" minSize: 200Gi type: hdd wipe: true partitions: - name: root_part1 size: 120Gi partitions: - name: rest_sda size: 0 - device: workBy: "by_id,by_wwn,by_path,by_name" minSize: 200Gi type: hdd wipe: true partitions: - name: root_part2 size: 120Gi partitions: - name: rest_sdb size: 0 volumeGroups: - name: vg-root devices: - partition: root_part1 - partition: root_part2 - name: vg-data devices: - partition: rest_sda - partition: rest_sdb logicalVolumes: - name: root type: raid1 ## <-- LVM raid1 vg: vg-root size: 119.9Gi - name: data type: linear vg: vg-data size: 0 fileSystems: - fileSystem: ext4 logicalVolume: root mountPoint: / mountOpts: "noatime,nodiratime" - fileSystem: ext4 logicalVolume: data mountPoint: /mnt/data
All data will be wiped during cluster deployment on devices
defined directly or indirectly in the
fileSystems list of
BareMetalHostProfile. For example:
A raw device partition with a file system on it
A device partition in a volume group with a logical volume that has a file system on it
An mdadm RAID device with a file system on it
An LVM RAID device with a file system on it
wipe field is always considered
true for these devices.
false value is ignored.
Therefore, to prevent data loss, move the necessary data from these file systems to another server beforehand, if required.