From d01e701fb760393f61eaa0c7e54f7db24ffa3522 Mon Sep 17 00:00:00 2001 From: eabdullin Date: Mon, 15 Sep 2025 12:34:46 +0000 Subject: [PATCH] import CS python-blivet-3.6.0-28.el9 --- ...ces-file-if-disabled-or-doesnt-exist.patch | 83 +++ ...nal-information-in-PartitioningError.patch | 85 +++ SOURCES/0035-LVMPV-format-size-fix.patch | 572 ++++++++++++++++++ ...36-Make-ActionDestroyFormat-optional.patch | 258 ++++++++ ...ore-creating-it-as-well-as-the-start.patch | 384 ++++++++++++ ...ixup-function-for-LVM-logical-volume.patch | 65 ++ SPECS/python-blivet.spec | 31 +- 7 files changed, 1477 insertions(+), 1 deletion(-) create mode 100644 SOURCES/0033-Do-not-remove-PVs-from-devices-file-if-disabled-or-doesnt-exist.patch create mode 100644 SOURCES/0034-Include-additional-information-in-PartitioningError.patch create mode 100644 SOURCES/0035-LVMPV-format-size-fix.patch create mode 100644 SOURCES/0036-Make-ActionDestroyFormat-optional.patch create mode 100644 SOURCES/0037-Wipe-end-partition-before-creating-it-as-well-as-the-start.patch create mode 100644 SOURCES/0038-Add-a-pre-wipe-fixup-function-for-LVM-logical-volume.patch diff --git a/SOURCES/0033-Do-not-remove-PVs-from-devices-file-if-disabled-or-doesnt-exist.patch b/SOURCES/0033-Do-not-remove-PVs-from-devices-file-if-disabled-or-doesnt-exist.patch new file mode 100644 index 0000000..39d5a12 --- /dev/null +++ b/SOURCES/0033-Do-not-remove-PVs-from-devices-file-if-disabled-or-doesnt-exist.patch @@ -0,0 +1,83 @@ +From 3e3b8d415ca50c4feaaf8d3688f0ebda2522d866 Mon Sep 17 00:00:00 2001 +From: Vojtech Trefny +Date: Mon, 20 Jan 2025 13:02:50 +0100 +Subject: [PATCH] Do not remove PVs from devices file if disabled or doesn't + exists + +When the file doesn't exists the 'lvmdevices --deldev' call will +fail but it will still create the devices file. This means we now +have an empty devices file and all subsequent LVM calls will fail. + +Resolves: RHEL-84662 +--- + blivet/formats/lvmpv.py | 5 +++++ + tests/unit_tests/formats_tests/lvmpv_test.py | 22 ++++++++++++++++++++ + 2 files changed, 27 insertions(+) + +diff --git a/blivet/formats/lvmpv.py b/blivet/formats/lvmpv.py +index 51fa4a3c8..f5d71dbd1 100644 +--- a/blivet/formats/lvmpv.py ++++ b/blivet/formats/lvmpv.py +@@ -166,6 +166,11 @@ def lvmdevices_remove(self): + if not lvm.HAVE_LVMDEVICES: + raise PhysicalVolumeError("LVM devices file feature is not supported") + ++ if not os.path.exists(lvm.LVM_DEVICES_FILE): ++ log.debug("Not removing %s from devices file: %s doesn't exist", ++ self.device, lvm.LVM_DEVICES_FILE) ++ return ++ + try: + blockdev.lvm.devices_delete(self.device) + except blockdev.LVMError as e: +diff --git a/tests/unit_tests/formats_tests/lvmpv_test.py b/tests/unit_tests/formats_tests/lvmpv_test.py +index 6490c7d48..54a59026d 100644 +--- a/tests/unit_tests/formats_tests/lvmpv_test.py ++++ b/tests/unit_tests/formats_tests/lvmpv_test.py +@@ -41,6 +41,11 @@ def test_lvm_devices(self): + + mock["blockdev"].lvm.devices_add.assert_not_called() + ++ # LVM devices file not enabled/supported -> devices_delete should not be called ++ fmt._destroy() ++ ++ mock["blockdev"].lvm.devices_delete.assert_not_called() ++ + with self.patches() as mock: + # LVM devices file enabled and devices file exists -> devices_add should be called + mock["lvm"].HAVE_LVMDEVICES = True +@@ -50,6 +55,11 @@ def test_lvm_devices(self): + + mock["blockdev"].lvm.devices_add.assert_called_with("/dev/test") + ++ # LVM devices file enabled and devices file exists -> devices_delete should be called ++ fmt._destroy() ++ ++ mock["blockdev"].lvm.devices_delete.assert_called_with("/dev/test") ++ + with self.patches() as mock: + # LVM devices file enabled and devices file doesn't exist + # and no existing VGs present -> devices_add should be called +@@ -61,6 +71,12 @@ def test_lvm_devices(self): + + mock["blockdev"].lvm.devices_add.assert_called_with("/dev/test") + ++ # LVM devices file enabled but devices file doesn't exist ++ # -> devices_delete should not be called ++ fmt._destroy() ++ ++ mock["blockdev"].lvm.devices_delete.assert_not_called() ++ + with self.patches() as mock: + # LVM devices file enabled and devices file doesn't exist + # and existing VGs present -> devices_add should not be called +@@ -71,3 +87,9 @@ def test_lvm_devices(self): + fmt._create() + + mock["blockdev"].lvm.devices_add.assert_not_called() ++ ++ # LVM devices file enabled but devices file doesn't exist ++ # -> devices_delete should not be called ++ fmt._destroy() ++ ++ mock["blockdev"].lvm.devices_delete.assert_not_called() diff --git a/SOURCES/0034-Include-additional-information-in-PartitioningError.patch b/SOURCES/0034-Include-additional-information-in-PartitioningError.patch new file mode 100644 index 0000000..33e6fa9 --- /dev/null +++ b/SOURCES/0034-Include-additional-information-in-PartitioningError.patch @@ -0,0 +1,85 @@ +From 6d2e5c70fecc68e0d62255d4e2a65e9d264578dd Mon Sep 17 00:00:00 2001 +From: Vojtech Trefny +Date: Wed, 22 Jan 2025 13:16:43 +0100 +Subject: [PATCH] Include additional information in PartitioningError + +The generic 'Unable to allocate requested partition scheme' is not +very helpful, we should try to include additional information if +possible. + +Resolves: RHEL-8005 +--- + blivet/partitioning.py | 25 ++++++++++++++++++++++--- + 1 file changed, 22 insertions(+), 3 deletions(-) + +diff --git a/blivet/partitioning.py b/blivet/partitioning.py +index ce77e4eb7..0a35c764d 100644 +--- a/blivet/partitioning.py ++++ b/blivet/partitioning.py +@@ -34,7 +34,7 @@ + from .flags import flags + from .devices import Device, PartitionDevice, device_path_to_name + from .size import Size +-from .i18n import _ ++from .i18n import _, N_ + from .util import stringize, unicodeize, compare + + import logging +@@ -681,6 +681,11 @@ def resolve_disk_tags(disks, tags): + return [disk for disk in disks if any(tag in disk.tags for tag in tags)] + + ++class PartitioningErrors: ++ NO_PRIMARY = N_("no primary partition slots available") ++ NO_SLOTS = N_("no free partition slots") ++ ++ + def allocate_partitions(storage, disks, partitions, freespace, boot_disk=None): + """ Allocate partitions based on requested features. + +@@ -763,6 +768,7 @@ def allocate_partitions(storage, disks, partitions, freespace, boot_disk=None): + part_type = None + growth = 0 # in sectors + # loop through disks ++ errors = {} + for _disk in req_disks: + try: + disklabel = disklabels[_disk.path] +@@ -798,6 +804,10 @@ def allocate_partitions(storage, disks, partitions, freespace, boot_disk=None): + if new_part_type is None: + # can't allocate any more partitions on this disk + log.debug("no free partition slots on %s", _disk.name) ++ if PartitioningErrors.NO_SLOTS in errors.keys(): ++ errors[PartitioningErrors.NO_SLOTS].append(_disk.name) ++ else: ++ errors[PartitioningErrors.NO_SLOTS] = [_disk.name] + continue + + if _part.req_primary and new_part_type != parted.PARTITION_NORMAL: +@@ -808,7 +818,11 @@ def allocate_partitions(storage, disks, partitions, freespace, boot_disk=None): + new_part_type = parted.PARTITION_NORMAL + else: + # we need a primary slot and none are free on this disk +- log.debug("no primary slots available on %s", _disk.name) ++ log.debug("no primary partition slots available on %s", _disk.name) ++ if PartitioningErrors.NO_PRIMARY in errors.keys(): ++ errors[PartitioningErrors.NO_PRIMARY].append(_disk.name) ++ else: ++ errors[PartitioningErrors.NO_PRIMARY] = [_disk.name] + continue + elif _part.req_part_type is not None and \ + new_part_type != _part.req_part_type: +@@ -968,7 +982,12 @@ def allocate_partitions(storage, disks, partitions, freespace, boot_disk=None): + break + + if free is None: +- raise PartitioningError(_("Unable to allocate requested partition scheme.")) ++ if not errors: ++ msg = _("Unable to allocate requested partition scheme.") ++ else: ++ errors_by_disk = (", ".join(disks) + ": " + _(error) for error, disks in errors.items()) ++ msg = _("Unable to allocate requested partition scheme on requested disks:\n%s") % "\n".join(errors_by_disk) ++ raise PartitioningError(msg) + + _disk = use_disk + disklabel = _disk.format diff --git a/SOURCES/0035-LVMPV-format-size-fix.patch b/SOURCES/0035-LVMPV-format-size-fix.patch new file mode 100644 index 0000000..d3e0c69 --- /dev/null +++ b/SOURCES/0035-LVMPV-format-size-fix.patch @@ -0,0 +1,572 @@ +From 6a54de2780aa3fd52b4a25dc8db7ab8c5b1b8d4d Mon Sep 17 00:00:00 2001 +From: Vojtech Trefny +Date: Tue, 21 Jan 2025 10:03:17 +0100 +Subject: [PATCH 1/7] Use pvs info from static data to get PV size in PVSize + +No need for a special code for this, we can reuse the existing +code from LVM static data. +--- + blivet/tasks/pvtask.py | 12 ++++++------ + 1 file changed, 6 insertions(+), 6 deletions(-) + +diff --git a/blivet/tasks/pvtask.py b/blivet/tasks/pvtask.py +index b5bd72e0d..e93a61bc7 100644 +--- a/blivet/tasks/pvtask.py ++++ b/blivet/tasks/pvtask.py +@@ -27,6 +27,7 @@ + + from ..errors import PhysicalVolumeError + from ..size import Size, B ++from ..static_data import pvs_info + + from . import availability + from . import task +@@ -55,13 +56,12 @@ def do_task(self): # pylint: disable=arguments-differ + :raises :class:`~.errors.PhysicalVolumeError`: if size cannot be obtained + """ + +- try: +- pv_info = blockdev.lvm.pvinfo(self.pv.device) +- pv_size = pv_info.pv_size +- except blockdev.LVMError as e: +- raise PhysicalVolumeError(e) ++ pvs_info.drop_cache() ++ pv_info = pvs_info.cache.get(self.pv.device) ++ if pv_info is None: ++ raise PhysicalVolumeError("Failed to get PV info for %s" % self.pv.device) + +- return Size(pv_size) ++ return Size(pv_info.pv_size) + + + class PVResize(task.BasicApplication, dfresize.DFResizeTask): + +From 0b8239470762cc3b3732d2f40910be7e84102fa0 Mon Sep 17 00:00:00 2001 +From: Vojtech Trefny +Date: Tue, 21 Jan 2025 10:05:13 +0100 +Subject: [PATCH 2/7] Get the actual PV format size for LVMPV format + +--- + blivet/formats/lvmpv.py | 2 ++ + blivet/populator/helpers/lvm.py | 2 ++ + tests/unit_tests/populator_test.py | 2 ++ + 3 files changed, 6 insertions(+) + +diff --git a/blivet/formats/lvmpv.py b/blivet/formats/lvmpv.py +index f5d71dbd1..769c96e1d 100644 +--- a/blivet/formats/lvmpv.py ++++ b/blivet/formats/lvmpv.py +@@ -101,6 +101,8 @@ def __init__(self, **kwargs): + # when set to True, blivet will try to resize the PV to fill all available space + self._grow_to_fill = False + ++ self._target_size = self._size ++ + def __repr__(self): + s = DeviceFormat.__repr__(self) + s += (" vg_name = %(vg_name)s vg_uuid = %(vg_uuid)s" +diff --git a/blivet/populator/helpers/lvm.py b/blivet/populator/helpers/lvm.py +index 6ef2f4174..74641bcf8 100644 +--- a/blivet/populator/helpers/lvm.py ++++ b/blivet/populator/helpers/lvm.py +@@ -112,6 +112,8 @@ def _get_kwargs(self): + log.warning("PV %s has no pe_start", name) + if pv_info.pv_free: + kwargs["free"] = Size(pv_info.pv_free) ++ if pv_info.pv_size: ++ kwargs["size"] = Size(pv_info.pv_size) + + return kwargs + +diff --git a/tests/unit_tests/populator_test.py b/tests/unit_tests/populator_test.py +index 1ee29b57f..55b6be8d8 100644 +--- a/tests/unit_tests/populator_test.py ++++ b/tests/unit_tests/populator_test.py +@@ -1064,6 +1064,7 @@ def test_run(self, *args): + pv_info.vg_uuid = sentinel.vg_uuid + pv_info.pe_start = 0 + pv_info.pv_free = 0 ++ pv_info.pv_size = "10g" + + vg_device = Mock() + vg_device.id = 0 +@@ -1095,6 +1096,7 @@ def test_run(self, *args): + pv_info.vg_extent_count = 2500 + pv_info.vg_free_count = 0 + pv_info.vg_pv_count = 1 ++ pv_info.pv_size = "10g" + + with patch("blivet.static_data.lvm_info.PVsInfo.cache", new_callable=PropertyMock) as mock_pvs_cache: + mock_pvs_cache.return_value = {device.path: pv_info} + +From 14b9538a8fd9f5bfc7d744902517739b6fae7a22 Mon Sep 17 00:00:00 2001 +From: Vojtech Trefny +Date: Tue, 21 Jan 2025 13:35:38 +0100 +Subject: [PATCH 3/7] Update PV format size after adding/removing the PV + to/from the VG + +Unfortunately LVM substracts VG metadata from the reported PV size +so we need to make sure to update the size after the vgextend and +vgreduce operation. +--- + blivet/devices/lvm.py | 12 ++++++++++++ + 1 file changed, 12 insertions(+) + +diff --git a/blivet/devices/lvm.py b/blivet/devices/lvm.py +index 62974443e..85850d8e8 100644 +--- a/blivet/devices/lvm.py ++++ b/blivet/devices/lvm.py +@@ -315,9 +315,21 @@ def _remove(self, member): + if lv.status and not status: + lv.teardown() + ++ # update LVMPV format size --> PV format has different size when in VG ++ try: ++ member.format._size = member.format._target_size = member.format._size_info.do_task() ++ except errors.PhysicalVolumeError as e: ++ log.warning("Failed to obtain current size for device %s: %s", member.format.device, e) ++ + def _add(self, member): + blockdev.lvm.vgextend(self.name, member.path) + ++ # update LVMPV format size --> PV format has different size when in VG ++ try: ++ member.format._size = member.format._target_size = member.format._size_info.do_task() ++ except errors.PhysicalVolumeError as e: ++ log.warning("Failed to obtain current size for device %s: %s", member.path, e) ++ + def _add_log_vol(self, lv): + """ Add an LV to this VG. """ + if lv in self._lvs: + +From d6b0c283eb3236f3578dc28d40182f48d05a5c24 Mon Sep 17 00:00:00 2001 +From: Vojtech Trefny +Date: Tue, 21 Jan 2025 14:22:07 +0100 +Subject: [PATCH 4/7] Use LVMPV format size when calculating VG size and free + space + +For existing PVs we need to check the format size instead of +simply expecting the format is fully resized to match the size of +the underlying block device. +--- + blivet/devices/lvm.py | 63 ++++++++++++++++++++++++++----------------- + 1 file changed, 39 insertions(+), 24 deletions(-) + +diff --git a/blivet/devices/lvm.py b/blivet/devices/lvm.py +index 85850d8e8..e3d08dbce 100644 +--- a/blivet/devices/lvm.py ++++ b/blivet/devices/lvm.py +@@ -500,40 +500,55 @@ def reserved_percent(self, value): + + self._reserved_percent = value + +- def _get_pv_usable_space(self, pv): ++ def _get_pv_metadata_space(self, pv): ++ """ Returns how much space will be used by VG metadata in given PV ++ This depends on type of the PV, PE size and PE start. ++ """ + if isinstance(pv, MDRaidArrayDevice): +- return self.align(pv.size - 2 * pv.format.pe_start) ++ return 2 * pv.format.pe_start ++ else: ++ return pv.format.pe_start ++ ++ def _get_pv_usable_space(self, pv): ++ """ Return how much space can be actually used on given PV. ++ This takes into account: ++ - VG metadata that is/will be stored in this PV ++ - the actual PV format size (which might differ from ++ the underlying block device size) ++ """ ++ ++ if pv.format.exists and pv.format.size and self.exists: ++ # PV format exists, we got its size and VG also exists ++ # -> all metadata is already accounted in the PV format size ++ return pv.format.size ++ elif pv.format.exists and pv.format.size and not self.exists: ++ # PV format exists, we got its size, but the VG doesn't exist ++ # -> metadata size is not accounted in the PV format size ++ return self.align(pv.format.size - self._get_pv_metadata_space(pv)) + else: +- return self.align(pv.size - pv.format.pe_start) ++ # something else -> either the PV format is not yet created or ++ # we for some reason failed to get size of the format, either way ++ # lets use the underlying block device size and calculate the ++ # metadata size ourselves ++ return self.align(pv.size - self._get_pv_metadata_space(pv)) + + @property + def lvm_metadata_space(self): +- """ The amount of the space LVM metadata cost us in this VG's PVs """ +- # NOTE: we either specify data alignment in a PV or the default is used +- # which is both handled by pv.format.pe_start, but LVM takes into +- # account also the underlying block device which means that e.g. +- # for an MD RAID device, it tries to align everything also to chunk +- # size and alignment offset of such device which may result in up +- # to a twice as big non-data area +- # TODO: move this to either LVMPhysicalVolume's pe_start property once +- # formats know about their devices or to a new LVMPhysicalVolumeDevice +- # class once it exists +- diff = Size(0) +- for pv in self.pvs: +- diff += pv.size - self._get_pv_usable_space(pv) +- +- return diff ++ """ The amount of the space LVM metadata cost us in this VG's PVs ++ Note: we either specify data alignment in a PV or the default is used ++ which is both handled by pv.format.pe_start, but LVM takes into ++ account also the underlying block device which means that e.g. ++ for an MD RAID device, it tries to align everything also to chunk ++ size and alignment offset of such device which may result in up ++ to a twice as big non-data area ++ """ ++ return sum(self._get_pv_metadata_space(pv) for pv in self.pvs) + + @property + def size(self): + """ The size of this VG """ + # TODO: just ask lvm if isModified returns False +- +- # sum up the sizes of the PVs, subtract the unusable (meta data) space +- size = sum(pv.size for pv in self.pvs) +- size -= self.lvm_metadata_space +- +- return size ++ return sum(self._get_pv_usable_space(pv) for pv in self.pvs) + + @property + def extents(self): + +From 4d033869de8c22f627cc23e70023e82d9c6e90ed Mon Sep 17 00:00:00 2001 +From: Vojtech Trefny +Date: Tue, 21 Jan 2025 14:28:56 +0100 +Subject: [PATCH 5/7] Add more tests for PV and VG size and free space + +--- + tests/storage_tests/devices_test/lvm_test.py | 104 ++++++++++++++++++- + 1 file changed, 103 insertions(+), 1 deletion(-) + +diff --git a/tests/storage_tests/devices_test/lvm_test.py b/tests/storage_tests/devices_test/lvm_test.py +index 97ef1c4b9..988201839 100644 +--- a/tests/storage_tests/devices_test/lvm_test.py ++++ b/tests/storage_tests/devices_test/lvm_test.py +@@ -22,6 +22,18 @@ def setUp(self): + self.assertIsNone(disk.format.type) + self.assertFalse(disk.children) + ++ def _get_pv_size(self, pv): ++ out = subprocess.check_output(["pvs", "-o", "pv_size", "--noheadings", "--nosuffix", "--units=b", pv]) ++ return blivet.size.Size(out.decode().strip()) ++ ++ def _get_vg_size(self, vg): ++ out = subprocess.check_output(["vgs", "-o", "vg_size", "--noheadings", "--nosuffix", "--units=b", vg]) ++ return blivet.size.Size(out.decode().strip()) ++ ++ def _get_vg_free(self, vg): ++ out = subprocess.check_output(["vgs", "-o", "vg_free", "--noheadings", "--nosuffix", "--units=b", vg]) ++ return blivet.size.Size(out.decode().strip()) ++ + def _clean_up(self): + self.storage.reset() + for disk in self.storage.disks: +@@ -63,6 +75,8 @@ def test_lvm_basic(self): + self.assertIsInstance(pv, blivet.devices.PartitionDevice) + self.assertIsNotNone(pv.format) + self.assertEqual(pv.format.type, "lvmpv") ++ pv_size = self._get_pv_size(pv.path) ++ self.assertEqual(pv.format.size, pv_size) + + vg = self.storage.devicetree.get_device_by_name("blivetTestVG") + self.assertIsNotNone(vg) +@@ -72,6 +86,10 @@ def test_lvm_basic(self): + self.assertEqual(pv.format.vg_name, vg.name) + self.assertEqual(len(vg.parents), 1) + self.assertEqual(vg.parents[0], pv) ++ vg_size = self._get_vg_size(vg.name) ++ self.assertEqual(vg_size, vg.size) ++ vg_free = self._get_vg_free(vg.name) ++ self.assertEqual(vg_free, vg.free_space) + + lv = self.storage.devicetree.get_device_by_name("blivetTestVG-blivetTestLV") + self.assertIsNotNone(lv) +@@ -112,6 +130,13 @@ def test_lvm_thin(self): + self.storage.do_it() + self.storage.reset() + ++ vg = self.storage.devicetree.get_device_by_name("blivetTestVG") ++ self.assertIsNotNone(vg) ++ vg_size = self._get_vg_size(vg.name) ++ self.assertEqual(vg_size, vg.size) ++ vg_free = self._get_vg_free(vg.name) ++ self.assertEqual(vg_free, vg.free_space) ++ + pool = self.storage.devicetree.get_device_by_name("blivetTestVG-blivetTestPool") + self.assertIsNotNone(pool) + self.assertTrue(pool.is_thin_pool) +@@ -158,6 +183,14 @@ def _test_lvm_raid(self, seg_type, raid_level, stripe_size=0): + self.storage.do_it() + self.storage.reset() + ++ vg = self.storage.devicetree.get_device_by_name("blivetTestVG") ++ self.assertIsNotNone(vg) ++ ++ vg_size = self._get_vg_size(vg.name) ++ self.assertEqual(vg_size, vg.size) ++ vg_free = self._get_vg_free(vg.name) ++ self.assertEqual(vg_free, vg.free_space + vg.reserved_space) ++ + raidlv = self.storage.devicetree.get_device_by_name("blivetTestVG-blivetTestRAIDLV") + self.assertIsNotNone(raidlv) + self.assertTrue(raidlv.is_raid_lv) +@@ -214,6 +247,13 @@ def test_lvm_cache(self): + self.storage.do_it() + self.storage.reset() + ++ vg = self.storage.devicetree.get_device_by_name("blivetTestVG") ++ self.assertIsNotNone(vg) ++ vg_size = self._get_vg_size(vg.name) ++ self.assertEqual(vg_size, vg.size) ++ vg_free = self._get_vg_free(vg.name) ++ self.assertEqual(vg_free, vg.free_space) ++ + cachedlv = self.storage.devicetree.get_device_by_name("blivetTestVG-blivetTestCachedLV") + self.assertIsNotNone(cachedlv) + self.assertTrue(cachedlv.cached) +@@ -253,6 +293,13 @@ def test_lvm_cache_attach(self): + self.storage.do_it() + self.storage.reset() + ++ vg = self.storage.devicetree.get_device_by_name("blivetTestVG") ++ self.assertIsNotNone(vg) ++ vg_size = self._get_vg_size(vg.name) ++ self.assertEqual(vg_size, vg.size) ++ vg_free = self._get_vg_free(vg.name) ++ self.assertEqual(vg_free, vg.free_space) ++ + cachedlv = self.storage.devicetree.get_device_by_name("blivetTestVG-blivetTestCachedLV") + self.assertIsNotNone(cachedlv) + cachepool = self.storage.devicetree.get_device_by_name("blivetTestVG-blivetTestFastLV") +@@ -308,6 +355,13 @@ def test_lvm_cache_create_and_attach(self): + self.storage.do_it() + self.storage.reset() + ++ vg = self.storage.devicetree.get_device_by_name("blivetTestVG") ++ self.assertIsNotNone(vg) ++ vg_size = self._get_vg_size(vg.name) ++ self.assertEqual(vg_size, vg.size) ++ vg_free = self._get_vg_free(vg.name) ++ self.assertEqual(vg_free, vg.free_space) ++ + cachedlv = self.storage.devicetree.get_device_by_name("blivetTestVG-blivetTestCachedLV") + self.assertIsNotNone(cachedlv) + +@@ -323,6 +377,13 @@ def test_lvm_cache_create_and_attach(self): + self.storage.do_it() + self.storage.reset() + ++ vg = self.storage.devicetree.get_device_by_name("blivetTestVG") ++ self.assertIsNotNone(vg) ++ vg_size = self._get_vg_size(vg.name) ++ self.assertEqual(vg_size, vg.size) ++ vg_free = self._get_vg_free(vg.name) ++ self.assertEqual(vg_free, vg.free_space) ++ + cachedlv = self.storage.devicetree.get_device_by_name("blivetTestVG-blivetTestCachedLV") + self.assertIsNotNone(cachedlv) + self.assertTrue(cachedlv.cached) +@@ -352,6 +413,13 @@ def test_lvm_pvs_add_remove(self): + + self.storage.do_it() + ++ vg = self.storage.devicetree.get_device_by_name("blivetTestVG") ++ self.assertIsNotNone(vg) ++ vg_size = self._get_vg_size(vg.name) ++ self.assertEqual(vg_size, vg.size) ++ vg_free = self._get_vg_free(vg.name) ++ self.assertEqual(vg_free, vg.free_space) ++ + # create a second PV + disk2 = self.storage.devicetree.get_device_by_path(self.vdevs[1]) + self.assertIsNotNone(disk2) +@@ -366,6 +434,17 @@ def test_lvm_pvs_add_remove(self): + self.storage.do_it() + self.storage.reset() + ++ pv1 = self.storage.devicetree.get_device_by_name(pv1.name) ++ pv1_size = self._get_pv_size(pv1.path) ++ self.assertEqual(pv1.format.size, pv1_size) ++ ++ vg = self.storage.devicetree.get_device_by_name("blivetTestVG") ++ self.assertIsNotNone(vg) ++ vg_size = self._get_vg_size(vg.name) ++ self.assertEqual(vg_size, vg.size) ++ vg_free = self._get_vg_free(vg.name) ++ self.assertEqual(vg_free, vg.free_space) ++ + # add the PV to the existing VG + vg = self.storage.devicetree.get_device_by_name("blivetTestVG") + pv2 = self.storage.devicetree.get_device_by_name(pv2.name) +@@ -374,6 +453,17 @@ def test_lvm_pvs_add_remove(self): + self.storage.devicetree.actions.add(ac) + self.storage.do_it() + ++ pv2 = self.storage.devicetree.get_device_by_name(pv2.name) ++ pv2_size = self._get_pv_size(pv2.path) ++ self.assertEqual(pv2.format.size, pv2_size) ++ ++ vg = self.storage.devicetree.get_device_by_name("blivetTestVG") ++ self.assertIsNotNone(vg) ++ vg_size = self._get_vg_size(vg.name) ++ self.assertEqual(vg_size, vg.size) ++ vg_free = self._get_vg_free(vg.name) ++ self.assertEqual(vg_free, vg.free_space) ++ + self.assertEqual(pv2.format.vg_name, vg.name) + + self.storage.reset() +@@ -387,7 +477,19 @@ def test_lvm_pvs_add_remove(self): + self.storage.devicetree.actions.add(ac) + self.storage.do_it() + +- self.assertIsNone(pv1.format.vg_name) ++ pv2 = self.storage.devicetree.get_device_by_name(pv2.name) ++ pv2_size = self._get_pv_size(pv2.path) ++ self.assertEqual(pv2.format.size, pv2_size) ++ ++ vg = self.storage.devicetree.get_device_by_name("blivetTestVG") ++ self.assertIsNotNone(vg) ++ vg_size = self._get_vg_size(vg.name) ++ self.assertEqual(vg_size, vg.size) ++ vg_free = self._get_vg_free(vg.name) ++ self.assertEqual(vg_free, vg.free_space) ++ ++ self.assertIsNone(pv1.format.type) ++ + self.storage.reset() + + self.storage.reset() + +From 4dfa8d699ed1216c18d0c7effa33580a3aa56606 Mon Sep 17 00:00:00 2001 +From: Vojtech Trefny +Date: Tue, 21 Jan 2025 15:16:29 +0100 +Subject: [PATCH 6/7] Add a separate test case for LVMPV smaller than the block + device + +--- + tests/storage_tests/devices_test/lvm_test.py | 55 ++++++++++++++++++++ + 1 file changed, 55 insertions(+) + +diff --git a/tests/storage_tests/devices_test/lvm_test.py b/tests/storage_tests/devices_test/lvm_test.py +index 988201839..a1064c9c4 100644 +--- a/tests/storage_tests/devices_test/lvm_test.py ++++ b/tests/storage_tests/devices_test/lvm_test.py +@@ -475,6 +475,11 @@ def test_lvm_pvs_add_remove(self): + pv1 = self.storage.devicetree.get_device_by_name(pv1.name) + ac = blivet.deviceaction.ActionRemoveMember(vg, pv1) + self.storage.devicetree.actions.add(ac) ++ ++ # schedule also removing the lvmpv format from the PV ++ ac = blivet.deviceaction.ActionDestroyFormat(pv1) ++ self.storage.devicetree.actions.add(ac) ++ + self.storage.do_it() + + pv2 = self.storage.devicetree.get_device_by_name(pv2.name) +@@ -497,3 +502,53 @@ def test_lvm_pvs_add_remove(self): + self.assertIsNotNone(vg) + self.assertEqual(len(vg.pvs), 1) + self.assertEqual(vg.pvs[0].name, pv2.name) ++ ++ def test_lvm_pv_size(self): ++ disk = self.storage.devicetree.get_device_by_path(self.vdevs[0]) ++ self.assertIsNotNone(disk) ++ self.storage.initialize_disk(disk) ++ ++ pv = self.storage.new_partition(size=blivet.size.Size("100 MiB"), fmt_type="lvmpv", ++ parents=[disk]) ++ self.storage.create_device(pv) ++ ++ blivet.partitioning.do_partitioning(self.storage) ++ ++ self.storage.do_it() ++ self.storage.reset() ++ ++ pv = self.storage.devicetree.get_device_by_name(pv.name) ++ self.assertIsNotNone(pv) ++ ++ pv.format.update_size_info() ++ self.assertTrue(pv.format.resizable) ++ ++ ac = blivet.deviceaction.ActionResizeFormat(pv, blivet.size.Size("50 MiB")) ++ self.storage.devicetree.actions.add(ac) ++ ++ self.storage.do_it() ++ self.storage.reset() ++ ++ pv = self.storage.devicetree.get_device_by_name(pv.name) ++ self.assertIsNotNone(pv) ++ self.assertEqual(pv.format.size, blivet.size.Size("50 MiB")) ++ pv_size = self._get_pv_size(pv.path) ++ self.assertEqual(pv_size, pv.format.size) ++ ++ vg = self.storage.new_vg(name="blivetTestVG", parents=[pv]) ++ self.storage.create_device(vg) ++ ++ self.storage.do_it() ++ self.storage.reset() ++ ++ pv = self.storage.devicetree.get_device_by_name(pv.name) ++ self.assertIsNotNone(pv) ++ pv_size = self._get_pv_size(pv.path) ++ self.assertEqual(pv_size, pv.format.size) ++ ++ vg = self.storage.devicetree.get_device_by_name("blivetTestVG") ++ self.assertIsNotNone(vg) ++ vg_size = self._get_vg_size(vg.name) ++ self.assertEqual(vg_size, vg.size) ++ vg_free = self._get_vg_free(vg.name) ++ self.assertEqual(vg_free, vg.free_space) + +From 6cfa9d0df6faa79b8ab471ba34aa0b3d6f0dc338 Mon Sep 17 00:00:00 2001 +From: Vojtech Trefny +Date: Mon, 14 Apr 2025 14:54:00 +0200 +Subject: [PATCH 7/7] Fix checking PV free space when removing it from a VG + +--- + blivet/devices/lvm.py | 14 ++++++++++---- + 1 file changed, 10 insertions(+), 4 deletions(-) + +diff --git a/blivet/devices/lvm.py b/blivet/devices/lvm.py +index e3d08dbce..a03d57f97 100644 +--- a/blivet/devices/lvm.py ++++ b/blivet/devices/lvm.py +@@ -305,9 +305,15 @@ def _remove(self, member): + if lv.exists: + lv.setup() + ++ # if format was already scheduled for removal, use original_format ++ if member.format != "lvmpv": ++ fmt = member.original_format ++ else: ++ fmt = member.format ++ + # do not run pvmove on empty PVs +- member.format.update_size_info() +- if member.format.free < member.format.current_size: ++ fmt.update_size_info() ++ if fmt.free < fmt.current_size: + blockdev.lvm.pvmove(member.path) + blockdev.lvm.vgreduce(self.name, member.path) + +@@ -317,9 +323,9 @@ def _remove(self, member): + + # update LVMPV format size --> PV format has different size when in VG + try: +- member.format._size = member.format._target_size = member.format._size_info.do_task() ++ fmt._size = fmt._target_size = fmt._size_info.do_task() + except errors.PhysicalVolumeError as e: +- log.warning("Failed to obtain current size for device %s: %s", member.format.device, e) ++ log.warning("Failed to obtain current size for device %s: %s", fmt.device, e) + + def _add(self, member): + blockdev.lvm.vgextend(self.name, member.path) diff --git a/SOURCES/0036-Make-ActionDestroyFormat-optional.patch b/SOURCES/0036-Make-ActionDestroyFormat-optional.patch new file mode 100644 index 0000000..f5077f6 --- /dev/null +++ b/SOURCES/0036-Make-ActionDestroyFormat-optional.patch @@ -0,0 +1,258 @@ +From 68db0569b3508bbedf33d9ee3b69e8fc6a309b65 Mon Sep 17 00:00:00 2001 +From: Vojtech Trefny +Date: Fri, 16 May 2025 17:15:17 +0200 +Subject: [PATCH 1/4] Allow ActionDestroyFormat to be marked as optional + +When we are also planning to remove the device, failing to remove +the format is not critical so we can ignore it in these cases. + +Resolves: RHEL-8008 +Resolves: RHEL-8012 +--- + blivet/deviceaction.py | 37 +++++++++++++++++++++++-------------- + 1 file changed, 23 insertions(+), 14 deletions(-) + +diff --git a/blivet/deviceaction.py b/blivet/deviceaction.py +index fc1ca4b65..a6fc211ea 100644 +--- a/blivet/deviceaction.py ++++ b/blivet/deviceaction.py +@@ -728,12 +728,13 @@ class ActionDestroyFormat(DeviceAction): + obj = ACTION_OBJECT_FORMAT + type_desc_str = N_("destroy format") + +- def __init__(self, device): ++ def __init__(self, device, optional=False): + if device.format_immutable: + raise ValueError("this device's formatting cannot be modified") + + DeviceAction.__init__(self, device) + self.orig_format = self.device.format ++ self.optional = optional + + if not device.format.destroyable: + raise ValueError("resource to destroy this format type %s is unavailable" % device.format.type) +@@ -752,21 +753,29 @@ def execute(self, callbacks=None): + """ wipe the filesystem signature from the device """ + # remove any flag if set + super(ActionDestroyFormat, self).execute(callbacks=callbacks) +- status = self.device.status +- self.device.setup(orig=True) +- if hasattr(self.device, 'set_rw'): +- self.device.set_rw() + +- self.format.destroy() +- udev.settle() +- if isinstance(self.device, PartitionDevice) and self.device.disklabel_supported: +- if self.format.parted_flag: +- self.device.unset_flag(self.format.parted_flag) +- self.device.disk.original_format.commit_to_disk() +- udev.settle() ++ try: ++ status = self.device.status ++ self.device.setup(orig=True) ++ if hasattr(self.device, 'set_rw'): ++ self.device.set_rw() + +- if not status: +- self.device.teardown() ++ self.format.destroy() ++ udev.settle() ++ if isinstance(self.device, PartitionDevice) and self.device.disklabel_supported: ++ if self.format.parted_flag: ++ self.device.unset_flag(self.format.parted_flag) ++ self.device.disk.original_format.commit_to_disk() ++ udev.settle() ++ ++ if not status: ++ self.device.teardown() ++ except Exception as e: # pylint: disable=broad-except ++ if self.optional: ++ log.error("Ignoring error when executing optional action: Failed to destroy format on %s: %s.", ++ self.device.name, str(e)) ++ else: ++ raise + + def cancel(self): + if not self._applied: + +From fca71515094840ab1ca8821641284cfb0b687d82 Mon Sep 17 00:00:00 2001 +From: Vojtech Trefny +Date: Fri, 16 May 2025 17:28:40 +0200 +Subject: [PATCH 2/4] Make ActionDestroyFormat optional when device is also + removed + +In both destroy_device and recursive_remove we try to remove both +the device and its format. In these cases the format destroy can +be considered to be optional and we don't need to fail just +because we failed to remove the format. + +Resolves: RHEL-8008 +Resolves: RHEL-8012 +--- + blivet/blivet.py | 2 +- + blivet/devicetree.py | 4 ++-- + 2 files changed, 3 insertions(+), 3 deletions(-) + +diff --git a/blivet/blivet.py b/blivet/blivet.py +index dc066b036..2e86f5bf6 100644 +--- a/blivet/blivet.py ++++ b/blivet/blivet.py +@@ -897,7 +897,7 @@ def destroy_device(self, device): + if device.format.exists and device.format.type and \ + not device.format_immutable: + # schedule destruction of any formatting while we're at it +- self.devicetree.actions.add(ActionDestroyFormat(device)) ++ self.devicetree.actions.add(ActionDestroyFormat(device, optional=True)) + + action = ActionDestroyDevice(device) + self.devicetree.actions.add(action) +diff --git a/blivet/devicetree.py b/blivet/devicetree.py +index c6c1b4400..f94e3ca30 100644 +--- a/blivet/devicetree.py ++++ b/blivet/devicetree.py +@@ -264,7 +264,7 @@ def recursive_remove(self, device, actions=True, remove_device=True, modparent=T + if actions: + if leaf.format.exists and not leaf.protected and \ + not leaf.format_immutable: +- self.actions.add(ActionDestroyFormat(leaf)) ++ self.actions.add(ActionDestroyFormat(leaf, optional=True)) + + self.actions.add(ActionDestroyDevice(leaf)) + else: +@@ -276,7 +276,7 @@ def recursive_remove(self, device, actions=True, remove_device=True, modparent=T + + if not device.format_immutable: + if actions: +- self.actions.add(ActionDestroyFormat(device)) ++ self.actions.add(ActionDestroyFormat(device, optional=True)) + else: + device.format = None + + +From 50efc63fa3053f863d03439a507b3e0a6d7b8168 Mon Sep 17 00:00:00 2001 +From: Vojtech Trefny +Date: Mon, 19 May 2025 14:24:06 +0200 +Subject: [PATCH 3/4] tests: Add a simple test case for optional format destroy + action + +Related: RHEL-8008 +Related: RHEL-8012 +--- + tests/unit_tests/devices_test/lvm_test.py | 29 +++++++++++++++++++++++ + 1 file changed, 29 insertions(+) + +diff --git a/tests/unit_tests/devices_test/lvm_test.py b/tests/unit_tests/devices_test/lvm_test.py +index e645309fc..34c2084a8 100644 +--- a/tests/unit_tests/devices_test/lvm_test.py ++++ b/tests/unit_tests/devices_test/lvm_test.py +@@ -1160,3 +1160,32 @@ def test_vdo_compression_deduplication_change(self): + with patch("blivet.devices.lvm.blockdev.lvm") as lvm: + b.do_it() + lvm.vdo_enable_deduplication.assert_called_with(vg.name, vdopool.lvname) ++ ++ ++@patch("blivet.devices.lvm.LVMLogicalVolumeDevice._external_dependencies", new=[]) ++@patch("blivet.devices.lvm.LVMLogicalVolumeBase._external_dependencies", new=[]) ++@patch("blivet.devices.dm.DMDevice._external_dependencies", new=[]) ++class BlivetLVMOptionalDestroyTest(unittest.TestCase): ++ ++ def test_optional_format_destroy(self, *args): # pylint: disable=unused-argument ++ b = blivet.Blivet() ++ pv = StorageDevice("pv1", fmt=blivet.formats.get_format("lvmpv"), ++ size=Size("10 GiB"), exists=True) ++ vg = LVMVolumeGroupDevice("testvg", parents=[pv], exists=True) ++ lv = LVMLogicalVolumeDevice("testlv", parents=[vg], exists=True, size=Size("5 GiB"), ++ fmt=blivet.formats.get_format("xfs", exists=True)) ++ ++ for dev in (pv, vg, lv): ++ b.devicetree._add_device(dev) ++ ++ b.destroy_device(lv) ++ fmt_ac = b.devicetree.actions.find(action_type="destroy", object_type="format") ++ self.assertTrue(fmt_ac) ++ self.assertTrue(fmt_ac[0].optional) ++ ++ with patch("blivet.devices.lvm.blockdev.lvm") as lvm: ++ lvm.lvactivate.side_effect = RuntimeError() ++ try: ++ b.do_it() ++ except RuntimeError: ++ self.fail("Optional format destroy action is not optional") + +From ea913c5fa8e60cd5c2fdd8196be51c067a2a73d8 Mon Sep 17 00:00:00 2001 +From: Vojtech Trefny +Date: Tue, 20 May 2025 13:02:00 +0200 +Subject: [PATCH 4/4] tests: Add test case for removing broken thin pool + +Related: RHEL-8008 +Related: RHEL-8012 +--- + tests/storage_tests/devices_test/lvm_test.py | 52 ++++++++++++++++++++ + 1 file changed, 52 insertions(+) + +diff --git a/tests/storage_tests/devices_test/lvm_test.py b/tests/storage_tests/devices_test/lvm_test.py +index a1064c9c4..10e7354ff 100644 +--- a/tests/storage_tests/devices_test/lvm_test.py ++++ b/tests/storage_tests/devices_test/lvm_test.py +@@ -1,5 +1,7 @@ + import os + import subprocess ++import tempfile ++from unittest.mock import patch + + from ..storagetestcase import StorageTestCase + +@@ -552,3 +554,53 @@ def test_lvm_pv_size(self): + self.assertEqual(vg_size, vg.size) + vg_free = self._get_vg_free(vg.name) + self.assertEqual(vg_free, vg.free_space) ++ ++ def _break_thin_pool(self, vgname): ++ os.system("vgchange -an %s >/dev/null 2>&1" % vgname) ++ ++ # changing transaction_id for the pool prevents it from being activated ++ with tempfile.NamedTemporaryFile(prefix="blivet_test") as temp: ++ os.system("vgcfgbackup -f %s %s >/dev/null 2>&1" % (temp.name, vgname)) ++ os.system("sed -i 's/transaction_id =.*/transaction_id = 123456/' %s >/dev/null 2>&1" % temp.name) ++ os.system("vgcfgrestore -f %s %s --force >/dev/null 2>&1" % (temp.name, vgname)) ++ ++ def test_lvm_broken_thin(self): ++ disk = self.storage.devicetree.get_device_by_path(self.vdevs[0]) ++ self.assertIsNotNone(disk) ++ ++ self.storage.initialize_disk(disk) ++ ++ pv = self.storage.new_partition(size=blivet.size.Size("100 MiB"), fmt_type="lvmpv", ++ parents=[disk]) ++ self.storage.create_device(pv) ++ ++ blivet.partitioning.do_partitioning(self.storage) ++ ++ vg = self.storage.new_vg(name="blivetTestVG", parents=[pv]) ++ self.storage.create_device(vg) ++ ++ pool = self.storage.new_lv(thin_pool=True, size=blivet.size.Size("50 MiB"), ++ parents=[vg], name="blivetTestPool") ++ self.storage.create_device(pool) ++ ++ self.storage.do_it() ++ ++ # intentionally break the thin pool created above ++ self._break_thin_pool("blivetTestVG") ++ ++ self.storage.reset() ++ ++ pool = self.storage.devicetree.get_device_by_name("blivetTestVG-blivetTestPool") ++ self.assertIsNotNone(pool) ++ ++ # check that the pool cannot be activated ++ try: ++ pool.setup() ++ except Exception: # pylint: disable=broad-except ++ pass ++ else: ++ self.fail("Failed to break thinpool for tests") ++ ++ # verify that the pool can be destroyed even if it cannot be activated ++ self.storage.recursive_remove(pool) ++ self.storage.do_it() diff --git a/SOURCES/0037-Wipe-end-partition-before-creating-it-as-well-as-the-start.patch b/SOURCES/0037-Wipe-end-partition-before-creating-it-as-well-as-the-start.patch new file mode 100644 index 0000000..fb476c5 --- /dev/null +++ b/SOURCES/0037-Wipe-end-partition-before-creating-it-as-well-as-the-start.patch @@ -0,0 +1,384 @@ +From c07938143a9906bc0e06e78c818227b4c06f64ad Mon Sep 17 00:00:00 2001 +From: Vojtech Trefny +Date: Tue, 27 May 2025 15:21:23 +0200 +Subject: [PATCH 1/3] Add some basic partitioning storage tests + +This supplements the existing tests which use sparse files. These +new test cases actually run do_it() and check the result after +reset. More test cases will follow. + +Related: RHEL-76917 +--- + .../devices_test/partition_test.py | 148 ++++++++++++++++++ + 1 file changed, 148 insertions(+) + +diff --git a/tests/storage_tests/devices_test/partition_test.py b/tests/storage_tests/devices_test/partition_test.py +index 679fded6e..6ad8a8f1a 100644 +--- a/tests/storage_tests/devices_test/partition_test.py ++++ b/tests/storage_tests/devices_test/partition_test.py +@@ -11,12 +11,15 @@ + except ImportError: + from mock import patch + ++import blivet + from blivet.devices import DiskFile + from blivet.devices import PartitionDevice + from blivet.formats import get_format + from blivet.size import Size + from blivet.util import sparsetmpfile + ++from ..storagetestcase import StorageTestCase ++ + + Weighted = namedtuple("Weighted", ["fstype", "mountpoint", "true_funcs", "weight"]) + +@@ -218,3 +221,148 @@ def test_extended_min_size(self): + end_free = (extended_end - logical_end) * sector_size + self.assertEqual(extended_device.min_size, + extended_device.align_target_size(extended_device.current_size - end_free)) ++ ++ ++class PartitionTestCase(StorageTestCase): ++ ++ def setUp(self): ++ super().setUp() ++ ++ disks = [os.path.basename(vdev) for vdev in self.vdevs] ++ self.storage = blivet.Blivet() ++ self.storage.exclusive_disks = disks ++ self.storage.reset() ++ ++ # make sure only the targetcli disks are in the devicetree ++ for disk in self.storage.disks: ++ self.assertTrue(disk.path in self.vdevs) ++ self.assertIsNone(disk.format.type) ++ self.assertFalse(disk.children) ++ ++ def _clean_up(self): ++ self.storage.reset() ++ for disk in self.storage.disks: ++ if disk.path not in self.vdevs: ++ raise RuntimeError("Disk %s found in devicetree but not in disks created for tests" % disk.name) ++ self.storage.recursive_remove(disk) ++ ++ self.storage.do_it() ++ ++ def test_msdos_basic(self): ++ disk = self.storage.devicetree.get_device_by_path(self.vdevs[0]) ++ self.assertIsNotNone(disk) ++ ++ self.storage.format_device(disk, blivet.formats.get_format("disklabel", label_type="msdos")) ++ ++ for i in range(4): ++ part = self.storage.new_partition(size=Size("100 MiB"), parents=[disk], ++ primary=True) ++ self.storage.create_device(part) ++ ++ blivet.partitioning.do_partitioning(self.storage) ++ ++ self.storage.do_it() ++ self.storage.reset() ++ ++ disk = self.storage.devicetree.get_device_by_path(self.vdevs[0]) ++ self.assertIsNotNone(disk) ++ self.assertEqual(disk.format.type, "disklabel") ++ self.assertEqual(disk.format.label_type, "msdos") ++ self.assertIsNotNone(disk.format.parted_disk) ++ self.assertIsNotNone(disk.format.parted_device) ++ self.assertEqual(len(disk.format.partitions), 4) ++ self.assertEqual(len(disk.format.primary_partitions), 4) ++ self.assertEqual(len(disk.children), 4) ++ ++ for i in range(4): ++ part = self.storage.devicetree.get_device_by_path(self.vdevs[0] + str(i + 1)) ++ self.assertIsNotNone(part) ++ self.assertEqual(part.type, "partition") ++ self.assertEqual(part.disk, disk) ++ self.assertEqual(part.size, Size("100 MiB")) ++ self.assertTrue(part.is_primary) ++ self.assertFalse(part.is_extended) ++ self.assertFalse(part.is_logical) ++ self.assertIsNotNone(part.parted_partition) ++ ++ def test_msdos_extended(self): ++ disk = self.storage.devicetree.get_device_by_path(self.vdevs[0]) ++ self.assertIsNotNone(disk) ++ ++ self.storage.format_device(disk, blivet.formats.get_format("disklabel", label_type="msdos")) ++ ++ part = self.storage.new_partition(size=Size("100 MiB"), parents=[disk]) ++ self.storage.create_device(part) ++ ++ part = self.storage.new_partition(size=Size("1 GiB"), parents=[disk], ++ part_type=parted.PARTITION_EXTENDED) ++ self.storage.create_device(part) ++ ++ blivet.partitioning.do_partitioning(self.storage) ++ ++ for i in range(4): ++ part = self.storage.new_partition(size=Size("100 MiB"), parents=[disk], ++ part_type=parted.PARTITION_LOGICAL) ++ self.storage.create_device(part) ++ ++ blivet.partitioning.do_partitioning(self.storage) ++ ++ self.storage.do_it() ++ self.storage.reset() ++ ++ disk = self.storage.devicetree.get_device_by_path(self.vdevs[0]) ++ self.assertIsNotNone(disk) ++ self.assertEqual(disk.format.type, "disklabel") ++ self.assertEqual(disk.format.label_type, "msdos") ++ self.assertIsNotNone(disk.format.parted_disk) ++ self.assertIsNotNone(disk.format.parted_device) ++ self.assertEqual(len(disk.format.partitions), 6) ++ self.assertEqual(len(disk.format.primary_partitions), 1) ++ self.assertEqual(len(disk.children), 6) ++ ++ for i in range(4, 8): ++ part = self.storage.devicetree.get_device_by_path(self.vdevs[0] + str(i + 1)) ++ self.assertIsNotNone(part) ++ self.assertEqual(part.type, "partition") ++ self.assertEqual(part.disk, disk) ++ self.assertEqual(part.size, Size("100 MiB")) ++ self.assertFalse(part.is_primary) ++ self.assertFalse(part.is_extended) ++ self.assertTrue(part.is_logical) ++ self.assertIsNotNone(part.parted_partition) ++ ++ def test_gpt_basic(self): ++ disk = self.storage.devicetree.get_device_by_path(self.vdevs[0]) ++ self.assertIsNotNone(disk) ++ ++ self.storage.format_device(disk, blivet.formats.get_format("disklabel", label_type="gpt")) ++ ++ for i in range(4): ++ part = self.storage.new_partition(size=Size("100 MiB"), parents=[disk],) ++ self.storage.create_device(part) ++ ++ blivet.partitioning.do_partitioning(self.storage) ++ ++ self.storage.do_it() ++ self.storage.reset() ++ ++ disk = self.storage.devicetree.get_device_by_path(self.vdevs[0]) ++ self.assertIsNotNone(disk) ++ self.assertEqual(disk.format.type, "disklabel") ++ self.assertEqual(disk.format.label_type, "gpt") ++ self.assertIsNotNone(disk.format.parted_disk) ++ self.assertIsNotNone(disk.format.parted_device) ++ self.assertEqual(len(disk.format.partitions), 4) ++ self.assertEqual(len(disk.format.primary_partitions), 4) ++ self.assertEqual(len(disk.children), 4) ++ ++ for i in range(4): ++ part = self.storage.devicetree.get_device_by_path(self.vdevs[0] + str(i + 1)) ++ self.assertIsNotNone(part) ++ self.assertEqual(part.type, "partition") ++ self.assertEqual(part.disk, disk) ++ self.assertEqual(part.size, Size("100 MiB")) ++ self.assertTrue(part.is_primary) ++ self.assertFalse(part.is_extended) ++ self.assertFalse(part.is_logical) ++ self.assertIsNotNone(part.parted_partition) + +From 1486d2d47d9b757694a3da88ccc13d29d8bb12fd Mon Sep 17 00:00:00 2001 +From: Vojtech Trefny +Date: Tue, 27 May 2025 14:10:49 +0200 +Subject: [PATCH 2/3] Wipe end partition before creating it as well as the + start + +We are currently overwritting start of the newly created partition +with zeroes to remove any filesystem metadata that might occupy +the space. This extends this functionality to end of the partition +to remove 1.0 MD metadata that might be there. + +Resolves: RHEL-76917 +--- + blivet/devices/partition.py | 20 +++++++++++++++++++- + 1 file changed, 19 insertions(+), 1 deletion(-) + +diff --git a/blivet/devices/partition.py b/blivet/devices/partition.py +index 6ae4b8d36..1dac75a5a 100644 +--- a/blivet/devices/partition.py ++++ b/blivet/devices/partition.py +@@ -599,7 +599,7 @@ def _wipe(self): + """ Wipe the partition metadata. + + Assumes that the partition metadata is located at the start +- of the partition and occupies no more than 1 MiB. ++ and end of the partition and occupies no more than 1 MiB. + + Erases in block increments. Erases the smallest number of blocks + such that at least 1 MiB is erased or the whole partition is +@@ -632,6 +632,24 @@ def _wipe(self): + # things to settle. + udev.settle() + ++ if count >= part_len: ++ # very small partition, we wiped it completely already ++ return ++ ++ # now do the end of the partition as well (RAID 1.0 metadata) ++ end = self.parted_partition.geometry.end ++ cmd = ["dd", "if=/dev/zero", "of=%s" % device, "bs=%d" % bs, ++ "seek=%d" % (end - count), "count=%d" % count] ++ try: ++ util.run_program(cmd) ++ except OSError as e: ++ log.error(str(e)) ++ finally: ++ # If a udev device is created with the watch option, then ++ # a change uevent is synthesized and we need to wait for ++ # things to settle. ++ udev.settle() ++ + def _create(self): + """ Create the device. """ + log_method_call(self, self.name, status=self.status) + +From f0f78b801fb52425c13d0384f6867bf55839d98f Mon Sep 17 00:00:00 2001 +From: Vojtech Trefny +Date: Wed, 28 May 2025 11:01:14 +0200 +Subject: [PATCH 3/3] tests: Add tests for wiping stale metadata from new + partitions + +Related: RHEL-76917 +--- + .../devices_test/partition_test.py | 119 ++++++++++++++++++ + 1 file changed, 119 insertions(+) + +diff --git a/tests/storage_tests/devices_test/partition_test.py b/tests/storage_tests/devices_test/partition_test.py +index 6ad8a8f1a..f4be3aa4c 100644 +--- a/tests/storage_tests/devices_test/partition_test.py ++++ b/tests/storage_tests/devices_test/partition_test.py +@@ -4,6 +4,7 @@ + import os + import six + import unittest ++import blivet.deviceaction + import parted + + try: +@@ -366,3 +367,121 @@ def test_gpt_basic(self): + self.assertFalse(part.is_extended) + self.assertFalse(part.is_logical) + self.assertIsNotNone(part.parted_partition) ++ ++ def _partition_wipe_check(self): ++ part1 = self.storage.devicetree.get_device_by_path(self.vdevs[0] + "1") ++ self.assertIsNotNone(part1) ++ self.assertIsNone(part1.format.type) ++ ++ out = blivet.util.capture_output(["blkid", "-p", "-sTYPE", "-ovalue", self.vdevs[0] + "1"]) ++ self.assertEqual(out.strip(), "") ++ ++ part2 = self.storage.devicetree.get_device_by_path(self.vdevs[0] + "2") ++ self.assertIsNotNone(part2) ++ self.assertEqual(part2.format.type, "ext4") ++ ++ try: ++ part2.format.do_check() ++ except blivet.errors.FSError as e: ++ self.fail("Partition wipe corrupted filesystem on an adjacent partition: %s" % str(e)) ++ ++ out = blivet.util.capture_output(["blkid", "-p", "-sTYPE", "-ovalue", self.vdevs[0] + "2"]) ++ self.assertEqual(out.strip(), "ext4") ++ ++ def test_partition_wipe_ext(self): ++ """ Check that any stray filesystem metadata are removed before creating a partition """ ++ disk = self.storage.devicetree.get_device_by_path(self.vdevs[0]) ++ self.assertIsNotNone(disk) ++ ++ self.storage.format_device(disk, blivet.formats.get_format("disklabel", label_type="gpt")) ++ ++ # create two partitions with ext4 ++ part1 = self.storage.new_partition(size=Size("100 MiB"), parents=[disk], ++ fmt=blivet.formats.get_format("ext4")) ++ self.storage.create_device(part1) ++ ++ part2 = self.storage.new_partition(size=Size("1 MiB"), parents=[disk], grow=True, ++ fmt=blivet.formats.get_format("ext4")) ++ self.storage.create_device(part2) ++ ++ blivet.partitioning.do_partitioning(self.storage) ++ ++ self.storage.do_it() ++ self.storage.reset() ++ ++ # remove the first partition (only the partition without removing the format) ++ part1 = self.storage.devicetree.get_device_by_path(self.vdevs[0] + "1") ++ ac = blivet.deviceaction.ActionDestroyDevice(part1) ++ self.storage.devicetree.actions.add(ac) ++ ++ self.storage.do_it() ++ self.storage.reset() ++ ++ # create the first partition again (without ext4) ++ disk = self.storage.devicetree.get_device_by_path(self.vdevs[0]) ++ part1 = self.storage.new_partition(size=Size("100 MiB"), parents=[disk]) ++ self.storage.create_device(part1) ++ ++ blivet.partitioning.do_partitioning(self.storage) ++ ++ # XXX PartitionDevice._post_create calls wipefs on the partition, we want to check that ++ # the _pre_create dd wipe works so we need to skip the _post_create wipefs call ++ part1._post_create = lambda: None ++ ++ self.storage.do_it() ++ self.storage.reset() ++ ++ # make sure the ext4 signature is not present on part1 (and untouched on part2) ++ self._partition_wipe_check() ++ ++ def test_partition_wipe_mdraid(self): ++ """ Check that any stray RAID metadata are removed before creating a partition """ ++ disk = self.storage.devicetree.get_device_by_path(self.vdevs[0]) ++ self.assertIsNotNone(disk) ++ ++ self.storage.format_device(disk, blivet.formats.get_format("disklabel", label_type="gpt")) ++ ++ # create two partitions, one empty, one with ext4 ++ part1 = self.storage.new_partition(size=Size("100 MiB"), parents=[disk]) ++ self.storage.create_device(part1) ++ ++ part2 = self.storage.new_partition(size=Size("1 MiB"), parents=[disk], grow=True, ++ fmt=blivet.formats.get_format("ext4")) ++ self.storage.create_device(part2) ++ ++ blivet.partitioning.do_partitioning(self.storage) ++ ++ self.storage.do_it() ++ self.storage.reset() ++ ++ # create MD RAID with metadata 1.0 on the first partition ++ ret = blivet.util.run_program(["mdadm", "--create", "blivetMDTest", "--level=linear", ++ "--metadata=1.0", "--raid-devices=1", "--force", part1.path]) ++ self.assertEqual(ret, 0, "Failed to create RAID array for partition wipe test") ++ ret = blivet.util.run_program(["mdadm", "--stop", "/dev/md/blivetMDTest"]) ++ self.assertEqual(ret, 0, "Failed to create RAID array for partition wipe test") ++ ++ # now remove the partition without removing the array first ++ part1 = self.storage.devicetree.get_device_by_path(self.vdevs[0] + "1") ++ ac = blivet.deviceaction.ActionDestroyDevice(part1) ++ self.storage.devicetree.actions.add(ac) ++ ++ self.storage.do_it() ++ self.storage.reset() ++ ++ # create the first partition again (without format) ++ disk = self.storage.devicetree.get_device_by_path(self.vdevs[0]) ++ part1 = self.storage.new_partition(size=Size("100 MiB"), parents=[disk]) ++ self.storage.create_device(part1) ++ ++ blivet.partitioning.do_partitioning(self.storage) ++ ++ # XXX PartitionDevice._post_create calls wipefs on the partition, we want to check that ++ # the _pre_create dd wipe works so we need to skip the _post_create wipefs call ++ part1._post_create = lambda: None ++ ++ self.storage.do_it() ++ self.storage.reset() ++ ++ # make sure the mdmember signature is not present on part1 (and ext4 is untouched on part2) ++ self._partition_wipe_check() diff --git a/SOURCES/0038-Add-a-pre-wipe-fixup-function-for-LVM-logical-volume.patch b/SOURCES/0038-Add-a-pre-wipe-fixup-function-for-LVM-logical-volume.patch new file mode 100644 index 0000000..613c6aa --- /dev/null +++ b/SOURCES/0038-Add-a-pre-wipe-fixup-function-for-LVM-logical-volume.patch @@ -0,0 +1,65 @@ +From f70ee1ef08c20485f49b30fe1072a7ccafaaa2fe Mon Sep 17 00:00:00 2001 +From: Vojtech Trefny +Date: Fri, 1 Aug 2025 15:03:09 +0200 +Subject: [PATCH] Add a pre-wipe fixup function for LVM logical volumes + +LVs scheduled to be removed are always activated to remove the +format during installation. If there is a read-only LV with the +skip activation flag with MD metadata this means after activating +the LV to remove the format the MD array is auto-assembled by udev +preventing us from removing it. For this special case, we simply +stop the array before removing the format. + +Resolves: RHEL-68368 +--- + blivet/deviceaction.py | 3 +++ + blivet/devices/lvm.py | 19 +++++++++++++++++++ + 2 files changed, 22 insertions(+) + +diff --git a/blivet/deviceaction.py b/blivet/deviceaction.py +index a6fc211e..169c3a10 100644 +--- a/blivet/deviceaction.py ++++ b/blivet/deviceaction.py +@@ -760,6 +760,9 @@ class ActionDestroyFormat(DeviceAction): + if hasattr(self.device, 'set_rw'): + self.device.set_rw() + ++ if hasattr(self.device, 'pre_format_destroy'): ++ self.device.pre_format_destroy() ++ + self.format.destroy() + udev.settle() + if isinstance(self.device, PartitionDevice) and self.device.disklabel_supported: +diff --git a/blivet/devices/lvm.py b/blivet/devices/lvm.py +index a03d57f9..6ea35212 100644 +--- a/blivet/devices/lvm.py ++++ b/blivet/devices/lvm.py +@@ -2695,6 +2695,25 @@ class LVMLogicalVolumeDevice(LVMLogicalVolumeBase, LVMInternalLogicalVolumeMixin + else: + blockdev.lvm.lvactivate(self.vg.name, self._name, ignore_skip=ignore_skip_activation) + ++ def pre_format_destroy(self): ++ """ Fixup needed to run before wiping this device """ ++ if self.ignore_skip_activation > 0: ++ # the LV was not activated during the initial scan so if there is an MD array on it ++ # it will now also get activated and we need to stop it to be able to remove the LV ++ try: ++ info = blockdev.md.examine(self.path) ++ except blockdev.MDRaidError: ++ pass ++ else: ++ # give udev a bit time to activate the array so we can deactivate it again ++ time.sleep(5) ++ log.info("MD metadata found on LV with skip activation, stopping the array %s", ++ info.device) ++ try: ++ blockdev.md.deactivate(info.device) ++ except blockdev.MDRaidError as err: ++ log.info("failed to deactivate %s: %s", info.device, str(err)) ++ + @type_specific + def _pre_create(self): + LVMLogicalVolumeBase._pre_create(self) +-- +2.50.1 + diff --git a/SPECS/python-blivet.spec b/SPECS/python-blivet.spec index 77407f9..2af0814 100644 --- a/SPECS/python-blivet.spec +++ b/SPECS/python-blivet.spec @@ -23,7 +23,7 @@ Version: 3.6.0 #%%global prerelease .b2 # prerelease, if defined, should be something like .a1, .b1, .b2.dev1, or .c2 -Release: 23%{?prerelease}%{?dist} +Release: 28%{?prerelease}%{?dist} Epoch: 1 License: LGPLv2+ %global realname blivet @@ -62,6 +62,12 @@ Patch28: 0029-Align-sizes-up-for-growable-LVs.patch Patch29: 0030-mod_pass_in_stratis_test.patch Patch30: 0031-Fix_running_tests_in_FIPS_mode.patch Patch31: 0032-Set-persistent-allow-discards-flag-for-new-LUKS-devices.patch +Patch32: 0033-Do-not-remove-PVs-from-devices-file-if-disabled-or-doesnt-exist.patch +Patch33: 0034-Include-additional-information-in-PartitioningError.patch +Patch34: 0035-LVMPV-format-size-fix.patch +Patch35: 0036-Make-ActionDestroyFormat-optional.patch +Patch36: 0037-Wipe-end-partition-before-creating-it-as-well-as-the-start.patch +Patch37: 0038-Add-a-pre-wipe-fixup-function-for-LVM-logical-volume.patch # Versions of required components (done so we make sure the buildrequires # match the requires versions of things). @@ -225,6 +231,29 @@ configuration. %endif %changelog +* Mon Aug 04 2025 Vojtech Trefny - 3.6.0-28 +- Add a pre-wipe fixup function for LVM logical volumes + Resolves: RHEL-68368 + +* Fri May 30 2025 Vojtech Trefny - 3.6.0-27 +- Wipe end partition before creating it as well as the start + Resolves: RHEL-76917 + +* Tue May 20 2025 Vojtech Trefny - 3.6.0-26 +- Make ActionDestroyFormat optional when the device is also scheduled to be removed + Resolves: RHEL-8008 + Resolves: RHEL-8012 + +* Mon Apr 14 2025 Vojtech Trefny - 3.6.0-25 +- Get the actual PV format size for LVMPV format + Resolves: RHEL-74078 +- Include additional information in PartitioningError + Resolves: RHEL-8005 + +* Thu Mar 27 2025 Vojtech Trefny - 3.6.0-24 +- Do not remove PVs from devices file if disabled or doesn't exist + Resolves: RHEL-84662 + * Tue Mar 11 2025 Vojtech Trefny - 3.6.0-23 - Set persistent allow-discards flag for newly created LUKS devices Resolves: RHEL-82430