import python-blivet-3.6.0-4.el8
This commit is contained in:
parent
1e883b6d67
commit
3d132c95da
4
.gitignore
vendored
4
.gitignore
vendored
@ -1,2 +1,2 @@
|
|||||||
SOURCES/blivet-3.4.0-tests.tar.gz
|
SOURCES/blivet-3.6.0-tests.tar.gz
|
||||||
SOURCES/blivet-3.4.0.tar.gz
|
SOURCES/blivet-3.6.0.tar.gz
|
||||||
|
@ -1,2 +1,2 @@
|
|||||||
d0a86df7bbaeda7be9990b7f7b15ec36b325ec7a SOURCES/blivet-3.4.0-tests.tar.gz
|
8393baa22cb433d1012e3923ad0bc232401116c6 SOURCES/blivet-3.6.0-tests.tar.gz
|
||||||
aafc429e224dfd204cb1c284bb70de52920f7b20 SOURCES/blivet-3.4.0.tar.gz
|
e9d95c1165703fed3da1f35a9199197bfff68f98 SOURCES/blivet-3.6.0.tar.gz
|
||||||
|
@ -1,4 +1,4 @@
|
|||||||
From 2f90040ff66eacc9715e370cd49ffb72d8d1f36f Mon Sep 17 00:00:00 2001
|
From 83ccc9f9f14845fcce7a5ba5fa21fbb97b1dbbb7 Mon Sep 17 00:00:00 2001
|
||||||
From: Vojtech Trefny <vtrefny@redhat.com>
|
From: Vojtech Trefny <vtrefny@redhat.com>
|
||||||
Date: Wed, 11 Jul 2018 15:36:24 +0200
|
Date: Wed, 11 Jul 2018 15:36:24 +0200
|
||||||
Subject: [PATCH] Force command line based libblockdev LVM plugin
|
Subject: [PATCH] Force command line based libblockdev LVM plugin
|
||||||
@ -8,7 +8,7 @@ Subject: [PATCH] Force command line based libblockdev LVM plugin
|
|||||||
1 file changed, 7 insertions(+), 2 deletions(-)
|
1 file changed, 7 insertions(+), 2 deletions(-)
|
||||||
|
|
||||||
diff --git a/blivet/__init__.py b/blivet/__init__.py
|
diff --git a/blivet/__init__.py b/blivet/__init__.py
|
||||||
index c5a75bb..cb75917 100644
|
index dd8d0f54..62cc539a 100644
|
||||||
--- a/blivet/__init__.py
|
--- a/blivet/__init__.py
|
||||||
+++ b/blivet/__init__.py
|
+++ b/blivet/__init__.py
|
||||||
@@ -63,11 +63,16 @@ gi.require_version("BlockDev", "2.0")
|
@@ -63,11 +63,16 @@ gi.require_version("BlockDev", "2.0")
|
||||||
@ -31,5 +31,5 @@ index c5a75bb..cb75917 100644
|
|||||||
# do not check for dependencies during libblockdev initializtion, do runtime
|
# do not check for dependencies during libblockdev initializtion, do runtime
|
||||||
# checks instead
|
# checks instead
|
||||||
--
|
--
|
||||||
1.8.3.1
|
2.38.1
|
||||||
|
|
||||||
|
@ -1,4 +1,4 @@
|
|||||||
From 6bf3378d3d2a1b6a4338df0c4dd36a783a641633 Mon Sep 17 00:00:00 2001
|
From c098d4112635b3ea55d5bd7e1817edbd519735fc Mon Sep 17 00:00:00 2001
|
||||||
From: Vojtech Trefny <vtrefny@redhat.com>
|
From: Vojtech Trefny <vtrefny@redhat.com>
|
||||||
Date: Mon, 16 Jul 2018 14:26:11 +0200
|
Date: Mon, 16 Jul 2018 14:26:11 +0200
|
||||||
Subject: [PATCH] Remove btrfs from requested libblockdev plugins
|
Subject: [PATCH] Remove btrfs from requested libblockdev plugins
|
||||||
@ -8,7 +8,7 @@ Subject: [PATCH] Remove btrfs from requested libblockdev plugins
|
|||||||
1 file changed, 2 insertions(+), 2 deletions(-)
|
1 file changed, 2 insertions(+), 2 deletions(-)
|
||||||
|
|
||||||
diff --git a/blivet/__init__.py b/blivet/__init__.py
|
diff --git a/blivet/__init__.py b/blivet/__init__.py
|
||||||
index cb75917..09f8b1c 100644
|
index 62cc539a..bbc7ea3a 100644
|
||||||
--- a/blivet/__init__.py
|
--- a/blivet/__init__.py
|
||||||
+++ b/blivet/__init__.py
|
+++ b/blivet/__init__.py
|
||||||
@@ -63,9 +63,9 @@ gi.require_version("BlockDev", "2.0")
|
@@ -63,9 +63,9 @@ gi.require_version("BlockDev", "2.0")
|
||||||
@ -24,5 +24,5 @@ index cb75917..09f8b1c 100644
|
|||||||
_requested_plugins = blockdev.plugin_specs_from_names(_REQUESTED_PLUGIN_NAMES)
|
_requested_plugins = blockdev.plugin_specs_from_names(_REQUESTED_PLUGIN_NAMES)
|
||||||
# XXX force non-dbus LVM plugin
|
# XXX force non-dbus LVM plugin
|
||||||
--
|
--
|
||||||
1.8.3.1
|
2.38.1
|
||||||
|
|
||||||
|
@ -1,19 +1,19 @@
|
|||||||
From 3a42d9e2afdf04dbbfd2c507f5b2392193fda25b Mon Sep 17 00:00:00 2001
|
From f6f90805020d7c6ac46f17a13a00f319fc4351f6 Mon Sep 17 00:00:00 2001
|
||||||
From: Vojtech Trefny <vtrefny@redhat.com>
|
From: Vojtech Trefny <vtrefny@redhat.com>
|
||||||
Date: Wed, 26 May 2021 12:15:54 +0200
|
Date: Wed, 26 May 2021 12:15:54 +0200
|
||||||
Subject: [PATCH] Revert "More consistent lvm errors (API break)"
|
Subject: [PATCH] Revert "More consistent lvm errors (API break)"
|
||||||
|
|
||||||
This reverts commit 49ec071c6d0673224a0774d613904387c52c7381.
|
This reverts commit 49ec071c6d0673224a0774d613904387c52c7381.
|
||||||
---
|
---
|
||||||
blivet/devices/lvm.py | 72 +++++++++++++++++-----------------
|
blivet/devices/lvm.py | 72 +++++++++++------------
|
||||||
tests/devices_test/lvm_test.py | 14 +++----
|
tests/unit_tests/devices_test/lvm_test.py | 14 ++---
|
||||||
2 files changed, 43 insertions(+), 43 deletions(-)
|
2 files changed, 43 insertions(+), 43 deletions(-)
|
||||||
|
|
||||||
diff --git a/blivet/devices/lvm.py b/blivet/devices/lvm.py
|
diff --git a/blivet/devices/lvm.py b/blivet/devices/lvm.py
|
||||||
index a55515fc..6d23bfba 100644
|
index 38e49e18..b8595d63 100644
|
||||||
--- a/blivet/devices/lvm.py
|
--- a/blivet/devices/lvm.py
|
||||||
+++ b/blivet/devices/lvm.py
|
+++ b/blivet/devices/lvm.py
|
||||||
@@ -307,7 +307,7 @@ class LVMVolumeGroupDevice(ContainerDevice):
|
@@ -304,7 +304,7 @@ class LVMVolumeGroupDevice(ContainerDevice):
|
||||||
def _add_log_vol(self, lv):
|
def _add_log_vol(self, lv):
|
||||||
""" Add an LV to this VG. """
|
""" Add an LV to this VG. """
|
||||||
if lv in self._lvs:
|
if lv in self._lvs:
|
||||||
@ -22,7 +22,7 @@ index a55515fc..6d23bfba 100644
|
|||||||
|
|
||||||
# verify we have the space, then add it
|
# verify we have the space, then add it
|
||||||
# do not verify for growing vg (because of ks)
|
# do not verify for growing vg (because of ks)
|
||||||
@@ -340,7 +340,7 @@ class LVMVolumeGroupDevice(ContainerDevice):
|
@@ -337,7 +337,7 @@ class LVMVolumeGroupDevice(ContainerDevice):
|
||||||
def _remove_log_vol(self, lv):
|
def _remove_log_vol(self, lv):
|
||||||
""" Remove an LV from this VG. """
|
""" Remove an LV from this VG. """
|
||||||
if lv not in self.lvs:
|
if lv not in self.lvs:
|
||||||
@ -31,7 +31,7 @@ index a55515fc..6d23bfba 100644
|
|||||||
|
|
||||||
self._lvs.remove(lv)
|
self._lvs.remove(lv)
|
||||||
|
|
||||||
@@ -415,7 +415,7 @@ class LVMVolumeGroupDevice(ContainerDevice):
|
@@ -430,7 +430,7 @@ class LVMVolumeGroupDevice(ContainerDevice):
|
||||||
@thpool_reserve.setter
|
@thpool_reserve.setter
|
||||||
def thpool_reserve(self, value):
|
def thpool_reserve(self, value):
|
||||||
if value is not None and not isinstance(value, ThPoolReserveSpec):
|
if value is not None and not isinstance(value, ThPoolReserveSpec):
|
||||||
@ -40,8 +40,8 @@ index a55515fc..6d23bfba 100644
|
|||||||
self._thpool_reserve = value
|
self._thpool_reserve = value
|
||||||
|
|
||||||
@property
|
@property
|
||||||
@@ -646,14 +646,14 @@ class LVMLogicalVolumeBase(DMDevice, RaidDevice):
|
@@ -665,14 +665,14 @@ class LVMLogicalVolumeBase(DMDevice, RaidDevice):
|
||||||
if seg_type not in [None, "linear", "thin", "thin-pool", "cache", "vdo-pool", "vdo"] + lvm.raid_seg_types:
|
if seg_type not in [None, "linear", "thin", "thin-pool", "cache", "vdo-pool", "vdo", "cache-pool"] + lvm.raid_seg_types:
|
||||||
raise ValueError("Invalid or unsupported segment type: %s" % seg_type)
|
raise ValueError("Invalid or unsupported segment type: %s" % seg_type)
|
||||||
if seg_type and seg_type in lvm.raid_seg_types and not pvs:
|
if seg_type and seg_type in lvm.raid_seg_types and not pvs:
|
||||||
- raise errors.DeviceError("List of PVs has to be given for every non-linear LV")
|
- raise errors.DeviceError("List of PVs has to be given for every non-linear LV")
|
||||||
@ -60,7 +60,7 @@ index a55515fc..6d23bfba 100644
|
|||||||
|
|
||||||
# When this device's format is set in the superclass constructor it will
|
# When this device's format is set in the superclass constructor it will
|
||||||
# try to access self.snapshots.
|
# try to access self.snapshots.
|
||||||
@@ -702,13 +702,13 @@ class LVMLogicalVolumeBase(DMDevice, RaidDevice):
|
@@ -721,13 +721,13 @@ class LVMLogicalVolumeBase(DMDevice, RaidDevice):
|
||||||
self._from_lvs = from_lvs
|
self._from_lvs = from_lvs
|
||||||
if self._from_lvs:
|
if self._from_lvs:
|
||||||
if exists:
|
if exists:
|
||||||
@ -78,7 +78,7 @@ index a55515fc..6d23bfba 100644
|
|||||||
|
|
||||||
self._cache = None
|
self._cache = None
|
||||||
if cache_request and not self.exists:
|
if cache_request and not self.exists:
|
||||||
@@ -723,13 +723,13 @@ class LVMLogicalVolumeBase(DMDevice, RaidDevice):
|
@@ -746,13 +746,13 @@ class LVMLogicalVolumeBase(DMDevice, RaidDevice):
|
||||||
elif isinstance(pv_spec, StorageDevice):
|
elif isinstance(pv_spec, StorageDevice):
|
||||||
self._pv_specs.append(LVPVSpec(pv_spec, Size(0)))
|
self._pv_specs.append(LVPVSpec(pv_spec, Size(0)))
|
||||||
else:
|
else:
|
||||||
@ -94,7 +94,7 @@ index a55515fc..6d23bfba 100644
|
|||||||
if self._pv_specs:
|
if self._pv_specs:
|
||||||
self._assign_pv_space()
|
self._assign_pv_space()
|
||||||
|
|
||||||
@@ -1072,7 +1072,7 @@ class LVMLogicalVolumeBase(DMDevice, RaidDevice):
|
@@ -1130,7 +1130,7 @@ class LVMLogicalVolumeBase(DMDevice, RaidDevice):
|
||||||
else:
|
else:
|
||||||
msg = "the specified internal LV '%s' doesn't belong to this LV ('%s')" % (int_lv.lv_name,
|
msg = "the specified internal LV '%s' doesn't belong to this LV ('%s')" % (int_lv.lv_name,
|
||||||
self.name)
|
self.name)
|
||||||
@ -103,7 +103,7 @@ index a55515fc..6d23bfba 100644
|
|||||||
|
|
||||||
def populate_ksdata(self, data):
|
def populate_ksdata(self, data):
|
||||||
super(LVMLogicalVolumeBase, self).populate_ksdata(data)
|
super(LVMLogicalVolumeBase, self).populate_ksdata(data)
|
||||||
@@ -1171,7 +1171,7 @@ class LVMInternalLogicalVolumeMixin(object):
|
@@ -1229,7 +1229,7 @@ class LVMInternalLogicalVolumeMixin(object):
|
||||||
def _init_check(self):
|
def _init_check(self):
|
||||||
# an internal LV should have no parents
|
# an internal LV should have no parents
|
||||||
if self._parent_lv and self._parents:
|
if self._parent_lv and self._parents:
|
||||||
@ -112,7 +112,7 @@ index a55515fc..6d23bfba 100644
|
|||||||
|
|
||||||
@property
|
@property
|
||||||
def is_internal_lv(self):
|
def is_internal_lv(self):
|
||||||
@@ -1231,7 +1231,7 @@ class LVMInternalLogicalVolumeMixin(object):
|
@@ -1289,7 +1289,7 @@ class LVMInternalLogicalVolumeMixin(object):
|
||||||
|
|
||||||
@readonly.setter
|
@readonly.setter
|
||||||
def readonly(self, value): # pylint: disable=unused-argument
|
def readonly(self, value): # pylint: disable=unused-argument
|
||||||
@ -121,7 +121,7 @@ index a55515fc..6d23bfba 100644
|
|||||||
|
|
||||||
@property
|
@property
|
||||||
def type(self):
|
def type(self):
|
||||||
@@ -1267,7 +1267,7 @@ class LVMInternalLogicalVolumeMixin(object):
|
@@ -1325,7 +1325,7 @@ class LVMInternalLogicalVolumeMixin(object):
|
||||||
def _check_parents(self):
|
def _check_parents(self):
|
||||||
# an internal LV should have no parents
|
# an internal LV should have no parents
|
||||||
if self._parents:
|
if self._parents:
|
||||||
@ -130,7 +130,7 @@ index a55515fc..6d23bfba 100644
|
|||||||
|
|
||||||
def _add_to_parents(self):
|
def _add_to_parents(self):
|
||||||
# nothing to do here, an internal LV has no parents (in the DeviceTree's
|
# nothing to do here, an internal LV has no parents (in the DeviceTree's
|
||||||
@@ -1277,13 +1277,13 @@ class LVMInternalLogicalVolumeMixin(object):
|
@@ -1335,13 +1335,13 @@ class LVMInternalLogicalVolumeMixin(object):
|
||||||
# internal LVs follow different rules limitting size
|
# internal LVs follow different rules limitting size
|
||||||
def _set_size(self, newsize):
|
def _set_size(self, newsize):
|
||||||
if not isinstance(newsize, Size):
|
if not isinstance(newsize, Size):
|
||||||
@ -146,7 +146,7 @@ index a55515fc..6d23bfba 100644
|
|||||||
else:
|
else:
|
||||||
# same rules apply as for any other LV
|
# same rules apply as for any other LV
|
||||||
raise NotTypeSpecific()
|
raise NotTypeSpecific()
|
||||||
@@ -1361,18 +1361,18 @@ class LVMSnapshotMixin(object):
|
@@ -1419,18 +1419,18 @@ class LVMSnapshotMixin(object):
|
||||||
return
|
return
|
||||||
|
|
||||||
if self.origin and not isinstance(self.origin, LVMLogicalVolumeDevice):
|
if self.origin and not isinstance(self.origin, LVMLogicalVolumeDevice):
|
||||||
@ -169,7 +169,7 @@ index a55515fc..6d23bfba 100644
|
|||||||
|
|
||||||
@property
|
@property
|
||||||
def is_snapshot_lv(self):
|
def is_snapshot_lv(self):
|
||||||
@@ -1544,7 +1544,7 @@ class LVMThinPoolMixin(object):
|
@@ -1606,7 +1606,7 @@ class LVMThinPoolMixin(object):
|
||||||
def _check_from_lvs(self):
|
def _check_from_lvs(self):
|
||||||
if self._from_lvs:
|
if self._from_lvs:
|
||||||
if len(self._from_lvs) != 2:
|
if len(self._from_lvs) != 2:
|
||||||
@ -178,7 +178,7 @@ index a55515fc..6d23bfba 100644
|
|||||||
|
|
||||||
def _convert_from_lvs(self):
|
def _convert_from_lvs(self):
|
||||||
data_lv, metadata_lv = self._from_lvs
|
data_lv, metadata_lv = self._from_lvs
|
||||||
@@ -1590,7 +1590,7 @@ class LVMThinPoolMixin(object):
|
@@ -1652,7 +1652,7 @@ class LVMThinPoolMixin(object):
|
||||||
def _add_log_vol(self, lv):
|
def _add_log_vol(self, lv):
|
||||||
""" Add an LV to this pool. """
|
""" Add an LV to this pool. """
|
||||||
if lv in self._lvs:
|
if lv in self._lvs:
|
||||||
@ -187,7 +187,7 @@ index a55515fc..6d23bfba 100644
|
|||||||
|
|
||||||
# TODO: add some checking to prevent overcommit for preexisting
|
# TODO: add some checking to prevent overcommit for preexisting
|
||||||
self.vg._add_log_vol(lv)
|
self.vg._add_log_vol(lv)
|
||||||
@@ -1601,7 +1601,7 @@ class LVMThinPoolMixin(object):
|
@@ -1663,7 +1663,7 @@ class LVMThinPoolMixin(object):
|
||||||
def _remove_log_vol(self, lv):
|
def _remove_log_vol(self, lv):
|
||||||
""" Remove an LV from this pool. """
|
""" Remove an LV from this pool. """
|
||||||
if lv not in self._lvs:
|
if lv not in self._lvs:
|
||||||
@ -196,7 +196,7 @@ index a55515fc..6d23bfba 100644
|
|||||||
|
|
||||||
self._lvs.remove(lv)
|
self._lvs.remove(lv)
|
||||||
self.vg._remove_log_vol(lv)
|
self.vg._remove_log_vol(lv)
|
||||||
@@ -1711,14 +1711,14 @@ class LVMThinLogicalVolumeMixin(object):
|
@@ -1772,14 +1772,14 @@ class LVMThinLogicalVolumeMixin(object):
|
||||||
"""Check that this device has parents as expected"""
|
"""Check that this device has parents as expected"""
|
||||||
if isinstance(self.parents, (list, ParentList)):
|
if isinstance(self.parents, (list, ParentList)):
|
||||||
if len(self.parents) != 1:
|
if len(self.parents) != 1:
|
||||||
@ -213,7 +213,7 @@ index a55515fc..6d23bfba 100644
|
|||||||
|
|
||||||
@property
|
@property
|
||||||
def is_thin_lv(self):
|
def is_thin_lv(self):
|
||||||
@@ -1755,7 +1755,7 @@ class LVMThinLogicalVolumeMixin(object):
|
@@ -1816,7 +1816,7 @@ class LVMThinLogicalVolumeMixin(object):
|
||||||
|
|
||||||
def _set_size(self, newsize):
|
def _set_size(self, newsize):
|
||||||
if not isinstance(newsize, Size):
|
if not isinstance(newsize, Size):
|
||||||
@ -222,7 +222,7 @@ index a55515fc..6d23bfba 100644
|
|||||||
|
|
||||||
newsize = self.vg.align(newsize)
|
newsize = self.vg.align(newsize)
|
||||||
newsize = self.vg.align(util.numeric_type(newsize))
|
newsize = self.vg.align(util.numeric_type(newsize))
|
||||||
@@ -2229,7 +2229,7 @@ class LVMLogicalVolumeDevice(LVMLogicalVolumeBase, LVMInternalLogicalVolumeMixin
|
@@ -2499,7 +2499,7 @@ class LVMLogicalVolumeDevice(LVMLogicalVolumeBase, LVMInternalLogicalVolumeMixin
|
||||||
container = self.parents
|
container = self.parents
|
||||||
|
|
||||||
if not isinstance(container, LVMVolumeGroupDevice):
|
if not isinstance(container, LVMVolumeGroupDevice):
|
||||||
@ -231,7 +231,7 @@ index a55515fc..6d23bfba 100644
|
|||||||
|
|
||||||
@type_specific
|
@type_specific
|
||||||
def _add_to_parents(self):
|
def _add_to_parents(self):
|
||||||
@@ -2240,12 +2240,12 @@ class LVMLogicalVolumeDevice(LVMLogicalVolumeBase, LVMInternalLogicalVolumeMixin
|
@@ -2510,12 +2510,12 @@ class LVMLogicalVolumeDevice(LVMLogicalVolumeBase, LVMInternalLogicalVolumeMixin
|
||||||
@type_specific
|
@type_specific
|
||||||
def _check_from_lvs(self):
|
def _check_from_lvs(self):
|
||||||
"""Check the LVs to create this LV from"""
|
"""Check the LVs to create this LV from"""
|
||||||
@ -246,7 +246,7 @@ index a55515fc..6d23bfba 100644
|
|||||||
|
|
||||||
@property
|
@property
|
||||||
def external_dependencies(self):
|
def external_dependencies(self):
|
||||||
@@ -2265,7 +2265,7 @@ class LVMLogicalVolumeDevice(LVMLogicalVolumeBase, LVMInternalLogicalVolumeMixin
|
@@ -2535,7 +2535,7 @@ class LVMLogicalVolumeDevice(LVMLogicalVolumeBase, LVMInternalLogicalVolumeMixin
|
||||||
@type_specific
|
@type_specific
|
||||||
def _set_size(self, newsize):
|
def _set_size(self, newsize):
|
||||||
if not isinstance(newsize, Size):
|
if not isinstance(newsize, Size):
|
||||||
@ -255,7 +255,7 @@ index a55515fc..6d23bfba 100644
|
|||||||
|
|
||||||
newsize = self.vg.align(newsize)
|
newsize = self.vg.align(newsize)
|
||||||
log.debug("trying to set lv %s size to %s", self.name, newsize)
|
log.debug("trying to set lv %s size to %s", self.name, newsize)
|
||||||
@@ -2274,7 +2274,7 @@ class LVMLogicalVolumeDevice(LVMLogicalVolumeBase, LVMInternalLogicalVolumeMixin
|
@@ -2544,7 +2544,7 @@ class LVMLogicalVolumeDevice(LVMLogicalVolumeBase, LVMInternalLogicalVolumeMixin
|
||||||
# space for it. A similar reasoning applies to shrinking the LV.
|
# space for it. A similar reasoning applies to shrinking the LV.
|
||||||
if not self.exists and newsize > self.size and newsize > self.vg.free_space + self.vg_space_used:
|
if not self.exists and newsize > self.size and newsize > self.vg.free_space + self.vg_space_used:
|
||||||
log.error("failed to set size: %s short", newsize - (self.vg.free_space + self.vg_space_used))
|
log.error("failed to set size: %s short", newsize - (self.vg.free_space + self.vg_space_used))
|
||||||
@ -264,7 +264,7 @@ index a55515fc..6d23bfba 100644
|
|||||||
|
|
||||||
LVMLogicalVolumeBase._set_size(self, newsize)
|
LVMLogicalVolumeBase._set_size(self, newsize)
|
||||||
|
|
||||||
@@ -2622,7 +2622,7 @@ class LVMCache(Cache):
|
@@ -2910,7 +2910,7 @@ class LVMCache(Cache):
|
||||||
spec.size = spec.pv.format.free
|
spec.size = spec.pv.format.free
|
||||||
space_to_assign -= spec.pv.format.free
|
space_to_assign -= spec.pv.format.free
|
||||||
if space_to_assign > 0:
|
if space_to_assign > 0:
|
||||||
@ -273,11 +273,11 @@ index a55515fc..6d23bfba 100644
|
|||||||
|
|
||||||
@property
|
@property
|
||||||
def size(self):
|
def size(self):
|
||||||
diff --git a/tests/devices_test/lvm_test.py b/tests/devices_test/lvm_test.py
|
diff --git a/tests/unit_tests/devices_test/lvm_test.py b/tests/unit_tests/devices_test/lvm_test.py
|
||||||
index 670c91c9..4156d0bf 100644
|
index 47613fdc..995c2da4 100644
|
||||||
--- a/tests/devices_test/lvm_test.py
|
--- a/tests/unit_tests/devices_test/lvm_test.py
|
||||||
+++ b/tests/devices_test/lvm_test.py
|
+++ b/tests/unit_tests/devices_test/lvm_test.py
|
||||||
@@ -36,10 +36,10 @@ class LVMDeviceTest(unittest.TestCase):
|
@@ -32,10 +32,10 @@ class LVMDeviceTest(unittest.TestCase):
|
||||||
lv = LVMLogicalVolumeDevice("testlv", parents=[vg],
|
lv = LVMLogicalVolumeDevice("testlv", parents=[vg],
|
||||||
fmt=blivet.formats.get_format("xfs"))
|
fmt=blivet.formats.get_format("xfs"))
|
||||||
|
|
||||||
@ -290,7 +290,7 @@ index 670c91c9..4156d0bf 100644
|
|||||||
LVMLogicalVolumeDevice("snap1", parents=[vg], vorigin=True)
|
LVMLogicalVolumeDevice("snap1", parents=[vg], vorigin=True)
|
||||||
|
|
||||||
lv.exists = True
|
lv.exists = True
|
||||||
@@ -64,7 +64,7 @@ class LVMDeviceTest(unittest.TestCase):
|
@@ -60,7 +60,7 @@ class LVMDeviceTest(unittest.TestCase):
|
||||||
pool = LVMLogicalVolumeDevice("pool1", parents=[vg], size=Size("500 MiB"), seg_type="thin-pool")
|
pool = LVMLogicalVolumeDevice("pool1", parents=[vg], size=Size("500 MiB"), seg_type="thin-pool")
|
||||||
thinlv = LVMLogicalVolumeDevice("thinlv", parents=[pool], size=Size("200 MiB"), seg_type="thin")
|
thinlv = LVMLogicalVolumeDevice("thinlv", parents=[pool], size=Size("200 MiB"), seg_type="thin")
|
||||||
|
|
||||||
@ -299,7 +299,7 @@ index 670c91c9..4156d0bf 100644
|
|||||||
LVMLogicalVolumeDevice("snap1", parents=[pool], origin=pv, seg_type="thin")
|
LVMLogicalVolumeDevice("snap1", parents=[pool], origin=pv, seg_type="thin")
|
||||||
|
|
||||||
# now make the constructor succeed so we can test some properties
|
# now make the constructor succeed so we can test some properties
|
||||||
@@ -258,21 +258,21 @@ class LVMDeviceTest(unittest.TestCase):
|
@@ -310,21 +310,21 @@ class LVMDeviceTest(unittest.TestCase):
|
||||||
vg = LVMVolumeGroupDevice("testvg", parents=[pv, pv2])
|
vg = LVMVolumeGroupDevice("testvg", parents=[pv, pv2])
|
||||||
|
|
||||||
# pvs have to be specified for non-linear LVs
|
# pvs have to be specified for non-linear LVs
|
||||||
@ -326,5 +326,5 @@ index 670c91c9..4156d0bf 100644
|
|||||||
pv_spec2 = LVPVSpec(pv2, Size("250 MiB"))
|
pv_spec2 = LVPVSpec(pv2, Size("250 MiB"))
|
||||||
lv = LVMLogicalVolumeDevice("testlv", parents=[vg], size=Size("512 MiB"),
|
lv = LVMLogicalVolumeDevice("testlv", parents=[vg], size=Size("512 MiB"),
|
||||||
--
|
--
|
||||||
2.31.1
|
2.38.1
|
||||||
|
|
||||||
|
@ -1,4 +1,4 @@
|
|||||||
From 42042e7fb6177d3cfe5568e358a38278925a2624 Mon Sep 17 00:00:00 2001
|
From d8a8d96450bf0d3458671b9b7d23d972aa540396 Mon Sep 17 00:00:00 2001
|
||||||
From: Vojtech Trefny <vtrefny@redhat.com>
|
From: Vojtech Trefny <vtrefny@redhat.com>
|
||||||
Date: Wed, 26 May 2021 12:27:34 +0200
|
Date: Wed, 26 May 2021 12:27:34 +0200
|
||||||
Subject: [PATCH] Revert "Terminology cleanups"
|
Subject: [PATCH] Revert "Terminology cleanups"
|
||||||
@ -8,32 +8,32 @@ This reverts following commits:
|
|||||||
- 63c9c7165e5cdfa4a47dcf0ed9d717b71e7921f2
|
- 63c9c7165e5cdfa4a47dcf0ed9d717b71e7921f2
|
||||||
- 8956b9af8a785ae25e0e7153d2ef0702ce2f567c
|
- 8956b9af8a785ae25e0e7153d2ef0702ce2f567c
|
||||||
---
|
---
|
||||||
blivet/devicefactory.py | 24 +++++-----
|
blivet/devicefactory.py | 24 +++----
|
||||||
blivet/devices/dm.py | 9 +++-
|
blivet/devices/dm.py | 9 ++-
|
||||||
blivet/devices/loop.py | 20 ++++----
|
blivet/devices/loop.py | 20 +++---
|
||||||
blivet/devices/luks.py | 26 +++++-----
|
blivet/devices/luks.py | 26 ++++---
|
||||||
blivet/errors.py | 2 +-
|
blivet/errors.py | 2 +-
|
||||||
blivet/partitioning.py | 22 +++++++--
|
blivet/partitioning.py | 22 +++++-
|
||||||
blivet/populator/helpers/dm.py | 4 +-
|
blivet/populator/helpers/dm.py | 4 +-
|
||||||
blivet/populator/helpers/luks.py | 4 +-
|
blivet/populator/helpers/luks.py | 4 +-
|
||||||
blivet/populator/helpers/lvm.py | 2 +-
|
blivet/populator/helpers/lvm.py | 2 +-
|
||||||
blivet/populator/helpers/mdraid.py | 14 +++---
|
blivet/populator/helpers/mdraid.py | 14 ++--
|
||||||
blivet/populator/helpers/multipath.py | 8 ++--
|
blivet/populator/helpers/multipath.py | 8 +--
|
||||||
blivet/populator/populator.py | 69 ++++++++++++++-------------
|
blivet/populator/populator.py | 67 ++++++++++---------
|
||||||
blivet/threads.py | 3 +-
|
blivet/threads.py | 3 +-
|
||||||
blivet/udev.py | 34 ++++++-------
|
blivet/udev.py | 34 +++++-----
|
||||||
tests/devicefactory_test.py | 10 ++--
|
tests/unit_tests/devicefactory_test.py | 10 +--
|
||||||
tests/devices_test/size_test.py | 6 +--
|
.../devices_test/device_size_test.py | 6 +-
|
||||||
tests/populator_test.py | 34 ++++++-------
|
tests/unit_tests/populator_test.py | 34 +++++-----
|
||||||
tests/udev_test.py | 12 ++---
|
tests/unit_tests/udev_test.py | 12 ++--
|
||||||
tests/vmtests/vmbackedtestcase.py | 2 +-
|
tests/vmtests/vmbackedtestcase.py | 2 +-
|
||||||
19 files changed, 168 insertions(+), 137 deletions(-)
|
19 files changed, 167 insertions(+), 136 deletions(-)
|
||||||
|
|
||||||
diff --git a/blivet/devicefactory.py b/blivet/devicefactory.py
|
diff --git a/blivet/devicefactory.py b/blivet/devicefactory.py
|
||||||
index 0f7fdfa1..f56bd9a3 100644
|
index 6f460f6d..90082c28 100644
|
||||||
--- a/blivet/devicefactory.py
|
--- a/blivet/devicefactory.py
|
||||||
+++ b/blivet/devicefactory.py
|
+++ b/blivet/devicefactory.py
|
||||||
@@ -849,12 +849,12 @@ class DeviceFactory(object):
|
@@ -859,12 +859,12 @@ class DeviceFactory(object):
|
||||||
parent_container.parents.remove(orig_device)
|
parent_container.parents.remove(orig_device)
|
||||||
|
|
||||||
if self.encrypted and isinstance(self.device, LUKSDevice) and \
|
if self.encrypted and isinstance(self.device, LUKSDevice) and \
|
||||||
@ -50,7 +50,7 @@ index 0f7fdfa1..f56bd9a3 100644
|
|||||||
|
|
||||||
def _set_name(self):
|
def _set_name(self):
|
||||||
if not self.device_name:
|
if not self.device_name:
|
||||||
@@ -1173,11 +1173,11 @@ class PartitionSetFactory(PartitionFactory):
|
@@ -1201,11 +1201,11 @@ class PartitionSetFactory(PartitionFactory):
|
||||||
container.parents.remove(member)
|
container.parents.remove(member)
|
||||||
self.storage.destroy_device(member)
|
self.storage.destroy_device(member)
|
||||||
members.remove(member)
|
members.remove(member)
|
||||||
@ -65,7 +65,7 @@ index 0f7fdfa1..f56bd9a3 100644
|
|||||||
|
|
||||||
continue
|
continue
|
||||||
|
|
||||||
@@ -1199,10 +1199,10 @@ class PartitionSetFactory(PartitionFactory):
|
@@ -1227,10 +1227,10 @@ class PartitionSetFactory(PartitionFactory):
|
||||||
|
|
||||||
continue
|
continue
|
||||||
|
|
||||||
@ -80,7 +80,7 @@ index 0f7fdfa1..f56bd9a3 100644
|
|||||||
|
|
||||||
##
|
##
|
||||||
# Prepare previously allocated member partitions for reallocation.
|
# Prepare previously allocated member partitions for reallocation.
|
||||||
@@ -1262,7 +1262,7 @@ class PartitionSetFactory(PartitionFactory):
|
@@ -1290,7 +1290,7 @@ class PartitionSetFactory(PartitionFactory):
|
||||||
|
|
||||||
if isinstance(member, LUKSDevice):
|
if isinstance(member, LUKSDevice):
|
||||||
self.storage.destroy_device(member)
|
self.storage.destroy_device(member)
|
||||||
@ -90,7 +90,7 @@ index 0f7fdfa1..f56bd9a3 100644
|
|||||||
self.storage.destroy_device(member)
|
self.storage.destroy_device(member)
|
||||||
|
|
||||||
diff --git a/blivet/devices/dm.py b/blivet/devices/dm.py
|
diff --git a/blivet/devices/dm.py b/blivet/devices/dm.py
|
||||||
index 3529f61c..508a6f89 100644
|
index 2f936170..ae25e8e6 100644
|
||||||
--- a/blivet/devices/dm.py
|
--- a/blivet/devices/dm.py
|
||||||
+++ b/blivet/devices/dm.py
|
+++ b/blivet/devices/dm.py
|
||||||
@@ -154,6 +154,11 @@ class DMDevice(StorageDevice):
|
@@ -154,6 +154,11 @@ class DMDevice(StorageDevice):
|
||||||
@ -178,7 +178,7 @@ index 0f4d7775..78f88d7d 100644
|
|||||||
+ def slave(self):
|
+ def slave(self):
|
||||||
+ return self.parents[0]
|
+ return self.parents[0]
|
||||||
diff --git a/blivet/devices/luks.py b/blivet/devices/luks.py
|
diff --git a/blivet/devices/luks.py b/blivet/devices/luks.py
|
||||||
index 5d6d6c65..555f1acd 100644
|
index 2eb1f130..5ab840ea 100644
|
||||||
--- a/blivet/devices/luks.py
|
--- a/blivet/devices/luks.py
|
||||||
+++ b/blivet/devices/luks.py
|
+++ b/blivet/devices/luks.py
|
||||||
@@ -66,13 +66,17 @@ class LUKSDevice(DMCryptDevice):
|
@@ -66,13 +66,17 @@ class LUKSDevice(DMCryptDevice):
|
||||||
@ -270,10 +270,10 @@ index 5d6d6c65..555f1acd 100644
|
|||||||
super(LUKSDevice, self).populate_ksdata(data)
|
super(LUKSDevice, self).populate_ksdata(data)
|
||||||
|
|
||||||
diff --git a/blivet/errors.py b/blivet/errors.py
|
diff --git a/blivet/errors.py b/blivet/errors.py
|
||||||
index fd51283f..f6bf853a 100644
|
index b886ffec..30c9921a 100644
|
||||||
--- a/blivet/errors.py
|
--- a/blivet/errors.py
|
||||||
+++ b/blivet/errors.py
|
+++ b/blivet/errors.py
|
||||||
@@ -192,7 +192,7 @@ class DeviceTreeError(StorageError):
|
@@ -201,7 +201,7 @@ class DeviceTreeError(StorageError):
|
||||||
pass
|
pass
|
||||||
|
|
||||||
|
|
||||||
@ -283,7 +283,7 @@ index fd51283f..f6bf853a 100644
|
|||||||
|
|
||||||
|
|
||||||
diff --git a/blivet/partitioning.py b/blivet/partitioning.py
|
diff --git a/blivet/partitioning.py b/blivet/partitioning.py
|
||||||
index 53f9cc3f..ca0a55d1 100644
|
index ce77e4eb..2cd6554c 100644
|
||||||
--- a/blivet/partitioning.py
|
--- a/blivet/partitioning.py
|
||||||
+++ b/blivet/partitioning.py
|
+++ b/blivet/partitioning.py
|
||||||
@@ -32,7 +32,7 @@ import _ped
|
@@ -32,7 +32,7 @@ import _ped
|
||||||
@ -295,7 +295,7 @@ index 53f9cc3f..ca0a55d1 100644
|
|||||||
from .size import Size
|
from .size import Size
|
||||||
from .i18n import _
|
from .i18n import _
|
||||||
from .util import stringize, unicodeize, compare
|
from .util import stringize, unicodeize, compare
|
||||||
@@ -1632,7 +1632,15 @@ class TotalSizeSet(object):
|
@@ -1635,7 +1635,15 @@ class TotalSizeSet(object):
|
||||||
:param size: the target combined size
|
:param size: the target combined size
|
||||||
:type size: :class:`~.size.Size`
|
:type size: :class:`~.size.Size`
|
||||||
"""
|
"""
|
||||||
@ -312,7 +312,7 @@ index 53f9cc3f..ca0a55d1 100644
|
|||||||
self.size = size
|
self.size = size
|
||||||
|
|
||||||
self.requests = []
|
self.requests = []
|
||||||
@@ -1670,7 +1678,15 @@ class SameSizeSet(object):
|
@@ -1673,7 +1681,15 @@ class SameSizeSet(object):
|
||||||
:keyword max_size: the maximum size for growable devices
|
:keyword max_size: the maximum size for growable devices
|
||||||
:type max_size: :class:`~.size.Size`
|
:type max_size: :class:`~.size.Size`
|
||||||
"""
|
"""
|
||||||
@ -330,10 +330,10 @@ index 53f9cc3f..ca0a55d1 100644
|
|||||||
self.grow = grow
|
self.grow = grow
|
||||||
self.max_size = max_size
|
self.max_size = max_size
|
||||||
diff --git a/blivet/populator/helpers/dm.py b/blivet/populator/helpers/dm.py
|
diff --git a/blivet/populator/helpers/dm.py b/blivet/populator/helpers/dm.py
|
||||||
index 30e99aa1..770736b0 100644
|
index 4721390e..0ad065e2 100644
|
||||||
--- a/blivet/populator/helpers/dm.py
|
--- a/blivet/populator/helpers/dm.py
|
||||||
+++ b/blivet/populator/helpers/dm.py
|
+++ b/blivet/populator/helpers/dm.py
|
||||||
@@ -46,13 +46,13 @@ class DMDevicePopulator(DevicePopulator):
|
@@ -47,13 +47,13 @@ class DMDevicePopulator(DevicePopulator):
|
||||||
name = udev.device_get_name(self.data)
|
name = udev.device_get_name(self.data)
|
||||||
log_method_call(self, name=name)
|
log_method_call(self, name=name)
|
||||||
sysfs_path = udev.device_get_sysfs_path(self.data)
|
sysfs_path = udev.device_get_sysfs_path(self.data)
|
||||||
@ -350,7 +350,7 @@ index 30e99aa1..770736b0 100644
|
|||||||
device.controllable = False
|
device.controllable = False
|
||||||
self._devicetree._add_device(device)
|
self._devicetree._add_device(device)
|
||||||
diff --git a/blivet/populator/helpers/luks.py b/blivet/populator/helpers/luks.py
|
diff --git a/blivet/populator/helpers/luks.py b/blivet/populator/helpers/luks.py
|
||||||
index 52795a98..51488691 100644
|
index 3221122a..9b5023f8 100644
|
||||||
--- a/blivet/populator/helpers/luks.py
|
--- a/blivet/populator/helpers/luks.py
|
||||||
+++ b/blivet/populator/helpers/luks.py
|
+++ b/blivet/populator/helpers/luks.py
|
||||||
@@ -43,7 +43,7 @@ class LUKSDevicePopulator(DevicePopulator):
|
@@ -43,7 +43,7 @@ class LUKSDevicePopulator(DevicePopulator):
|
||||||
@ -368,14 +368,14 @@ index 52795a98..51488691 100644
|
|||||||
def run(self):
|
def run(self):
|
||||||
- parents = self._devicetree._add_parent_devices(self.data)
|
- parents = self._devicetree._add_parent_devices(self.data)
|
||||||
+ parents = self._devicetree._add_slave_devices(self.data)
|
+ parents = self._devicetree._add_slave_devices(self.data)
|
||||||
device = IntegrityDevice(udev.device_get_name(self.data),
|
name = udev.device_get_name(self.data)
|
||||||
sysfs_path=udev.device_get_sysfs_path(self.data),
|
|
||||||
parents=parents,
|
try:
|
||||||
diff --git a/blivet/populator/helpers/lvm.py b/blivet/populator/helpers/lvm.py
|
diff --git a/blivet/populator/helpers/lvm.py b/blivet/populator/helpers/lvm.py
|
||||||
index c7adfa4e..b1626306 100644
|
index 6ef2f417..b549e8d3 100644
|
||||||
--- a/blivet/populator/helpers/lvm.py
|
--- a/blivet/populator/helpers/lvm.py
|
||||||
+++ b/blivet/populator/helpers/lvm.py
|
+++ b/blivet/populator/helpers/lvm.py
|
||||||
@@ -57,7 +57,7 @@ class LVMDevicePopulator(DevicePopulator):
|
@@ -58,7 +58,7 @@ class LVMDevicePopulator(DevicePopulator):
|
||||||
log.warning("found non-vg device with name %s", vg_name)
|
log.warning("found non-vg device with name %s", vg_name)
|
||||||
device = None
|
device = None
|
||||||
|
|
||||||
@ -385,7 +385,7 @@ index c7adfa4e..b1626306 100644
|
|||||||
# LVM provides no means to resolve conflicts caused by duplicated VG
|
# LVM provides no means to resolve conflicts caused by duplicated VG
|
||||||
# names, so we're just being optimistic here. Woo!
|
# names, so we're just being optimistic here. Woo!
|
||||||
diff --git a/blivet/populator/helpers/mdraid.py b/blivet/populator/helpers/mdraid.py
|
diff --git a/blivet/populator/helpers/mdraid.py b/blivet/populator/helpers/mdraid.py
|
||||||
index 3479e3f7..76aebf25 100644
|
index a7602d20..9bec11ef 100644
|
||||||
--- a/blivet/populator/helpers/mdraid.py
|
--- a/blivet/populator/helpers/mdraid.py
|
||||||
+++ b/blivet/populator/helpers/mdraid.py
|
+++ b/blivet/populator/helpers/mdraid.py
|
||||||
@@ -31,7 +31,7 @@ from ... import udev
|
@@ -31,7 +31,7 @@ from ... import udev
|
||||||
@ -448,7 +448,7 @@ index 96c0a9ad..10c745bf 100644
|
|||||||
|
|
||||||
return device
|
return device
|
||||||
diff --git a/blivet/populator/populator.py b/blivet/populator/populator.py
|
diff --git a/blivet/populator/populator.py b/blivet/populator/populator.py
|
||||||
index 75bb1741..d252281d 100644
|
index 3a419418..068270b2 100644
|
||||||
--- a/blivet/populator/populator.py
|
--- a/blivet/populator/populator.py
|
||||||
+++ b/blivet/populator/populator.py
|
+++ b/blivet/populator/populator.py
|
||||||
@@ -31,7 +31,7 @@ gi.require_version("BlockDev", "2.0")
|
@@ -31,7 +31,7 @@ gi.require_version("BlockDev", "2.0")
|
||||||
@ -549,17 +549,8 @@ index 75bb1741..d252281d 100644
|
|||||||
|
|
||||||
def _add_name(self, name):
|
def _add_name(self, name):
|
||||||
if name not in self.names:
|
if name not in self.names:
|
||||||
@@ -317,7 +318,7 @@ class PopulatorMixin(object):
|
|
||||||
continue
|
|
||||||
|
|
||||||
# Make sure lvm doesn't get confused by PVs that belong to
|
|
||||||
- # incomplete VGs. We will remove the PVs from the reject list when/if
|
|
||||||
+ # incomplete VGs. We will remove the PVs from the blacklist when/if
|
|
||||||
# the time comes to remove the incomplete VG and its PVs.
|
|
||||||
for pv in vg.pvs:
|
|
||||||
lvm.lvm_cc_addFilterRejectRegexp(pv.name)
|
|
||||||
diff --git a/blivet/threads.py b/blivet/threads.py
|
diff --git a/blivet/threads.py b/blivet/threads.py
|
||||||
index 7e6d3105..a70deb69 100644
|
index 5e2dff3f..1a5cc6db 100644
|
||||||
--- a/blivet/threads.py
|
--- a/blivet/threads.py
|
||||||
+++ b/blivet/threads.py
|
+++ b/blivet/threads.py
|
||||||
@@ -63,11 +63,12 @@ class SynchronizedMeta(type):
|
@@ -63,11 +63,12 @@ class SynchronizedMeta(type):
|
||||||
@ -577,7 +568,7 @@ index 7e6d3105..a70deb69 100644
|
|||||||
elif isinstance(obj, FunctionType):
|
elif isinstance(obj, FunctionType):
|
||||||
obj = exclusive(obj)
|
obj = exclusive(obj)
|
||||||
diff --git a/blivet/udev.py b/blivet/udev.py
|
diff --git a/blivet/udev.py b/blivet/udev.py
|
||||||
index a8297f3f..e1b67845 100644
|
index efbc53d6..ddc49a37 100644
|
||||||
--- a/blivet/udev.py
|
--- a/blivet/udev.py
|
||||||
+++ b/blivet/udev.py
|
+++ b/blivet/udev.py
|
||||||
@@ -39,7 +39,7 @@ from gi.repository import BlockDev as blockdev
|
@@ -39,7 +39,7 @@ from gi.repository import BlockDev as blockdev
|
||||||
@ -615,16 +606,16 @@ index a8297f3f..e1b67845 100644
|
|||||||
return True
|
return True
|
||||||
|
|
||||||
dev_path = "/sys/class/block/%s" % dev_name
|
dev_path = "/sys/class/block/%s" % dev_name
|
||||||
@@ -374,7 +374,7 @@ def device_is_disk(info):
|
@@ -375,7 +375,7 @@ def device_is_disk(info):
|
||||||
device_is_dm_lvm(info) or
|
|
||||||
device_is_dm_crypt(info) or
|
device_is_dm_crypt(info) or
|
||||||
|
device_is_dm_stratis(info) or
|
||||||
(device_is_md(info) and
|
(device_is_md(info) and
|
||||||
- (not device_get_md_container(info) and not all(device_is_disk(d) for d in device_get_parents(info))))))
|
- (not device_get_md_container(info) and not all(device_is_disk(d) for d in device_get_parents(info))))))
|
||||||
+ (not device_get_md_container(info) and not all(device_is_disk(d) for d in device_get_slaves(info))))))
|
+ (not device_get_md_container(info) and not all(device_is_disk(d) for d in device_get_slaves(info))))))
|
||||||
|
|
||||||
|
|
||||||
def device_is_partition(info):
|
def device_is_partition(info):
|
||||||
@@ -453,18 +453,18 @@ def device_get_devname(info):
|
@@ -454,18 +454,18 @@ def device_get_devname(info):
|
||||||
return info.get('DEVNAME')
|
return info.get('DEVNAME')
|
||||||
|
|
||||||
|
|
||||||
@ -651,7 +642,7 @@ index a8297f3f..e1b67845 100644
|
|||||||
|
|
||||||
|
|
||||||
def device_get_holders(info):
|
def device_get_holders(info):
|
||||||
@@ -736,7 +736,7 @@ def device_get_partition_disk(info):
|
@@ -742,7 +742,7 @@ def device_get_partition_disk(info):
|
||||||
disk = None
|
disk = None
|
||||||
majorminor = info.get("ID_PART_ENTRY_DISK")
|
majorminor = info.get("ID_PART_ENTRY_DISK")
|
||||||
sysfs_path = device_get_sysfs_path(info)
|
sysfs_path = device_get_sysfs_path(info)
|
||||||
@ -660,7 +651,7 @@ index a8297f3f..e1b67845 100644
|
|||||||
if majorminor:
|
if majorminor:
|
||||||
major, minor = majorminor.split(":")
|
major, minor = majorminor.split(":")
|
||||||
for device in get_devices():
|
for device in get_devices():
|
||||||
@@ -744,8 +744,8 @@ def device_get_partition_disk(info):
|
@@ -750,8 +750,8 @@ def device_get_partition_disk(info):
|
||||||
disk = device_get_name(device)
|
disk = device_get_name(device)
|
||||||
break
|
break
|
||||||
elif device_is_dm_partition(info):
|
elif device_is_dm_partition(info):
|
||||||
@ -671,11 +662,11 @@ index a8297f3f..e1b67845 100644
|
|||||||
if len(parents) == 1:
|
if len(parents) == 1:
|
||||||
disk = resolve_devspec(parents[0].replace('!', '/'))
|
disk = resolve_devspec(parents[0].replace('!', '/'))
|
||||||
else:
|
else:
|
||||||
diff --git a/tests/devicefactory_test.py b/tests/devicefactory_test.py
|
diff --git a/tests/unit_tests/devicefactory_test.py b/tests/unit_tests/devicefactory_test.py
|
||||||
index dc0d6408..dfd78a7a 100644
|
index ff6bcb9e..552aadc1 100644
|
||||||
--- a/tests/devicefactory_test.py
|
--- a/tests/unit_tests/devicefactory_test.py
|
||||||
+++ b/tests/devicefactory_test.py
|
+++ b/tests/unit_tests/devicefactory_test.py
|
||||||
@@ -112,9 +112,9 @@ class DeviceFactoryTestCase(unittest.TestCase):
|
@@ -115,9 +115,9 @@ class DeviceFactoryTestCase(unittest.TestCase):
|
||||||
kwargs.get("encrypted", False) or
|
kwargs.get("encrypted", False) or
|
||||||
kwargs.get("container_encrypted", False))
|
kwargs.get("container_encrypted", False))
|
||||||
if kwargs.get("encrypted", False):
|
if kwargs.get("encrypted", False):
|
||||||
@ -687,7 +678,7 @@ index dc0d6408..dfd78a7a 100644
|
|||||||
kwargs.get("luks_sector_size", 0))
|
kwargs.get("luks_sector_size", 0))
|
||||||
|
|
||||||
self.assertTrue(set(device.disks).issubset(kwargs["disks"]))
|
self.assertTrue(set(device.disks).issubset(kwargs["disks"]))
|
||||||
@@ -354,7 +354,7 @@ class LVMFactoryTestCase(DeviceFactoryTestCase):
|
@@ -357,7 +357,7 @@ class LVMFactoryTestCase(DeviceFactoryTestCase):
|
||||||
device = args[0]
|
device = args[0]
|
||||||
|
|
||||||
if kwargs.get("encrypted"):
|
if kwargs.get("encrypted"):
|
||||||
@ -696,7 +687,7 @@ index dc0d6408..dfd78a7a 100644
|
|||||||
else:
|
else:
|
||||||
container = device.container
|
container = device.container
|
||||||
|
|
||||||
@@ -373,7 +373,7 @@ class LVMFactoryTestCase(DeviceFactoryTestCase):
|
@@ -376,7 +376,7 @@ class LVMFactoryTestCase(DeviceFactoryTestCase):
|
||||||
self.assertIsInstance(pv, member_class)
|
self.assertIsInstance(pv, member_class)
|
||||||
|
|
||||||
if pv.encrypted:
|
if pv.encrypted:
|
||||||
@ -705,7 +696,7 @@ index dc0d6408..dfd78a7a 100644
|
|||||||
kwargs.get("luks_version", crypto.DEFAULT_LUKS_VERSION))
|
kwargs.get("luks_version", crypto.DEFAULT_LUKS_VERSION))
|
||||||
|
|
||||||
@patch("blivet.formats.lvmpv.LVMPhysicalVolume.formattable", return_value=True)
|
@patch("blivet.formats.lvmpv.LVMPhysicalVolume.formattable", return_value=True)
|
||||||
@@ -589,7 +589,7 @@ class LVMThinPFactoryTestCase(LVMFactoryTestCase):
|
@@ -592,7 +592,7 @@ class LVMThinPFactoryTestCase(LVMFactoryTestCase):
|
||||||
device = args[0]
|
device = args[0]
|
||||||
|
|
||||||
if kwargs.get("encrypted", False):
|
if kwargs.get("encrypted", False):
|
||||||
@ -714,10 +705,10 @@ index dc0d6408..dfd78a7a 100644
|
|||||||
else:
|
else:
|
||||||
thinlv = device
|
thinlv = device
|
||||||
|
|
||||||
diff --git a/tests/devices_test/size_test.py b/tests/devices_test/size_test.py
|
diff --git a/tests/unit_tests/devices_test/device_size_test.py b/tests/unit_tests/devices_test/device_size_test.py
|
||||||
index d0c0a3f4..a1efa86d 100644
|
index d0c0a3f4..a1efa86d 100644
|
||||||
--- a/tests/devices_test/size_test.py
|
--- a/tests/unit_tests/devices_test/device_size_test.py
|
||||||
+++ b/tests/devices_test/size_test.py
|
+++ b/tests/unit_tests/devices_test/device_size_test.py
|
||||||
@@ -107,8 +107,8 @@ class LUKSDeviceSizeTest(StorageDeviceSizeTest):
|
@@ -107,8 +107,8 @@ class LUKSDeviceSizeTest(StorageDeviceSizeTest):
|
||||||
|
|
||||||
def _get_device(self, *args, **kwargs):
|
def _get_device(self, *args, **kwargs):
|
||||||
@ -735,11 +726,11 @@ index d0c0a3f4..a1efa86d 100644
|
|||||||
self.assertEqual(dev.size, initial_size)
|
self.assertEqual(dev.size, initial_size)
|
||||||
- self.assertEqual(dev.raw_device.size, initial_size + crypto.LUKS_METADATA_SIZE)
|
- self.assertEqual(dev.raw_device.size, initial_size + crypto.LUKS_METADATA_SIZE)
|
||||||
+ self.assertEqual(dev.slave.size, initial_size + crypto.LUKS_METADATA_SIZE)
|
+ self.assertEqual(dev.slave.size, initial_size + crypto.LUKS_METADATA_SIZE)
|
||||||
diff --git a/tests/populator_test.py b/tests/populator_test.py
|
diff --git a/tests/unit_tests/populator_test.py b/tests/unit_tests/populator_test.py
|
||||||
index a7748a9d..531ec74b 100644
|
index 369fe878..7ba04bac 100644
|
||||||
--- a/tests/populator_test.py
|
--- a/tests/unit_tests/populator_test.py
|
||||||
+++ b/tests/populator_test.py
|
+++ b/tests/unit_tests/populator_test.py
|
||||||
@@ -81,7 +81,7 @@ class DMDevicePopulatorTestCase(PopulatorHelperTestCase):
|
@@ -86,7 +86,7 @@ class DMDevicePopulatorTestCase(PopulatorHelperTestCase):
|
||||||
@patch.object(DeviceTree, "get_device_by_name")
|
@patch.object(DeviceTree, "get_device_by_name")
|
||||||
@patch.object(DMDevice, "status", return_value=True)
|
@patch.object(DMDevice, "status", return_value=True)
|
||||||
@patch.object(DMDevice, "update_sysfs_path")
|
@patch.object(DMDevice, "update_sysfs_path")
|
||||||
@ -748,7 +739,7 @@ index a7748a9d..531ec74b 100644
|
|||||||
@patch("blivet.udev.device_get_name")
|
@patch("blivet.udev.device_get_name")
|
||||||
@patch("blivet.udev.device_get_sysfs_path", return_value=sentinel.sysfs_path)
|
@patch("blivet.udev.device_get_sysfs_path", return_value=sentinel.sysfs_path)
|
||||||
def test_run(self, *args):
|
def test_run(self, *args):
|
||||||
@@ -90,7 +90,7 @@ class DMDevicePopulatorTestCase(PopulatorHelperTestCase):
|
@@ -95,7 +95,7 @@ class DMDevicePopulatorTestCase(PopulatorHelperTestCase):
|
||||||
|
|
||||||
devicetree = DeviceTree()
|
devicetree = DeviceTree()
|
||||||
|
|
||||||
@ -756,17 +747,17 @@ index a7748a9d..531ec74b 100644
|
|||||||
+ # The general case for dm devices is that adding the slave/parent devices
|
+ # The general case for dm devices is that adding the slave/parent devices
|
||||||
# will result in the dm device itself being in the tree.
|
# will result in the dm device itself being in the tree.
|
||||||
device = Mock()
|
device = Mock()
|
||||||
devicetree.get_device_by_name.return_value = device
|
device.id = 0
|
||||||
@@ -99,7 +99,7 @@ class DMDevicePopulatorTestCase(PopulatorHelperTestCase):
|
@@ -106,7 +106,7 @@ class DMDevicePopulatorTestCase(PopulatorHelperTestCase):
|
||||||
|
|
||||||
parent = Mock()
|
parent = Mock()
|
||||||
|
parent.id = 0
|
||||||
parent.parents = []
|
parent.parents = []
|
||||||
- devicetree._add_parent_devices.return_value = [parent]
|
- devicetree._add_parent_devices.return_value = [parent]
|
||||||
+ devicetree._add_slave_devices.return_value = [parent]
|
+ devicetree._add_slave_devices.return_value = [parent]
|
||||||
devicetree._add_device(parent)
|
devicetree._add_device(parent)
|
||||||
devicetree.get_device_by_name.return_value = None
|
devicetree.get_device_by_name.return_value = None
|
||||||
device_name = "dmdevice"
|
device_name = "dmdevice"
|
||||||
@@ -228,7 +228,7 @@ class LVMDevicePopulatorTestCase(PopulatorHelperTestCase):
|
@@ -235,7 +235,7 @@ class LVMDevicePopulatorTestCase(PopulatorHelperTestCase):
|
||||||
# could be the first helper class checked.
|
# could be the first helper class checked.
|
||||||
|
|
||||||
@patch.object(DeviceTree, "get_device_by_name")
|
@patch.object(DeviceTree, "get_device_by_name")
|
||||||
@ -775,7 +766,7 @@ index a7748a9d..531ec74b 100644
|
|||||||
@patch("blivet.udev.device_get_name")
|
@patch("blivet.udev.device_get_name")
|
||||||
@patch("blivet.udev.device_get_lv_vg_name")
|
@patch("blivet.udev.device_get_lv_vg_name")
|
||||||
def test_run(self, *args):
|
def test_run(self, *args):
|
||||||
@@ -240,7 +240,7 @@ class LVMDevicePopulatorTestCase(PopulatorHelperTestCase):
|
@@ -247,7 +247,7 @@ class LVMDevicePopulatorTestCase(PopulatorHelperTestCase):
|
||||||
devicetree = DeviceTree()
|
devicetree = DeviceTree()
|
||||||
data = Mock()
|
data = Mock()
|
||||||
|
|
||||||
@ -784,7 +775,7 @@ index a7748a9d..531ec74b 100644
|
|||||||
device_get_name.return_value = sentinel.lv_name
|
device_get_name.return_value = sentinel.lv_name
|
||||||
devicetree.get_device_by_name.return_value = None
|
devicetree.get_device_by_name.return_value = None
|
||||||
|
|
||||||
@@ -260,7 +260,7 @@ class LVMDevicePopulatorTestCase(PopulatorHelperTestCase):
|
@@ -267,7 +267,7 @@ class LVMDevicePopulatorTestCase(PopulatorHelperTestCase):
|
||||||
call(sentinel.vg_name),
|
call(sentinel.vg_name),
|
||||||
call(sentinel.lv_name)])
|
call(sentinel.lv_name)])
|
||||||
|
|
||||||
@ -793,7 +784,7 @@ index a7748a9d..531ec74b 100644
|
|||||||
get_device_by_name.side_effect = None
|
get_device_by_name.side_effect = None
|
||||||
get_device_by_name.return_value = None
|
get_device_by_name.return_value = None
|
||||||
self.assertEqual(helper.run(), None)
|
self.assertEqual(helper.run(), None)
|
||||||
@@ -625,7 +625,7 @@ class MDDevicePopulatorTestCase(PopulatorHelperTestCase):
|
@@ -639,7 +639,7 @@ class MDDevicePopulatorTestCase(PopulatorHelperTestCase):
|
||||||
# could be the first helper class checked.
|
# could be the first helper class checked.
|
||||||
|
|
||||||
@patch.object(DeviceTree, "get_device_by_name")
|
@patch.object(DeviceTree, "get_device_by_name")
|
||||||
@ -802,7 +793,7 @@ index a7748a9d..531ec74b 100644
|
|||||||
@patch("blivet.udev.device_get_name")
|
@patch("blivet.udev.device_get_name")
|
||||||
@patch("blivet.udev.device_get_md_uuid")
|
@patch("blivet.udev.device_get_md_uuid")
|
||||||
@patch("blivet.udev.device_get_md_name")
|
@patch("blivet.udev.device_get_md_name")
|
||||||
@@ -636,7 +636,7 @@ class MDDevicePopulatorTestCase(PopulatorHelperTestCase):
|
@@ -650,7 +650,7 @@ class MDDevicePopulatorTestCase(PopulatorHelperTestCase):
|
||||||
|
|
||||||
devicetree = DeviceTree()
|
devicetree = DeviceTree()
|
||||||
|
|
||||||
@ -811,7 +802,7 @@ index a7748a9d..531ec74b 100644
|
|||||||
data = Mock()
|
data = Mock()
|
||||||
device = Mock()
|
device = Mock()
|
||||||
device.parents = []
|
device.parents = []
|
||||||
@@ -699,12 +699,12 @@ class MultipathDevicePopulatorTestCase(PopulatorHelperTestCase):
|
@@ -713,12 +713,12 @@ class MultipathDevicePopulatorTestCase(PopulatorHelperTestCase):
|
||||||
# could be the first helper class checked.
|
# could be the first helper class checked.
|
||||||
|
|
||||||
@patch("blivet.udev.device_get_sysfs_path")
|
@patch("blivet.udev.device_get_sysfs_path")
|
||||||
@ -826,20 +817,20 @@ index a7748a9d..531ec74b 100644
|
|||||||
|
|
||||||
devicetree = DeviceTree()
|
devicetree = DeviceTree()
|
||||||
# set up some fake udev data to verify handling of specific entries
|
# set up some fake udev data to verify handling of specific entries
|
||||||
@@ -719,13 +719,13 @@ class MultipathDevicePopulatorTestCase(PopulatorHelperTestCase):
|
@@ -733,13 +733,13 @@ class MultipathDevicePopulatorTestCase(PopulatorHelperTestCase):
|
||||||
|
|
||||||
device_name = "mpathtest"
|
device_name = "mpathtest"
|
||||||
device_get_name.return_value = device_name
|
device_get_name.return_value = device_name
|
||||||
- parent_1 = Mock(tags=set(), wwn=wwn[2:])
|
- parent_1 = Mock(tags=set(), wwn=wwn[2:], id=0)
|
||||||
- parent_1.parents = []
|
- parent_1.parents = []
|
||||||
- parent_2 = Mock(tags=set(), wwn=wwn[2:])
|
- parent_2 = Mock(tags=set(), wwn=wwn[2:], id=0)
|
||||||
- parent_2.parents = []
|
- parent_2.parents = []
|
||||||
- devicetree._add_device(parent_1)
|
- devicetree._add_device(parent_1)
|
||||||
- devicetree._add_device(parent_2)
|
- devicetree._add_device(parent_2)
|
||||||
- add_parent_devices.return_value = [parent_1, parent_2]
|
- add_parent_devices.return_value = [parent_1, parent_2]
|
||||||
+ slave_1 = Mock(tags=set(), wwn=wwn[2:])
|
+ slave_1 = Mock(tags=set(), wwn=wwn[2:], id=0)
|
||||||
+ slave_1.parents = []
|
+ slave_1.parents = []
|
||||||
+ slave_2 = Mock(tags=set(), wwn=wwn[2:])
|
+ slave_2 = Mock(tags=set(), wwn=wwn[2:], id=0)
|
||||||
+ slave_2.parents = []
|
+ slave_2.parents = []
|
||||||
+ devicetree._add_device(slave_1)
|
+ devicetree._add_device(slave_1)
|
||||||
+ devicetree._add_device(slave_2)
|
+ devicetree._add_device(slave_2)
|
||||||
@ -847,11 +838,11 @@ index a7748a9d..531ec74b 100644
|
|||||||
|
|
||||||
helper = self.helper_class(devicetree, data)
|
helper = self.helper_class(devicetree, data)
|
||||||
|
|
||||||
diff --git a/tests/udev_test.py b/tests/udev_test.py
|
diff --git a/tests/unit_tests/udev_test.py b/tests/unit_tests/udev_test.py
|
||||||
index f9b10620..d30a647b 100644
|
index b208efa8..ebcd59e2 100644
|
||||||
--- a/tests/udev_test.py
|
--- a/tests/unit_tests/udev_test.py
|
||||||
+++ b/tests/udev_test.py
|
+++ b/tests/unit_tests/udev_test.py
|
||||||
@@ -45,11 +45,11 @@ class UdevTest(unittest.TestCase):
|
@@ -49,11 +49,11 @@ class UdevTest(unittest.TestCase):
|
||||||
@mock.patch('blivet.udev.device_is_dm_crypt', return_value=False)
|
@mock.patch('blivet.udev.device_is_dm_crypt', return_value=False)
|
||||||
@mock.patch('blivet.udev.device_is_md')
|
@mock.patch('blivet.udev.device_is_md')
|
||||||
@mock.patch('blivet.udev.device_get_md_container')
|
@mock.patch('blivet.udev.device_get_md_container')
|
||||||
@ -865,7 +856,7 @@ index f9b10620..d30a647b 100644
|
|||||||
|
|
||||||
disk_parents = [dict(DEVTYPE="disk", SYS_PATH='/fake/path/2'),
|
disk_parents = [dict(DEVTYPE="disk", SYS_PATH='/fake/path/2'),
|
||||||
dict(DEVTYPE="disk", SYS_PATH='/fake/path/3')]
|
dict(DEVTYPE="disk", SYS_PATH='/fake/path/3')]
|
||||||
@@ -64,20 +64,20 @@ class UdevTest(unittest.TestCase):
|
@@ -68,20 +68,20 @@ class UdevTest(unittest.TestCase):
|
||||||
# Intel FW RAID (MD RAID w/ container layer)
|
# Intel FW RAID (MD RAID w/ container layer)
|
||||||
# device_get_container will return some mock value which will evaluate to True
|
# device_get_container will return some mock value which will evaluate to True
|
||||||
device_get_md_container.return_value = mock.sentinel.md_container
|
device_get_md_container.return_value = mock.sentinel.md_container
|
||||||
@ -904,5 +895,5 @@ index 797bac85..6255104f 100644
|
|||||||
#
|
#
|
||||||
# create disk images
|
# create disk images
|
||||||
--
|
--
|
||||||
2.31.1
|
2.38.1
|
||||||
|
|
||||||
|
@ -1,7 +1,7 @@
|
|||||||
From 898178047ac4bc97ddccb193cb0e11f7fdf18196 Mon Sep 17 00:00:00 2001
|
From f6490c469904f4808c63a170210e53acc908b018 Mon Sep 17 00:00:00 2001
|
||||||
From: Vojtech Trefny <vtrefny@redhat.com>
|
From: Vojtech Trefny <vtrefny@redhat.com>
|
||||||
Date: Wed, 17 Aug 2022 14:24:21 +0200
|
Date: Wed, 17 Aug 2022 14:24:21 +0200
|
||||||
Subject: [PATCH 1/3] Use MD populator instead of DM to handle DDF RAID format
|
Subject: [PATCH 1/2] Use MD populator instead of DM to handle DDF RAID format
|
||||||
|
|
||||||
---
|
---
|
||||||
blivet/formats/dmraid.py | 2 +-
|
blivet/formats/dmraid.py | 2 +-
|
||||||
@ -9,7 +9,7 @@ Subject: [PATCH 1/3] Use MD populator instead of DM to handle DDF RAID format
|
|||||||
2 files changed, 2 insertions(+), 2 deletions(-)
|
2 files changed, 2 insertions(+), 2 deletions(-)
|
||||||
|
|
||||||
diff --git a/blivet/formats/dmraid.py b/blivet/formats/dmraid.py
|
diff --git a/blivet/formats/dmraid.py b/blivet/formats/dmraid.py
|
||||||
index 2ba9dcfe5..ce15905dc 100644
|
index 2ba9dcfe..ce15905d 100644
|
||||||
--- a/blivet/formats/dmraid.py
|
--- a/blivet/formats/dmraid.py
|
||||||
+++ b/blivet/formats/dmraid.py
|
+++ b/blivet/formats/dmraid.py
|
||||||
@@ -43,7 +43,7 @@ class DMRaidMember(DeviceFormat):
|
@@ -43,7 +43,7 @@ class DMRaidMember(DeviceFormat):
|
||||||
@ -22,7 +22,7 @@ index 2ba9dcfe5..ce15905dc 100644
|
|||||||
"isw_raid_member",
|
"isw_raid_member",
|
||||||
"jmicron_raid_member", "lsi_mega_raid_member",
|
"jmicron_raid_member", "lsi_mega_raid_member",
|
||||||
diff --git a/blivet/formats/mdraid.py b/blivet/formats/mdraid.py
|
diff --git a/blivet/formats/mdraid.py b/blivet/formats/mdraid.py
|
||||||
index 41ddef810..4aa3f3b07 100644
|
index 41ddef81..4aa3f3b0 100644
|
||||||
--- a/blivet/formats/mdraid.py
|
--- a/blivet/formats/mdraid.py
|
||||||
+++ b/blivet/formats/mdraid.py
|
+++ b/blivet/formats/mdraid.py
|
||||||
@@ -41,7 +41,7 @@ class MDRaidMember(DeviceFormat):
|
@@ -41,7 +41,7 @@ class MDRaidMember(DeviceFormat):
|
||||||
@ -34,11 +34,14 @@ index 41ddef810..4aa3f3b07 100644
|
|||||||
parted_flag = PARTITION_RAID
|
parted_flag = PARTITION_RAID
|
||||||
_formattable = True # can be formatted
|
_formattable = True # can be formatted
|
||||||
_supported = True # is supported
|
_supported = True # is supported
|
||||||
|
--
|
||||||
|
2.38.1
|
||||||
|
|
||||||
From c487c6178ee9859163379946c1bdc3b2df1857b1 Mon Sep 17 00:00:00 2001
|
|
||||||
|
From 5fadd850aae217d7692a6c8a50b2dcd5e61a63cd Mon Sep 17 00:00:00 2001
|
||||||
From: Vojtech Trefny <vtrefny@redhat.com>
|
From: Vojtech Trefny <vtrefny@redhat.com>
|
||||||
Date: Wed, 17 Aug 2022 14:24:58 +0200
|
Date: Wed, 17 Aug 2022 14:24:58 +0200
|
||||||
Subject: [PATCH 2/3] Do not read DDF RAID UUID from udev
|
Subject: [PATCH 2/2] Do not read DDF RAID UUID from udev
|
||||||
|
|
||||||
The UUID we get from udev isn't the array UUID, we need to get
|
The UUID we get from udev isn't the array UUID, we need to get
|
||||||
that using libblockdev.
|
that using libblockdev.
|
||||||
@ -47,7 +50,7 @@ that using libblockdev.
|
|||||||
1 file changed, 10 insertions(+), 6 deletions(-)
|
1 file changed, 10 insertions(+), 6 deletions(-)
|
||||||
|
|
||||||
diff --git a/blivet/populator/helpers/mdraid.py b/blivet/populator/helpers/mdraid.py
|
diff --git a/blivet/populator/helpers/mdraid.py b/blivet/populator/helpers/mdraid.py
|
||||||
index 76aebf250..9bec11efb 100644
|
index 3479e3f7..a7602d20 100644
|
||||||
--- a/blivet/populator/helpers/mdraid.py
|
--- a/blivet/populator/helpers/mdraid.py
|
||||||
+++ b/blivet/populator/helpers/mdraid.py
|
+++ b/blivet/populator/helpers/mdraid.py
|
||||||
@@ -98,17 +98,21 @@ class MDFormatPopulator(FormatPopulator):
|
@@ -98,17 +98,21 @@ class MDFormatPopulator(FormatPopulator):
|
||||||
@ -78,30 +81,6 @@ index 76aebf250..9bec11efb 100644
|
|||||||
return kwargs
|
return kwargs
|
||||||
|
|
||||||
def run(self):
|
def run(self):
|
||||||
|
--
|
||||||
|
2.38.1
|
||||||
|
|
||||||
From 325681bcd40fc4f0e13a4d23c889e1f7cc043cc1 Mon Sep 17 00:00:00 2001
|
|
||||||
From: Vojtech Trefny <vtrefny@redhat.com>
|
|
||||||
Date: Thu, 17 Mar 2022 15:48:25 +0100
|
|
||||||
Subject: [PATCH 3/3] Do not crash when a disk populator doesn't return kwargs
|
|
||||||
|
|
||||||
This happens when trying to use Blivet on a system with a BIOS
|
|
||||||
RAID without dmraid installed. Because we don't fully support
|
|
||||||
BIOS RAIDs using MD the MDBiosRaidDevicePopulator helper fails
|
|
||||||
to get kwargs for the BIOS RAID "disk" and populate fails.
|
|
||||||
---
|
|
||||||
blivet/populator/helpers/disk.py | 2 ++
|
|
||||||
1 file changed, 2 insertions(+)
|
|
||||||
|
|
||||||
diff --git a/blivet/populator/helpers/disk.py b/blivet/populator/helpers/disk.py
|
|
||||||
index 2e5026f7e..9db7b810d 100644
|
|
||||||
--- a/blivet/populator/helpers/disk.py
|
|
||||||
+++ b/blivet/populator/helpers/disk.py
|
|
||||||
@@ -68,6 +68,8 @@ def run(self):
|
|
||||||
log_method_call(self, name=name)
|
|
||||||
|
|
||||||
kwargs = self._get_kwargs()
|
|
||||||
+ if not kwargs:
|
|
||||||
+ return
|
|
||||||
device = self._device_class(name, **kwargs)
|
|
||||||
self._devicetree._add_device(device)
|
|
||||||
return device
|
|
@ -1,35 +0,0 @@
|
|||||||
From 8ece3da18b1abb89320d02f4475002e6a3ed7875 Mon Sep 17 00:00:00 2001
|
|
||||||
From: Vojtech Trefny <vtrefny@redhat.com>
|
|
||||||
Date: Thu, 20 May 2021 13:40:26 +0200
|
|
||||||
Subject: [PATCH] Fix activating old style LVM snapshots
|
|
||||||
|
|
||||||
The old style snapshots are activated together with the origin LV
|
|
||||||
so we need to make sure it is activated to be able to remove the
|
|
||||||
snapshot or its format.
|
|
||||||
|
|
||||||
Resolves: rhbz#1961739
|
|
||||||
---
|
|
||||||
blivet/devices/lvm.py | 10 +++++++---
|
|
||||||
1 file changed, 7 insertions(+), 3 deletions(-)
|
|
||||||
|
|
||||||
diff --git a/blivet/devices/lvm.py b/blivet/devices/lvm.py
|
|
||||||
index a55515fcb..fb57804d9 100644
|
|
||||||
--- a/blivet/devices/lvm.py
|
|
||||||
+++ b/blivet/devices/lvm.py
|
|
||||||
@@ -1461,9 +1461,13 @@ def _set_format(self, fmt): # pylint: disable=unused-argument
|
|
||||||
self._update_format_from_origin()
|
|
||||||
|
|
||||||
@old_snapshot_specific
|
|
||||||
- def setup(self, orig=False):
|
|
||||||
- # the old snapshot cannot be setup and torn down
|
|
||||||
- pass
|
|
||||||
+ def setup(self, orig=False): # pylint: disable=unused-argument
|
|
||||||
+ # the old snapshot is activated together with the origin
|
|
||||||
+ if self.origin and not self.origin.status:
|
|
||||||
+ try:
|
|
||||||
+ self.origin.setup()
|
|
||||||
+ except blockdev.LVMError as lvmerr:
|
|
||||||
+ log.error("failed to activate origin LV: %s", lvmerr)
|
|
||||||
|
|
||||||
@old_snapshot_specific
|
|
||||||
def teardown(self, recursive=False):
|
|
@ -1,75 +0,0 @@
|
|||||||
From 344e624f91010b6041c22ee8a24c9305b82af969 Mon Sep 17 00:00:00 2001
|
|
||||||
From: Vojtech Trefny <vtrefny@redhat.com>
|
|
||||||
Date: Tue, 18 May 2021 12:54:02 +0200
|
|
||||||
Subject: [PATCH] Fix resolving devices with names that look like BIOS drive
|
|
||||||
number
|
|
||||||
|
|
||||||
A RAID array named "10" will not be resolved because we try to
|
|
||||||
resolve it using EDD data and after this lookup fails, we don't
|
|
||||||
try the name.
|
|
||||||
|
|
||||||
Resolves: rhbz#1960798
|
|
||||||
---
|
|
||||||
blivet/devicetree.py | 18 +++++++++---------
|
|
||||||
tests/devicetree_test.py | 4 ++++
|
|
||||||
2 files changed, 13 insertions(+), 9 deletions(-)
|
|
||||||
|
|
||||||
diff --git a/blivet/devicetree.py b/blivet/devicetree.py
|
|
||||||
index 88e9f0e5..f4ae1968 100644
|
|
||||||
--- a/blivet/devicetree.py
|
|
||||||
+++ b/blivet/devicetree.py
|
|
||||||
@@ -634,20 +634,20 @@ class DeviceTreeBase(object):
|
|
||||||
(label.startswith("'") and label.endswith("'"))):
|
|
||||||
label = label[1:-1]
|
|
||||||
device = self.labels.get(label)
|
|
||||||
- elif re.match(r'(0x)?[A-Fa-f0-9]{2}(p\d+)?$', devspec):
|
|
||||||
- # BIOS drive number
|
|
||||||
- (drive, _p, partnum) = devspec.partition("p")
|
|
||||||
- spec = int(drive, 16)
|
|
||||||
- for (edd_name, edd_number) in self.edd_dict.items():
|
|
||||||
- if edd_number == spec:
|
|
||||||
- device = self.get_device_by_name(edd_name + partnum)
|
|
||||||
- break
|
|
||||||
elif options and "nodev" in options.split(","):
|
|
||||||
device = self.get_device_by_name(devspec)
|
|
||||||
if not device:
|
|
||||||
device = self.get_device_by_path(devspec)
|
|
||||||
else:
|
|
||||||
- if not devspec.startswith("/dev/"):
|
|
||||||
+ if re.match(r'(0x)?[A-Fa-f0-9]{2}(p\d+)?$', devspec):
|
|
||||||
+ # BIOS drive number
|
|
||||||
+ (drive, _p, partnum) = devspec.partition("p")
|
|
||||||
+ spec = int(drive, 16)
|
|
||||||
+ for (edd_name, edd_number) in self.edd_dict.items():
|
|
||||||
+ if edd_number == spec:
|
|
||||||
+ device = self.get_device_by_name(edd_name + partnum)
|
|
||||||
+ break
|
|
||||||
+ if not device and not devspec.startswith("/dev/"):
|
|
||||||
device = self.get_device_by_name(devspec)
|
|
||||||
if not device:
|
|
||||||
devspec = "/dev/" + devspec
|
|
||||||
diff --git a/tests/devicetree_test.py b/tests/devicetree_test.py
|
|
||||||
index 11f8469d..b033343d 100644
|
|
||||||
--- a/tests/devicetree_test.py
|
|
||||||
+++ b/tests/devicetree_test.py
|
|
||||||
@@ -49,6 +49,9 @@ class DeviceTreeTestCase(unittest.TestCase):
|
|
||||||
dev3 = StorageDevice("sdp2", exists=True)
|
|
||||||
dt._add_device(dev3)
|
|
||||||
|
|
||||||
+ dev4 = StorageDevice("10", exists=True)
|
|
||||||
+ dt._add_device(dev4)
|
|
||||||
+
|
|
||||||
dt.edd_dict.update({"dev1": 0x81,
|
|
||||||
"dev2": 0x82})
|
|
||||||
|
|
||||||
@@ -62,6 +65,7 @@ class DeviceTreeTestCase(unittest.TestCase):
|
|
||||||
self.assertEqual(dt.resolve_device("0x82"), dev2)
|
|
||||||
|
|
||||||
self.assertEqual(dt.resolve_device(dev3.name), dev3)
|
|
||||||
+ self.assertEqual(dt.resolve_device(dev4.name), dev4)
|
|
||||||
|
|
||||||
def test_device_name(self):
|
|
||||||
# check that devicetree.names property contains all device's names
|
|
||||||
--
|
|
||||||
2.31.1
|
|
||||||
|
|
77
SOURCES/0006-Revert-Remove-the-Blivet.roots-attribute.patch
Normal file
77
SOURCES/0006-Revert-Remove-the-Blivet.roots-attribute.patch
Normal file
@ -0,0 +1,77 @@
|
|||||||
|
From 62af1d7f96b8ed8eb8f2732787576161ae5da79f Mon Sep 17 00:00:00 2001
|
||||||
|
From: Vojtech Trefny <vtrefny@redhat.com>
|
||||||
|
Date: Thu, 13 Oct 2022 10:47:52 +0200
|
||||||
|
Subject: [PATCH] Revert "Remove the Blivet.roots attribute"
|
||||||
|
|
||||||
|
This reverts commit 19a826073345ca6b57a8f9a95ec855892320300e.
|
||||||
|
---
|
||||||
|
blivet/blivet.py | 21 +++++++++++++++++++++
|
||||||
|
blivet/devicefactory.py | 3 +++
|
||||||
|
2 files changed, 24 insertions(+)
|
||||||
|
|
||||||
|
diff --git a/blivet/blivet.py b/blivet/blivet.py
|
||||||
|
index bf72ee9c..dc066b03 100644
|
||||||
|
--- a/blivet/blivet.py
|
||||||
|
+++ b/blivet/blivet.py
|
||||||
|
@@ -88,6 +88,7 @@ class Blivet(object):
|
||||||
|
self.devicetree = DeviceTree(ignored_disks=self.ignored_disks,
|
||||||
|
exclusive_disks=self.exclusive_disks,
|
||||||
|
disk_images=self.disk_images)
|
||||||
|
+ self.roots = []
|
||||||
|
|
||||||
|
@property
|
||||||
|
def short_product_name(self):
|
||||||
|
@@ -1314,5 +1315,25 @@ class Blivet(object):
|
||||||
|
p = partition.disk.format.parted_disk.getPartitionByPath(partition.path)
|
||||||
|
partition.parted_partition = p
|
||||||
|
|
||||||
|
+ for root in new.roots:
|
||||||
|
+ root.swaps = [new.devicetree.get_device_by_id(d.id, hidden=True) for d in root.swaps]
|
||||||
|
+ root.swaps = [s for s in root.swaps if s]
|
||||||
|
+
|
||||||
|
+ removed = set()
|
||||||
|
+ for (mountpoint, old_dev) in root.mounts.items():
|
||||||
|
+ if old_dev is None:
|
||||||
|
+ continue
|
||||||
|
+
|
||||||
|
+ new_dev = new.devicetree.get_device_by_id(old_dev.id, hidden=True)
|
||||||
|
+ if new_dev is None:
|
||||||
|
+ # if the device has been removed don't include this
|
||||||
|
+ # mountpoint at all
|
||||||
|
+ removed.add(mountpoint)
|
||||||
|
+ else:
|
||||||
|
+ root.mounts[mountpoint] = new_dev
|
||||||
|
+
|
||||||
|
+ for mnt in removed:
|
||||||
|
+ del root.mounts[mnt]
|
||||||
|
+
|
||||||
|
log.debug("finished Blivet copy")
|
||||||
|
return new
|
||||||
|
diff --git a/blivet/devicefactory.py b/blivet/devicefactory.py
|
||||||
|
index 8105bfc7..6f460f6d 100644
|
||||||
|
--- a/blivet/devicefactory.py
|
||||||
|
+++ b/blivet/devicefactory.py
|
||||||
|
@@ -383,6 +383,7 @@ class DeviceFactory(object):
|
||||||
|
# used for error recovery
|
||||||
|
self.__devices = []
|
||||||
|
self.__actions = []
|
||||||
|
+ self.__roots = []
|
||||||
|
|
||||||
|
def _is_container_encrypted(self):
|
||||||
|
return all(isinstance(p, LUKSDevice) for p in self.device.container.parents)
|
||||||
|
@@ -994,10 +995,12 @@ class DeviceFactory(object):
|
||||||
|
_blivet_copy = self.storage.copy()
|
||||||
|
self.__devices = _blivet_copy.devicetree._devices
|
||||||
|
self.__actions = _blivet_copy.devicetree._actions
|
||||||
|
+ self.__roots = _blivet_copy.roots
|
||||||
|
|
||||||
|
def _revert_devicetree(self):
|
||||||
|
self.storage.devicetree._devices = self.__devices
|
||||||
|
self.storage.devicetree._actions = self.__actions
|
||||||
|
+ self.storage.roots = self.__roots
|
||||||
|
|
||||||
|
|
||||||
|
class PartitionFactory(DeviceFactory):
|
||||||
|
--
|
||||||
|
2.38.1
|
||||||
|
|
@ -1,151 +0,0 @@
|
|||||||
From dc1e2fe7783748528cac2f7aa516c89d1959b052 Mon Sep 17 00:00:00 2001
|
|
||||||
From: Vojtech Trefny <vtrefny@redhat.com>
|
|
||||||
Date: Thu, 29 Jul 2021 14:44:22 +0200
|
|
||||||
Subject: [PATCH] Do not set chunk size for RAID 1
|
|
||||||
|
|
||||||
Setting chunk size for RAID 1 doesn't make sense and latest
|
|
||||||
mdadm started returning error instead of ignoring the --chunk
|
|
||||||
option when creating an array.
|
|
||||||
|
|
||||||
Resolves: rhbz#1987170
|
|
||||||
---
|
|
||||||
blivet/devicelibs/raid.py | 12 ++++++++++
|
|
||||||
blivet/devices/md.py | 15 ++++++++++---
|
|
||||||
tests/devices_test/md_test.py | 41 +++++++++++++++++++++++++++++++++--
|
|
||||||
3 files changed, 63 insertions(+), 5 deletions(-)
|
|
||||||
|
|
||||||
diff --git a/blivet/devicelibs/raid.py b/blivet/devicelibs/raid.py
|
|
||||||
index 19c3fae98..a9e241c7a 100644
|
|
||||||
--- a/blivet/devicelibs/raid.py
|
|
||||||
+++ b/blivet/devicelibs/raid.py
|
|
||||||
@@ -462,6 +462,18 @@ def _pad(self, size, chunk_size):
|
|
||||||
def _get_recommended_stride(self, member_count):
|
|
||||||
return None
|
|
||||||
|
|
||||||
+ def get_size(self, member_sizes, num_members=None, chunk_size=None, superblock_size_func=None):
|
|
||||||
+ if not member_sizes:
|
|
||||||
+ return Size(0)
|
|
||||||
+
|
|
||||||
+ if num_members is None:
|
|
||||||
+ num_members = len(member_sizes)
|
|
||||||
+
|
|
||||||
+ min_size = min(member_sizes)
|
|
||||||
+ superblock_size = superblock_size_func(min_size)
|
|
||||||
+ min_data_size = self._trim(min_size - superblock_size, chunk_size)
|
|
||||||
+ return self.get_net_array_size(num_members, min_data_size)
|
|
||||||
+
|
|
||||||
|
|
||||||
RAID1 = RAID1()
|
|
||||||
ALL_LEVELS.add_raid_level(RAID1)
|
|
||||||
diff --git a/blivet/devices/md.py b/blivet/devices/md.py
|
|
||||||
index 69eee93a5..d1a2faf1f 100644
|
|
||||||
--- a/blivet/devices/md.py
|
|
||||||
+++ b/blivet/devices/md.py
|
|
||||||
@@ -138,7 +138,7 @@ def __init__(self, name, level=None, major=None, minor=None, size=None,
|
|
||||||
if self.exists:
|
|
||||||
self._chunk_size = self.read_chunk_size()
|
|
||||||
else:
|
|
||||||
- self._chunk_size = chunk_size or mdraid.MD_CHUNK_SIZE
|
|
||||||
+ self.chunk_size = chunk_size or Size(0)
|
|
||||||
|
|
||||||
if not self.exists and not isinstance(metadata_version, str):
|
|
||||||
self.metadata_version = "default"
|
|
||||||
@@ -208,8 +208,14 @@ def sector_size(self):
|
|
||||||
|
|
||||||
@property
|
|
||||||
def chunk_size(self):
|
|
||||||
- if self.exists and self._chunk_size == Size(0):
|
|
||||||
- self._chunk_size = self.read_chunk_size()
|
|
||||||
+ if self._chunk_size == Size(0):
|
|
||||||
+ if self.exists:
|
|
||||||
+ return self.read_chunk_size()
|
|
||||||
+ else:
|
|
||||||
+ if self.level == raid.RAID1:
|
|
||||||
+ return self._chunk_size
|
|
||||||
+ else:
|
|
||||||
+ return mdraid.MD_CHUNK_SIZE
|
|
||||||
return self._chunk_size
|
|
||||||
|
|
||||||
@chunk_size.setter
|
|
||||||
@@ -223,6 +229,9 @@ def chunk_size(self, newsize):
|
|
||||||
if self.exists:
|
|
||||||
raise ValueError("cannot set chunk size for an existing device")
|
|
||||||
|
|
||||||
+ if self.level == raid.RAID1 and newsize != Size(0):
|
|
||||||
+ raise ValueError("specifying chunk size is not allowed for raid1")
|
|
||||||
+
|
|
||||||
self._chunk_size = newsize
|
|
||||||
|
|
||||||
def read_chunk_size(self):
|
|
||||||
diff --git a/tests/devices_test/md_test.py b/tests/devices_test/md_test.py
|
|
||||||
index 46df76d3d..47a0fa0cc 100644
|
|
||||||
--- a/tests/devices_test/md_test.py
|
|
||||||
+++ b/tests/devices_test/md_test.py
|
|
||||||
@@ -1,6 +1,11 @@
|
|
||||||
import six
|
|
||||||
import unittest
|
|
||||||
|
|
||||||
+try:
|
|
||||||
+ from unittest.mock import patch
|
|
||||||
+except ImportError:
|
|
||||||
+ from mock import patch
|
|
||||||
+
|
|
||||||
import blivet
|
|
||||||
|
|
||||||
from blivet.devices import StorageDevice
|
|
||||||
@@ -27,9 +32,27 @@ def test_chunk_size1(self):
|
|
||||||
raid_array = MDRaidArrayDevice(name="raid", level="raid0", member_devices=2,
|
|
||||||
total_devices=2, parents=[member1, member2])
|
|
||||||
|
|
||||||
- # no chunk_size specified -- default value
|
|
||||||
+ # no chunk_size specified and RAID0 -- default value
|
|
||||||
self.assertEqual(raid_array.chunk_size, mdraid.MD_CHUNK_SIZE)
|
|
||||||
|
|
||||||
+ with patch("blivet.devices.md.blockdev.md.create") as md_create:
|
|
||||||
+ raid_array._create()
|
|
||||||
+ md_create.assert_called_with("/dev/md/raid", "raid0", ["/dev/member1", "/dev/member2"],
|
|
||||||
+ 0, version="default", bitmap=False,
|
|
||||||
+ chunk_size=mdraid.MD_CHUNK_SIZE)
|
|
||||||
+
|
|
||||||
+ raid_array = MDRaidArrayDevice(name="raid", level="raid1", member_devices=2,
|
|
||||||
+ total_devices=2, parents=[member1, member2])
|
|
||||||
+
|
|
||||||
+ # no chunk_size specified and RAID1 -- no chunk size set (0)
|
|
||||||
+ self.assertEqual(raid_array.chunk_size, Size(0))
|
|
||||||
+
|
|
||||||
+ with patch("blivet.devices.md.blockdev.md.create") as md_create:
|
|
||||||
+ raid_array._create()
|
|
||||||
+ md_create.assert_called_with("/dev/md/raid", "raid1", ["/dev/member1", "/dev/member2"],
|
|
||||||
+ 0, version="default", bitmap=True,
|
|
||||||
+ chunk_size=0)
|
|
||||||
+
|
|
||||||
def test_chunk_size2(self):
|
|
||||||
|
|
||||||
member1 = StorageDevice("member1", fmt=blivet.formats.get_format("mdmember"),
|
|
||||||
@@ -40,11 +63,25 @@ def test_chunk_size2(self):
|
|
||||||
raid_array = MDRaidArrayDevice(name="raid", level="raid0", member_devices=2,
|
|
||||||
total_devices=2, parents=[member1, member2],
|
|
||||||
chunk_size=Size("1024 KiB"))
|
|
||||||
-
|
|
||||||
self.assertEqual(raid_array.chunk_size, Size("1024 KiB"))
|
|
||||||
|
|
||||||
+ # for raid0 setting chunk_size = 0 means "default"
|
|
||||||
+ raid_array.chunk_size = Size(0)
|
|
||||||
+ self.assertEqual(raid_array.chunk_size, mdraid.MD_CHUNK_SIZE)
|
|
||||||
+
|
|
||||||
with six.assertRaisesRegex(self, ValueError, "new chunk size must be of type Size"):
|
|
||||||
raid_array.chunk_size = 1
|
|
||||||
|
|
||||||
with six.assertRaisesRegex(self, ValueError, "new chunk size must be multiple of 4 KiB"):
|
|
||||||
raid_array.chunk_size = Size("5 KiB")
|
|
||||||
+
|
|
||||||
+ with six.assertRaisesRegex(self, ValueError, "specifying chunk size is not allowed for raid1"):
|
|
||||||
+ MDRaidArrayDevice(name="raid", level="raid1", member_devices=2,
|
|
||||||
+ total_devices=2, parents=[member1, member2],
|
|
||||||
+ chunk_size=Size("1024 KiB"))
|
|
||||||
+
|
|
||||||
+ raid_array = MDRaidArrayDevice(name="raid", level="raid1", member_devices=2,
|
|
||||||
+ total_devices=2, parents=[member1, member2])
|
|
||||||
+
|
|
||||||
+ with six.assertRaisesRegex(self, ValueError, "specifying chunk size is not allowed for raid1"):
|
|
||||||
+ raid_array.chunk_size = Size("512 KiB")
|
|
@ -0,0 +1,45 @@
|
|||||||
|
From 1561bfe8820118178bbb07021adc1cacd875c4c7 Mon Sep 17 00:00:00 2001
|
||||||
|
From: Vojtech Trefny <vtrefny@redhat.com>
|
||||||
|
Date: Tue, 18 Oct 2022 12:28:37 +0200
|
||||||
|
Subject: [PATCH] Fix potential AttributeError when getting stratis blockdev
|
||||||
|
info
|
||||||
|
|
||||||
|
---
|
||||||
|
blivet/static_data/stratis_info.py | 12 +++++++-----
|
||||||
|
1 file changed, 7 insertions(+), 5 deletions(-)
|
||||||
|
|
||||||
|
diff --git a/blivet/static_data/stratis_info.py b/blivet/static_data/stratis_info.py
|
||||||
|
index bd1c5a18..42f230ee 100644
|
||||||
|
--- a/blivet/static_data/stratis_info.py
|
||||||
|
+++ b/blivet/static_data/stratis_info.py
|
||||||
|
@@ -124,20 +124,22 @@ class StratisInfo(object):
|
||||||
|
log.error("Failed to get DBus properties of '%s'", blockdev_path)
|
||||||
|
return None
|
||||||
|
|
||||||
|
+ blockdev_uuid = str(uuid.UUID(properties["Uuid"]))
|
||||||
|
+
|
||||||
|
pool_path = properties["Pool"]
|
||||||
|
if pool_path == "/":
|
||||||
|
pool_name = ""
|
||||||
|
+ return StratisBlockdevInfo(path=properties["Devnode"], uuid=blockdev_uuid,
|
||||||
|
+ pool_name="", pool_uuid="", object_path=blockdev_path)
|
||||||
|
else:
|
||||||
|
pool_info = self._get_pool_info(properties["Pool"])
|
||||||
|
if not pool_info:
|
||||||
|
return None
|
||||||
|
pool_name = pool_info.name
|
||||||
|
|
||||||
|
- blockdev_uuid = str(uuid.UUID(properties["Uuid"]))
|
||||||
|
-
|
||||||
|
- return StratisBlockdevInfo(path=properties["Devnode"], uuid=blockdev_uuid,
|
||||||
|
- pool_name=pool_name, pool_uuid=pool_info.uuid,
|
||||||
|
- object_path=blockdev_path)
|
||||||
|
+ return StratisBlockdevInfo(path=properties["Devnode"], uuid=blockdev_uuid,
|
||||||
|
+ pool_name=pool_name, pool_uuid=pool_info.uuid,
|
||||||
|
+ object_path=blockdev_path)
|
||||||
|
|
||||||
|
def _get_locked_pools_info(self):
|
||||||
|
locked_pools = []
|
||||||
|
--
|
||||||
|
2.38.1
|
||||||
|
|
@ -1,240 +0,0 @@
|
|||||||
From 3b9a781e138830a190d16c8dd970b800a086de46 Mon Sep 17 00:00:00 2001
|
|
||||||
From: Vojtech Trefny <vtrefny@redhat.com>
|
|
||||||
Date: Wed, 4 Aug 2021 13:00:53 +0200
|
|
||||||
Subject: [PATCH 1/3] edd_test: Locate the edd_data based on the test file
|
|
||||||
location
|
|
||||||
|
|
||||||
We can't use the blivet.edd module location when running tests
|
|
||||||
against installed version of blivet.
|
|
||||||
---
|
|
||||||
tests/devicelibs_test/edd_test.py | 5 ++---
|
|
||||||
1 file changed, 2 insertions(+), 3 deletions(-)
|
|
||||||
|
|
||||||
diff --git a/tests/devicelibs_test/edd_test.py b/tests/devicelibs_test/edd_test.py
|
|
||||||
index 23d736f4..7ec8d1e6 100644
|
|
||||||
--- a/tests/devicelibs_test/edd_test.py
|
|
||||||
+++ b/tests/devicelibs_test/edd_test.py
|
|
||||||
@@ -1,7 +1,6 @@
|
|
||||||
import unittest
|
|
||||||
import mock
|
|
||||||
import os
|
|
||||||
-import inspect
|
|
||||||
import logging
|
|
||||||
import copy
|
|
||||||
|
|
||||||
@@ -110,9 +109,9 @@ class EddTestCase(unittest.TestCase):
|
|
||||||
name = name[:-1]
|
|
||||||
if name.startswith("/"):
|
|
||||||
name = name[1:]
|
|
||||||
- dirname = os.path.dirname(inspect.getfile(edd))
|
|
||||||
+ dirname = os.path.abspath(os.path.dirname(__file__))
|
|
||||||
return os.path.join(dirname,
|
|
||||||
- "../../tests/devicelibs_test/edd_data/",
|
|
||||||
+ "edd_data/",
|
|
||||||
name)
|
|
||||||
|
|
||||||
def edd_debug(self, *args):
|
|
||||||
--
|
|
||||||
2.31.1
|
|
||||||
|
|
||||||
|
|
||||||
From 7ad3824fceb98e2741820b76a9cfea5add338343 Mon Sep 17 00:00:00 2001
|
|
||||||
From: Vojtech Trefny <vtrefny@redhat.com>
|
|
||||||
Date: Wed, 4 Aug 2021 13:02:08 +0200
|
|
||||||
Subject: [PATCH 2/3] tests: Allow running tests without the tests directory in
|
|
||||||
PYTHONPATH
|
|
||||||
|
|
||||||
When running the tests against installed version of blivet, the
|
|
||||||
"tests" directory is not in PYTHONPATH so we need to import all
|
|
||||||
helper modules using relative path.
|
|
||||||
---
|
|
||||||
tests/action_test.py | 2 +-
|
|
||||||
tests/devicelibs_test/edd_test.py | 2 +-
|
|
||||||
tests/{ => devicelibs_test}/lib.py | 0
|
|
||||||
tests/formats_test/fs_test.py | 2 +-
|
|
||||||
tests/formats_test/fslabeling.py | 2 +-
|
|
||||||
tests/formats_test/fstesting.py | 2 +-
|
|
||||||
tests/formats_test/fsuuid.py | 2 +-
|
|
||||||
tests/formats_test/labeling_test.py | 2 +-
|
|
||||||
tests/{ => formats_test}/loopbackedtestcase.py | 0
|
|
||||||
tests/formats_test/luks_test.py | 2 +-
|
|
||||||
tests/formats_test/lvmpv_test.py | 2 +-
|
|
||||||
tests/partitioning_test.py | 2 +-
|
|
||||||
12 files changed, 10 insertions(+), 10 deletions(-)
|
|
||||||
rename tests/{ => devicelibs_test}/lib.py (100%)
|
|
||||||
rename tests/{ => formats_test}/loopbackedtestcase.py (100%)
|
|
||||||
|
|
||||||
diff --git a/tests/action_test.py b/tests/action_test.py
|
|
||||||
index 38a2e872..1e84c20b 100644
|
|
||||||
--- a/tests/action_test.py
|
|
||||||
+++ b/tests/action_test.py
|
|
||||||
@@ -6,7 +6,7 @@ if PY3:
|
|
||||||
else:
|
|
||||||
from mock import Mock
|
|
||||||
|
|
||||||
-from tests.storagetestcase import StorageTestCase
|
|
||||||
+from storagetestcase import StorageTestCase
|
|
||||||
import blivet
|
|
||||||
from blivet.formats import get_format
|
|
||||||
from blivet.size import Size
|
|
||||||
diff --git a/tests/devicelibs_test/edd_test.py b/tests/devicelibs_test/edd_test.py
|
|
||||||
index 7ec8d1e6..379c7aeb 100644
|
|
||||||
--- a/tests/devicelibs_test/edd_test.py
|
|
||||||
+++ b/tests/devicelibs_test/edd_test.py
|
|
||||||
@@ -6,7 +6,7 @@ import copy
|
|
||||||
|
|
||||||
from blivet import arch
|
|
||||||
from blivet.devicelibs import edd
|
|
||||||
-from tests import lib
|
|
||||||
+from . import lib
|
|
||||||
|
|
||||||
|
|
||||||
class FakeDevice(object):
|
|
||||||
diff --git a/tests/lib.py b/tests/devicelibs_test/lib.py
|
|
||||||
similarity index 100%
|
|
||||||
rename from tests/lib.py
|
|
||||||
rename to tests/devicelibs_test/lib.py
|
|
||||||
diff --git a/tests/formats_test/fs_test.py b/tests/formats_test/fs_test.py
|
|
||||||
index ab3499a7..bd643370 100644
|
|
||||||
--- a/tests/formats_test/fs_test.py
|
|
||||||
+++ b/tests/formats_test/fs_test.py
|
|
||||||
@@ -10,7 +10,7 @@ from blivet.errors import DeviceFormatError
|
|
||||||
from blivet.formats import get_format
|
|
||||||
from blivet.devices import PartitionDevice, DiskDevice
|
|
||||||
|
|
||||||
-from tests import loopbackedtestcase
|
|
||||||
+from . import loopbackedtestcase
|
|
||||||
|
|
||||||
from . import fstesting
|
|
||||||
|
|
||||||
diff --git a/tests/formats_test/fslabeling.py b/tests/formats_test/fslabeling.py
|
|
||||||
index fbb28eee..0e0dc261 100644
|
|
||||||
--- a/tests/formats_test/fslabeling.py
|
|
||||||
+++ b/tests/formats_test/fslabeling.py
|
|
||||||
@@ -2,7 +2,7 @@
|
|
||||||
import abc
|
|
||||||
import six
|
|
||||||
|
|
||||||
-from tests import loopbackedtestcase
|
|
||||||
+from . import loopbackedtestcase
|
|
||||||
from blivet.errors import FSError, FSReadLabelError
|
|
||||||
from blivet.size import Size
|
|
||||||
|
|
||||||
diff --git a/tests/formats_test/fstesting.py b/tests/formats_test/fstesting.py
|
|
||||||
index 86b2a116..e34584d8 100644
|
|
||||||
--- a/tests/formats_test/fstesting.py
|
|
||||||
+++ b/tests/formats_test/fstesting.py
|
|
||||||
@@ -5,7 +5,7 @@ from six import add_metaclass
|
|
||||||
import os
|
|
||||||
import tempfile
|
|
||||||
|
|
||||||
-from tests import loopbackedtestcase
|
|
||||||
+from . import loopbackedtestcase
|
|
||||||
from blivet.errors import FSError, FSResizeError, DeviceFormatError
|
|
||||||
from blivet.size import Size, ROUND_DOWN
|
|
||||||
from blivet.formats import fs
|
|
||||||
diff --git a/tests/formats_test/fsuuid.py b/tests/formats_test/fsuuid.py
|
|
||||||
index c8003945..16aa19a6 100644
|
|
||||||
--- a/tests/formats_test/fsuuid.py
|
|
||||||
+++ b/tests/formats_test/fsuuid.py
|
|
||||||
@@ -3,7 +3,7 @@ import abc
|
|
||||||
import six
|
|
||||||
from unittest import skipIf
|
|
||||||
|
|
||||||
-from tests import loopbackedtestcase
|
|
||||||
+from . import loopbackedtestcase
|
|
||||||
from blivet.errors import FSError, FSWriteUUIDError
|
|
||||||
from blivet.size import Size
|
|
||||||
from blivet.util import capture_output
|
|
||||||
diff --git a/tests/formats_test/labeling_test.py b/tests/formats_test/labeling_test.py
|
|
||||||
index e26cb7df..d24e6619 100644
|
|
||||||
--- a/tests/formats_test/labeling_test.py
|
|
||||||
+++ b/tests/formats_test/labeling_test.py
|
|
||||||
@@ -1,10 +1,10 @@
|
|
||||||
import unittest
|
|
||||||
|
|
||||||
-from tests import loopbackedtestcase
|
|
||||||
from blivet.formats import device_formats
|
|
||||||
import blivet.formats.fs as fs
|
|
||||||
import blivet.formats.swap as swap
|
|
||||||
|
|
||||||
+from . import loopbackedtestcase
|
|
||||||
from . import fslabeling
|
|
||||||
|
|
||||||
|
|
||||||
diff --git a/tests/loopbackedtestcase.py b/tests/formats_test/loopbackedtestcase.py
|
|
||||||
similarity index 100%
|
|
||||||
rename from tests/loopbackedtestcase.py
|
|
||||||
rename to tests/formats_test/loopbackedtestcase.py
|
|
||||||
diff --git a/tests/formats_test/luks_test.py b/tests/formats_test/luks_test.py
|
|
||||||
index be0d50b0..5423ebdf 100644
|
|
||||||
--- a/tests/formats_test/luks_test.py
|
|
||||||
+++ b/tests/formats_test/luks_test.py
|
|
||||||
@@ -7,7 +7,7 @@ from blivet.formats.luks import LUKS
|
|
||||||
|
|
||||||
from blivet.size import Size
|
|
||||||
|
|
||||||
-from tests import loopbackedtestcase
|
|
||||||
+from . import loopbackedtestcase
|
|
||||||
|
|
||||||
|
|
||||||
class LUKSTestCase(loopbackedtestcase.LoopBackedTestCase):
|
|
||||||
diff --git a/tests/formats_test/lvmpv_test.py b/tests/formats_test/lvmpv_test.py
|
|
||||||
index 792a2f1d..da7270d9 100644
|
|
||||||
--- a/tests/formats_test/lvmpv_test.py
|
|
||||||
+++ b/tests/formats_test/lvmpv_test.py
|
|
||||||
@@ -4,7 +4,7 @@ from blivet.formats.lvmpv import LVMPhysicalVolume
|
|
||||||
|
|
||||||
from blivet.size import Size
|
|
||||||
|
|
||||||
-from tests import loopbackedtestcase
|
|
||||||
+from . import loopbackedtestcase
|
|
||||||
|
|
||||||
|
|
||||||
class LVMPVTestCase(loopbackedtestcase.LoopBackedTestCase):
|
|
||||||
diff --git a/tests/partitioning_test.py b/tests/partitioning_test.py
|
|
||||||
index b7aa5045..a713aaa1 100644
|
|
||||||
--- a/tests/partitioning_test.py
|
|
||||||
+++ b/tests/partitioning_test.py
|
|
||||||
@@ -29,7 +29,7 @@ from blivet.devices.lvm import LVMCacheRequest
|
|
||||||
|
|
||||||
from blivet.errors import PartitioningError
|
|
||||||
|
|
||||||
-from tests.imagebackedtestcase import ImageBackedTestCase
|
|
||||||
+from imagebackedtestcase import ImageBackedTestCase
|
|
||||||
from blivet.blivet import Blivet
|
|
||||||
from blivet.util import sparsetmpfile
|
|
||||||
from blivet.formats import get_format
|
|
||||||
--
|
|
||||||
2.31.1
|
|
||||||
|
|
||||||
|
|
||||||
From 9ee41c8b60c56ce752e305be73001c7089f43011 Mon Sep 17 00:00:00 2001
|
|
||||||
From: Vojtech Trefny <vtrefny@redhat.com>
|
|
||||||
Date: Fri, 6 Aug 2021 14:51:01 +0200
|
|
||||||
Subject: [PATCH 3/3] tests: Print version and blivet location when running
|
|
||||||
tests
|
|
||||||
|
|
||||||
---
|
|
||||||
tests/run_tests.py | 5 +++++
|
|
||||||
1 file changed, 5 insertions(+)
|
|
||||||
|
|
||||||
diff --git a/tests/run_tests.py b/tests/run_tests.py
|
|
||||||
index 32e3f2d3..8ad8b61a 100644
|
|
||||||
--- a/tests/run_tests.py
|
|
||||||
+++ b/tests/run_tests.py
|
|
||||||
@@ -32,6 +32,11 @@ if __name__ == '__main__':
|
|
||||||
|
|
||||||
testdir = os.path.abspath(os.path.dirname(__file__))
|
|
||||||
|
|
||||||
+ import blivet
|
|
||||||
+ print("Running tests with Blivet %s from %s" % (blivet.__version__,
|
|
||||||
+ os.path.abspath(os.path.dirname(blivet.__file__))),
|
|
||||||
+ file=sys.stderr)
|
|
||||||
+
|
|
||||||
if args.testname:
|
|
||||||
for n in args.testname:
|
|
||||||
suite.addTests(unittest.TestLoader().loadTestsFromName(n))
|
|
||||||
--
|
|
||||||
2.31.1
|
|
||||||
|
|
@ -0,0 +1,27 @@
|
|||||||
|
From b747c4ed07937f54a546ffb2f2c8c95e0797dd6c Mon Sep 17 00:00:00 2001
|
||||||
|
From: Vojtech Trefny <vtrefny@redhat.com>
|
||||||
|
Date: Thu, 20 Oct 2022 15:19:29 +0200
|
||||||
|
Subject: [PATCH] tests: Skip XFS resize test on CentOS/RHEL 8
|
||||||
|
|
||||||
|
Partitions on loop devices are broken on CentOS/RHEL 8.
|
||||||
|
---
|
||||||
|
tests/skip.yml | 6 ++++++
|
||||||
|
1 file changed, 6 insertions(+)
|
||||||
|
|
||||||
|
diff --git a/tests/skip.yml b/tests/skip.yml
|
||||||
|
index 568c3fff..66b34493 100644
|
||||||
|
--- a/tests/skip.yml
|
||||||
|
+++ b/tests/skip.yml
|
||||||
|
@@ -29,3 +29,9 @@
|
||||||
|
- distro: "centos"
|
||||||
|
version: "9"
|
||||||
|
reason: "Creating RAID 1 LV on CentOS/RHEL 9 causes a system deadlock"
|
||||||
|
+
|
||||||
|
+- test: storage_tests.formats_test.fs_test.XFSTestCase.test_resize
|
||||||
|
+ skip_on:
|
||||||
|
+ - distro: ["centos", "enterprise_linux"]
|
||||||
|
+ version: "8"
|
||||||
|
+ reason: "Creating partitions on loop devices is broken on CentOS/RHEL 8 latest kernel"
|
||||||
|
--
|
||||||
|
2.37.3
|
||||||
|
|
@ -1,65 +0,0 @@
|
|||||||
From 46335861073882b7162221fc0995dc1df3c67749 Mon Sep 17 00:00:00 2001
|
|
||||||
From: Vojtech Trefny <vtrefny@redhat.com>
|
|
||||||
Date: Fri, 6 Aug 2021 16:37:51 +0200
|
|
||||||
Subject: [PATCH] Improve error message printed for missing dependecies
|
|
||||||
|
|
||||||
The existing error message can be confusing for people that don't
|
|
||||||
know internals of blivet and libblockdev and the information what
|
|
||||||
is actually broken or not installed on the system is missing
|
|
||||||
completely. Example for LVM VDO with missing kvdo module:
|
|
||||||
|
|
||||||
Before:
|
|
||||||
|
|
||||||
device type lvmvdopool requires unavailable_dependencies:
|
|
||||||
libblockdev lvm plugin (vdo technology)
|
|
||||||
|
|
||||||
After:
|
|
||||||
|
|
||||||
device type lvmvdopool requires unavailable_dependencies:
|
|
||||||
libblockdev lvm plugin (vdo technology):
|
|
||||||
libblockdev plugin lvm is loaded but some required technologies
|
|
||||||
are not available (BD_LVM_TECH_VDO: Kernel module 'kvdo' not
|
|
||||||
available)
|
|
||||||
---
|
|
||||||
blivet/deviceaction.py | 2 +-
|
|
||||||
blivet/tasks/availability.py | 4 ++--
|
|
||||||
2 files changed, 3 insertions(+), 3 deletions(-)
|
|
||||||
|
|
||||||
diff --git a/blivet/deviceaction.py b/blivet/deviceaction.py
|
|
||||||
index 56e29215..0458e4be 100644
|
|
||||||
--- a/blivet/deviceaction.py
|
|
||||||
+++ b/blivet/deviceaction.py
|
|
||||||
@@ -173,7 +173,7 @@ class DeviceAction(util.ObjectID):
|
|
||||||
def _check_device_dependencies(self):
|
|
||||||
unavailable_dependencies = self.device.unavailable_dependencies
|
|
||||||
if unavailable_dependencies:
|
|
||||||
- dependencies_str = ", ".join(str(d) for d in unavailable_dependencies)
|
|
||||||
+ dependencies_str = ", ".join("%s:\n%s" % (str(d), ", ".join(d.availability_errors)) for d in unavailable_dependencies)
|
|
||||||
raise DependencyError("device type %s requires unavailable_dependencies: %s" % (self.device.type, dependencies_str))
|
|
||||||
|
|
||||||
def apply(self):
|
|
||||||
diff --git a/blivet/tasks/availability.py b/blivet/tasks/availability.py
|
|
||||||
index 1fd80590..1537f3f5 100644
|
|
||||||
--- a/blivet/tasks/availability.py
|
|
||||||
+++ b/blivet/tasks/availability.py
|
|
||||||
@@ -224,7 +224,7 @@ class BlockDevMethod(Method):
|
|
||||||
try:
|
|
||||||
self._tech_info.check_fn(tech, mode)
|
|
||||||
except GLib.GError as e:
|
|
||||||
- errors.append(str(e))
|
|
||||||
+ errors.append("%s: %s" % (tech.value_name, e.message))
|
|
||||||
return errors
|
|
||||||
|
|
||||||
def availability_errors(self, resource):
|
|
||||||
@@ -242,7 +242,7 @@ class BlockDevMethod(Method):
|
|
||||||
tech_missing = self._check_technologies()
|
|
||||||
if tech_missing:
|
|
||||||
return ["libblockdev plugin %s is loaded but some required "
|
|
||||||
- "technologies are not available:\n%s" % (self._tech_info.plugin_name, tech_missing)]
|
|
||||||
+ "technologies are not available (%s)" % (self._tech_info.plugin_name, "; ".join(tech_missing))]
|
|
||||||
else:
|
|
||||||
return []
|
|
||||||
|
|
||||||
--
|
|
||||||
2.31.1
|
|
||||||
|
|
160
SOURCES/0009-Revert-Adjust-to-new-XFS-min-size.patch
Normal file
160
SOURCES/0009-Revert-Adjust-to-new-XFS-min-size.patch
Normal file
@ -0,0 +1,160 @@
|
|||||||
|
From 9618b84f94187efddc7316c2546bed923a91ecf9 Mon Sep 17 00:00:00 2001
|
||||||
|
From: Vojtech Trefny <vtrefny@redhat.com>
|
||||||
|
Date: Thu, 3 Nov 2022 08:36:27 +0100
|
||||||
|
Subject: [PATCH 1/2] Revert "Set XFS minimal size to 300 MiB"
|
||||||
|
|
||||||
|
This reverts commit 307d49833771d161314bae50c68e70dc35c3bb36.
|
||||||
|
---
|
||||||
|
blivet/formats/fs.py | 2 +-
|
||||||
|
1 file changed, 1 insertion(+), 1 deletion(-)
|
||||||
|
|
||||||
|
diff --git a/blivet/formats/fs.py b/blivet/formats/fs.py
|
||||||
|
index 8c346aa5..33922f3a 100644
|
||||||
|
--- a/blivet/formats/fs.py
|
||||||
|
+++ b/blivet/formats/fs.py
|
||||||
|
@@ -1091,7 +1091,7 @@ class XFS(FS):
|
||||||
|
_modules = ["xfs"]
|
||||||
|
_labelfs = fslabeling.XFSLabeling()
|
||||||
|
_uuidfs = fsuuid.XFSUUID()
|
||||||
|
- _min_size = Size("300 MiB")
|
||||||
|
+ _min_size = Size("16 MiB")
|
||||||
|
_max_size = Size("16 EiB")
|
||||||
|
_formattable = True
|
||||||
|
_linux_native = True
|
||||||
|
--
|
||||||
|
2.38.1
|
||||||
|
|
||||||
|
|
||||||
|
From 24d94922d6879baa85aaa101f6b21efa568a9cbc Mon Sep 17 00:00:00 2001
|
||||||
|
From: Vojtech Trefny <vtrefny@redhat.com>
|
||||||
|
Date: Thu, 3 Nov 2022 08:36:39 +0100
|
||||||
|
Subject: [PATCH 2/2] Revert "tests: Create bigger devices for XFS testing"
|
||||||
|
|
||||||
|
This reverts commit 467cb8024010b2cabb1e92d9e64f6d3cbe949ad9.
|
||||||
|
---
|
||||||
|
tests/storage_tests/formats_test/fs_test.py | 7 +++----
|
||||||
|
tests/storage_tests/formats_test/fslabeling.py | 4 +---
|
||||||
|
tests/storage_tests/formats_test/fsuuid.py | 4 +---
|
||||||
|
tests/storage_tests/formats_test/labeling_test.py | 2 --
|
||||||
|
tests/storage_tests/formats_test/uuid_test.py | 3 ---
|
||||||
|
5 files changed, 5 insertions(+), 15 deletions(-)
|
||||||
|
|
||||||
|
diff --git a/tests/storage_tests/formats_test/fs_test.py b/tests/storage_tests/formats_test/fs_test.py
|
||||||
|
index cf8fb441..97f4cbbe 100644
|
||||||
|
--- a/tests/storage_tests/formats_test/fs_test.py
|
||||||
|
+++ b/tests/storage_tests/formats_test/fs_test.py
|
||||||
|
@@ -54,7 +54,6 @@ class ReiserFSTestCase(fstesting.FSAsRoot):
|
||||||
|
|
||||||
|
class XFSTestCase(fstesting.FSAsRoot):
|
||||||
|
_fs_class = fs.XFS
|
||||||
|
- _DEVICE_SIZE = Size("500 MiB")
|
||||||
|
|
||||||
|
def can_resize(self, an_fs):
|
||||||
|
resize_tasks = (an_fs._resize, an_fs._size_info)
|
||||||
|
@@ -96,12 +95,12 @@ class XFSTestCase(fstesting.FSAsRoot):
|
||||||
|
self.assertFalse(an_fs.resizable)
|
||||||
|
# Not resizable, so can not do resizing actions.
|
||||||
|
with self.assertRaises(DeviceFormatError):
|
||||||
|
- an_fs.target_size = Size("300 MiB")
|
||||||
|
+ an_fs.target_size = Size("64 MiB")
|
||||||
|
with self.assertRaises(DeviceFormatError):
|
||||||
|
an_fs.do_resize()
|
||||||
|
else:
|
||||||
|
disk = DiskDevice(os.path.basename(self.loop_devices[0]))
|
||||||
|
- part = self._create_partition(disk, Size("300 MiB"))
|
||||||
|
+ part = self._create_partition(disk, Size("50 MiB"))
|
||||||
|
an_fs = self._fs_class()
|
||||||
|
an_fs.device = part.path
|
||||||
|
self.assertIsNone(an_fs.create())
|
||||||
|
@@ -114,7 +113,7 @@ class XFSTestCase(fstesting.FSAsRoot):
|
||||||
|
part = self._create_partition(disk, size=part.size + Size("40 MiB"))
|
||||||
|
|
||||||
|
# Try a reasonable target size
|
||||||
|
- TARGET_SIZE = Size("325 MiB")
|
||||||
|
+ TARGET_SIZE = Size("64 MiB")
|
||||||
|
an_fs.target_size = TARGET_SIZE
|
||||||
|
self.assertEqual(an_fs.target_size, TARGET_SIZE)
|
||||||
|
self.assertNotEqual(an_fs._size, TARGET_SIZE)
|
||||||
|
diff --git a/tests/storage_tests/formats_test/fslabeling.py b/tests/storage_tests/formats_test/fslabeling.py
|
||||||
|
index ebe0b70a..0e0dc261 100644
|
||||||
|
--- a/tests/storage_tests/formats_test/fslabeling.py
|
||||||
|
+++ b/tests/storage_tests/formats_test/fslabeling.py
|
||||||
|
@@ -21,10 +21,8 @@ class LabelingAsRoot(loopbackedtestcase.LoopBackedTestCase):
|
||||||
|
_invalid_label = abc.abstractproperty(
|
||||||
|
doc="A label which is invalid for this filesystem.")
|
||||||
|
|
||||||
|
- _DEVICE_SIZE = Size("100 MiB")
|
||||||
|
-
|
||||||
|
def __init__(self, methodName='run_test'):
|
||||||
|
- super(LabelingAsRoot, self).__init__(methodName=methodName, device_spec=[self._DEVICE_SIZE])
|
||||||
|
+ super(LabelingAsRoot, self).__init__(methodName=methodName, device_spec=[Size("100 MiB")])
|
||||||
|
|
||||||
|
def setUp(self):
|
||||||
|
an_fs = self._fs_class()
|
||||||
|
diff --git a/tests/storage_tests/formats_test/fsuuid.py b/tests/storage_tests/formats_test/fsuuid.py
|
||||||
|
index 0b9762fd..16aa19a6 100644
|
||||||
|
--- a/tests/storage_tests/formats_test/fsuuid.py
|
||||||
|
+++ b/tests/storage_tests/formats_test/fsuuid.py
|
||||||
|
@@ -23,11 +23,9 @@ class SetUUID(loopbackedtestcase.LoopBackedTestCase):
|
||||||
|
_invalid_uuid = abc.abstractproperty(
|
||||||
|
doc="An invalid UUID for this filesystem.")
|
||||||
|
|
||||||
|
- _DEVICE_SIZE = Size("100 MiB")
|
||||||
|
-
|
||||||
|
def __init__(self, methodName='run_test'):
|
||||||
|
super(SetUUID, self).__init__(methodName=methodName,
|
||||||
|
- device_spec=[self._DEVICE_SIZE])
|
||||||
|
+ device_spec=[Size("100 MiB")])
|
||||||
|
|
||||||
|
def setUp(self):
|
||||||
|
an_fs = self._fs_class()
|
||||||
|
diff --git a/tests/storage_tests/formats_test/labeling_test.py b/tests/storage_tests/formats_test/labeling_test.py
|
||||||
|
index 0702260a..d24e6619 100644
|
||||||
|
--- a/tests/storage_tests/formats_test/labeling_test.py
|
||||||
|
+++ b/tests/storage_tests/formats_test/labeling_test.py
|
||||||
|
@@ -1,7 +1,6 @@
|
||||||
|
import unittest
|
||||||
|
|
||||||
|
from blivet.formats import device_formats
|
||||||
|
-from blivet.size import Size
|
||||||
|
import blivet.formats.fs as fs
|
||||||
|
import blivet.formats.swap as swap
|
||||||
|
|
||||||
|
@@ -62,7 +61,6 @@ class InitializationTestCase(unittest.TestCase):
|
||||||
|
class XFSTestCase(fslabeling.CompleteLabelingAsRoot):
|
||||||
|
_fs_class = fs.XFS
|
||||||
|
_invalid_label = "root filesystem"
|
||||||
|
- _DEVICE_SIZE = Size("500 MiB")
|
||||||
|
|
||||||
|
|
||||||
|
class FATFSTestCase(fslabeling.CompleteLabelingAsRoot):
|
||||||
|
diff --git a/tests/storage_tests/formats_test/uuid_test.py b/tests/storage_tests/formats_test/uuid_test.py
|
||||||
|
index af35c0ee..ee8d452e 100644
|
||||||
|
--- a/tests/storage_tests/formats_test/uuid_test.py
|
||||||
|
+++ b/tests/storage_tests/formats_test/uuid_test.py
|
||||||
|
@@ -2,7 +2,6 @@ import unittest
|
||||||
|
|
||||||
|
import blivet.formats.fs as fs
|
||||||
|
import blivet.formats.swap as swap
|
||||||
|
-from blivet.size import Size
|
||||||
|
|
||||||
|
from . import fsuuid
|
||||||
|
|
||||||
|
@@ -53,14 +52,12 @@ class XFSTestCase(fsuuid.SetUUIDWithMkFs):
|
||||||
|
_fs_class = fs.XFS
|
||||||
|
_invalid_uuid = "abcdefgh-ijkl-mnop-qrst-uvwxyz123456"
|
||||||
|
_valid_uuid = "97e3d40f-dca8-497d-8b86-92f257402465"
|
||||||
|
- _DEVICE_SIZE = Size("500 MiB")
|
||||||
|
|
||||||
|
|
||||||
|
class XFSAfterTestCase(fsuuid.SetUUIDAfterMkFs):
|
||||||
|
_fs_class = fs.XFS
|
||||||
|
_invalid_uuid = "abcdefgh-ijkl-mnop-qrst-uvwxyz123456"
|
||||||
|
_valid_uuid = "97e3d40f-dca8-497d-8b86-92f257402465"
|
||||||
|
- _DEVICE_SIZE = Size("500 MiB")
|
||||||
|
|
||||||
|
|
||||||
|
class FATFSTestCase(fsuuid.SetUUIDWithMkFs):
|
||||||
|
--
|
||||||
|
2.38.1
|
||||||
|
|
@ -0,0 +1,55 @@
|
|||||||
|
From fed62af06eb1584adbacd821dfe79c2df52c6aa4 Mon Sep 17 00:00:00 2001
|
||||||
|
From: Vojtech Trefny <vtrefny@redhat.com>
|
||||||
|
Date: Wed, 2 Nov 2022 12:14:28 +0100
|
||||||
|
Subject: [PATCH] Catch BlockDevNotImplementedError for btrfs plugin calls
|
||||||
|
|
||||||
|
This is a workaround for RHEL where the btrfs plugin is not
|
||||||
|
available and where we might still try to call some libblockdev
|
||||||
|
functions to gather information about preexisting btrfs devices.
|
||||||
|
---
|
||||||
|
blivet/devices/btrfs.py | 8 ++++----
|
||||||
|
1 file changed, 4 insertions(+), 4 deletions(-)
|
||||||
|
|
||||||
|
diff --git a/blivet/devices/btrfs.py b/blivet/devices/btrfs.py
|
||||||
|
index 0e029715..1ae6a04d 100644
|
||||||
|
--- a/blivet/devices/btrfs.py
|
||||||
|
+++ b/blivet/devices/btrfs.py
|
||||||
|
@@ -362,7 +362,7 @@ class BTRFSVolumeDevice(BTRFSDevice, ContainerDevice, RaidDevice):
|
||||||
|
try:
|
||||||
|
subvols = blockdev.btrfs.list_subvolumes(mountpoint,
|
||||||
|
snapshots_only=snapshots_only)
|
||||||
|
- except blockdev.BtrfsError as e:
|
||||||
|
+ except (blockdev.BtrfsError, blockdev.BlockDevNotImplementedError) as e:
|
||||||
|
log.debug("failed to list subvolumes: %s", e)
|
||||||
|
else:
|
||||||
|
self._get_default_subvolume_id()
|
||||||
|
@@ -400,7 +400,7 @@ class BTRFSVolumeDevice(BTRFSDevice, ContainerDevice, RaidDevice):
|
||||||
|
with self._do_temp_mount() as mountpoint:
|
||||||
|
try:
|
||||||
|
subvolid = blockdev.btrfs.get_default_subvolume_id(mountpoint)
|
||||||
|
- except blockdev.BtrfsError as e:
|
||||||
|
+ except (blockdev.BtrfsError, blockdev.BlockDevNotImplementedError) as e:
|
||||||
|
log.debug("failed to get default subvolume id: %s", e)
|
||||||
|
|
||||||
|
self._default_subvolume_id = subvolid
|
||||||
|
@@ -413,7 +413,7 @@ class BTRFSVolumeDevice(BTRFSDevice, ContainerDevice, RaidDevice):
|
||||||
|
with self._do_temp_mount() as mountpoint:
|
||||||
|
try:
|
||||||
|
blockdev.btrfs.set_default_subvolume(mountpoint, vol_id)
|
||||||
|
- except blockdev.BtrfsError as e:
|
||||||
|
+ except (blockdev.BtrfsError, blockdev.BlockDevNotImplementedError) as e:
|
||||||
|
log.error("failed to set new default subvolume id (%s): %s",
|
||||||
|
vol_id, e)
|
||||||
|
# The only time we set a new default subvolume is so we can remove
|
||||||
|
@@ -471,7 +471,7 @@ class BTRFSVolumeDevice(BTRFSDevice, ContainerDevice, RaidDevice):
|
||||||
|
if not self.format.vol_uuid:
|
||||||
|
try:
|
||||||
|
bd_info = blockdev.btrfs.filesystem_info(self.parents[0].path)
|
||||||
|
- except blockdev.BtrfsError as e:
|
||||||
|
+ except (blockdev.BtrfsError, blockdev.BlockDevNotImplementedError) as e:
|
||||||
|
log.error("failed to get filesystem info for new btrfs volume %s", e)
|
||||||
|
else:
|
||||||
|
self.format.vol_uuid = bd_info.uuid
|
||||||
|
--
|
||||||
|
2.38.1
|
||||||
|
|
@ -1,90 +0,0 @@
|
|||||||
From 06cafbbbbff0aae3634eb2908d25d0dc46c2048b Mon Sep 17 00:00:00 2001
|
|
||||||
From: Vojtech Trefny <vtrefny@redhat.com>
|
|
||||||
Date: Tue, 9 Nov 2021 15:52:48 +0100
|
|
||||||
Subject: [PATCH] Use bigger chunk size for thinpools bigger than ~15.88 TiB
|
|
||||||
|
|
||||||
With our default chunk size of 64 KiB we cannot create bigger
|
|
||||||
thin pools than 15.88 TiB. Unfortunately we need to specify chunk
|
|
||||||
size to be able to calculate thin metadata properly so we can't
|
|
||||||
simply leave this to LVM to determine the correct chunk size.
|
|
||||||
---
|
|
||||||
blivet/devicelibs/lvm.py | 11 +++++++++++
|
|
||||||
blivet/devices/lvm.py | 6 +++---
|
|
||||||
tests/devices_test/lvm_test.py | 11 +++++++++++
|
|
||||||
3 files changed, 25 insertions(+), 3 deletions(-)
|
|
||||||
|
|
||||||
diff --git a/blivet/devicelibs/lvm.py b/blivet/devicelibs/lvm.py
|
|
||||||
index d56a76ed..cb6f655e 100644
|
|
||||||
--- a/blivet/devicelibs/lvm.py
|
|
||||||
+++ b/blivet/devicelibs/lvm.py
|
|
||||||
@@ -20,6 +20,7 @@
|
|
||||||
# Author(s): Dave Lehman <dlehman@redhat.com>
|
|
||||||
#
|
|
||||||
|
|
||||||
+import math
|
|
||||||
import os
|
|
||||||
import re
|
|
||||||
|
|
||||||
@@ -51,6 +52,7 @@ LVM_THINP_MIN_METADATA_SIZE = Size("2 MiB")
|
|
||||||
LVM_THINP_MAX_METADATA_SIZE = Size("16 GiB")
|
|
||||||
LVM_THINP_MIN_CHUNK_SIZE = Size("64 KiB")
|
|
||||||
LVM_THINP_MAX_CHUNK_SIZE = Size("1 GiB")
|
|
||||||
+LVM_THINP_ADDRESSABLE_CHUNK_SIZE = Size("17455015526400 B") # 15.88 TiB
|
|
||||||
|
|
||||||
raid_levels = raid.RAIDLevels(["linear", "striped", "raid1", "raid4", "raid5", "raid6", "raid10"])
|
|
||||||
raid_seg_types = list(itertools.chain.from_iterable([level.names for level in raid_levels if level.name != "linear"]))
|
|
||||||
@@ -225,3 +227,12 @@ def is_lvm_name_valid(name):
|
|
||||||
return False
|
|
||||||
|
|
||||||
return True
|
|
||||||
+
|
|
||||||
+
|
|
||||||
+def recommend_thpool_chunk_size(thpool_size):
|
|
||||||
+ # calculation of the recommended chunk size by LVM is so complicated that we
|
|
||||||
+ # can't really replicate it, but we know that 64 KiB chunk size gives us
|
|
||||||
+ # upper limit of ~15.88 TiB so we will just add 64 KiB to the chunk size
|
|
||||||
+ # for every ~15.88 TiB of thinpool data size
|
|
||||||
+ return min(math.ceil(thpool_size / LVM_THINP_ADDRESSABLE_CHUNK_SIZE) * LVM_THINP_MIN_CHUNK_SIZE,
|
|
||||||
+ LVM_THINP_MAX_CHUNK_SIZE)
|
|
||||||
diff --git a/blivet/devices/lvm.py b/blivet/devices/lvm.py
|
|
||||||
index 51d785d9..c61eeb4b 100644
|
|
||||||
--- a/blivet/devices/lvm.py
|
|
||||||
+++ b/blivet/devices/lvm.py
|
|
||||||
@@ -1634,9 +1634,9 @@ class LVMThinPoolMixin(object):
|
|
||||||
return
|
|
||||||
|
|
||||||
# we need to know chunk size to calculate recommended metadata size
|
|
||||||
- if self._chunk_size == 0:
|
|
||||||
- self._chunk_size = Size(blockdev.LVM_DEFAULT_CHUNK_SIZE)
|
|
||||||
- log.debug("Using default chunk size: %s", self._chunk_size)
|
|
||||||
+ if self._chunk_size == 0 or enforced:
|
|
||||||
+ self._chunk_size = lvm.recommend_thpool_chunk_size(self._size)
|
|
||||||
+ log.debug("Using recommended chunk size: %s", self._chunk_size)
|
|
||||||
|
|
||||||
old_md_size = self._metadata_size
|
|
||||||
old_pmspare_size = self.vg.pmspare_size
|
|
||||||
diff --git a/tests/devices_test/lvm_test.py b/tests/devices_test/lvm_test.py
|
|
||||||
index 4156d0bf..336c5b99 100644
|
|
||||||
--- a/tests/devices_test/lvm_test.py
|
|
||||||
+++ b/tests/devices_test/lvm_test.py
|
|
||||||
@@ -442,6 +442,17 @@ class LVMDeviceTest(unittest.TestCase):
|
|
||||||
self.assertFalse(pool.exists)
|
|
||||||
self.assertTrue(lvm.lvremove.called)
|
|
||||||
|
|
||||||
+ def test_lvmthinpool_chunk_size(self):
|
|
||||||
+ pv = StorageDevice("pv1", fmt=blivet.formats.get_format("lvmpv"),
|
|
||||||
+ size=Size("100 TiB"))
|
|
||||||
+ vg = LVMVolumeGroupDevice("testvg", parents=[pv])
|
|
||||||
+ pool = LVMLogicalVolumeDevice("pool1", parents=[vg], size=Size("500 MiB"), seg_type="thin-pool")
|
|
||||||
+ self.assertEqual(pool.chunk_size, Size("64 KiB"))
|
|
||||||
+
|
|
||||||
+ pool.size = Size("16 TiB")
|
|
||||||
+ pool.autoset_md_size(enforced=True)
|
|
||||||
+ self.assertEqual(pool.chunk_size, Size("128 KiB"))
|
|
||||||
+
|
|
||||||
|
|
||||||
class TypeSpecificCallsTest(unittest.TestCase):
|
|
||||||
def test_type_specific_calls(self):
|
|
||||||
--
|
|
||||||
2.31.1
|
|
||||||
|
|
@ -0,0 +1,57 @@
|
|||||||
|
From 2aba050e74dc5df483da022dcf436b101c7a4301 Mon Sep 17 00:00:00 2001
|
||||||
|
From: Vojtech Trefny <vtrefny@redhat.com>
|
||||||
|
Date: Wed, 11 Jan 2023 14:59:24 +0100
|
||||||
|
Subject: [PATCH] Default to encryption sector size 512 for LUKS devices
|
||||||
|
|
||||||
|
We are currently letting cryptsetup decide the optimal encryption
|
||||||
|
sector size for LUKS. The problem is that for disks with physical
|
||||||
|
sector size 4096 cryptsetup will default to 4096 encryption sector
|
||||||
|
size even if the drive logical sector size is 512 which means
|
||||||
|
these disks cannot be combined with other 512 logical sector size
|
||||||
|
disks in LVM. This requires a more sophisticated solution in the
|
||||||
|
future, but for now just default to 512 if not specified by the
|
||||||
|
user otherwise.
|
||||||
|
|
||||||
|
Resolves: rhbz#2103800
|
||||||
|
---
|
||||||
|
blivet/formats/luks.py | 10 +++++++---
|
||||||
|
tests/unit_tests/formats_tests/luks_test.py | 2 +-
|
||||||
|
2 files changed, 8 insertions(+), 4 deletions(-)
|
||||||
|
|
||||||
|
diff --git a/blivet/formats/luks.py b/blivet/formats/luks.py
|
||||||
|
index 8de4911f..2637e0c5 100644
|
||||||
|
--- a/blivet/formats/luks.py
|
||||||
|
+++ b/blivet/formats/luks.py
|
||||||
|
@@ -166,9 +166,13 @@ class LUKS(DeviceFormat):
|
||||||
|
if self.pbkdf_args.type == "pbkdf2" and self.pbkdf_args.max_memory_kb:
|
||||||
|
log.warning("Memory limit is not used for pbkdf2 and it will be ignored.")
|
||||||
|
|
||||||
|
- self.luks_sector_size = kwargs.get("luks_sector_size") or 0
|
||||||
|
- if self.luks_sector_size and self.luks_version != "luks2":
|
||||||
|
- raise ValueError("Sector size argument is valid only for LUKS version 2.")
|
||||||
|
+ self.luks_sector_size = kwargs.get("luks_sector_size")
|
||||||
|
+ if self.luks_version == "luks2":
|
||||||
|
+ if self.luks_sector_size is None:
|
||||||
|
+ self.luks_sector_size = 512 # XXX we don't want cryptsetup choose automatically here so fallback to 512
|
||||||
|
+ else:
|
||||||
|
+ if self.luks_sector_size:
|
||||||
|
+ raise ValueError("Sector size argument is valid only for LUKS version 2.")
|
||||||
|
|
||||||
|
def __repr__(self):
|
||||||
|
s = DeviceFormat.__repr__(self)
|
||||||
|
diff --git a/tests/unit_tests/formats_tests/luks_test.py b/tests/unit_tests/formats_tests/luks_test.py
|
||||||
|
index 5ae6acfe..ec7b7592 100644
|
||||||
|
--- a/tests/unit_tests/formats_tests/luks_test.py
|
||||||
|
+++ b/tests/unit_tests/formats_tests/luks_test.py
|
||||||
|
@@ -53,7 +53,7 @@ class LUKSNodevTestCase(unittest.TestCase):
|
||||||
|
|
||||||
|
def test_sector_size(self):
|
||||||
|
fmt = LUKS()
|
||||||
|
- self.assertEqual(fmt.luks_sector_size, 0)
|
||||||
|
+ self.assertEqual(fmt.luks_sector_size, 512)
|
||||||
|
|
||||||
|
with self.assertRaises(ValueError):
|
||||||
|
fmt = LUKS(luks_version="luks1", luks_sector_size=4096)
|
||||||
|
--
|
||||||
|
2.39.0
|
||||||
|
|
@ -1,53 +0,0 @@
|
|||||||
From b938e224c41021c19775d8675dc4337f1e10d4e3 Mon Sep 17 00:00:00 2001
|
|
||||||
From: Vojtech Trefny <vtrefny@redhat.com>
|
|
||||||
Date: Wed, 1 Dec 2021 16:28:15 +0100
|
|
||||||
Subject: [PATCH] iscsi: Replace all log_exception_info calls with log.info
|
|
||||||
|
|
||||||
We don't get any useful information from the exception, it's
|
|
||||||
always the same traceback from a failed DBus call and we only use
|
|
||||||
these when a called failed because firmware ISCSI is not supported.
|
|
||||||
The resulting log message also looks like a failure with the
|
|
||||||
traceback logged and not just as a debug information.
|
|
||||||
|
|
||||||
Resolves: rhbz#2028134
|
|
||||||
---
|
|
||||||
blivet/iscsi.py | 9 ++++-----
|
|
||||||
1 file changed, 4 insertions(+), 5 deletions(-)
|
|
||||||
|
|
||||||
diff --git a/blivet/iscsi.py b/blivet/iscsi.py
|
|
||||||
index 5ee2082b..bc77ca62 100644
|
|
||||||
--- a/blivet/iscsi.py
|
|
||||||
+++ b/blivet/iscsi.py
|
|
||||||
@@ -22,7 +22,6 @@ from . import udev
|
|
||||||
from . import util
|
|
||||||
from .flags import flags
|
|
||||||
from .i18n import _
|
|
||||||
-from .storage_log import log_exception_info
|
|
||||||
from . import safe_dbus
|
|
||||||
import os
|
|
||||||
import re
|
|
||||||
@@ -277,8 +276,8 @@ class iSCSI(object):
|
|
||||||
'org.freedesktop.DBus.ObjectManager',
|
|
||||||
'GetManagedObjects',
|
|
||||||
None)[0]
|
|
||||||
- except safe_dbus.DBusCallError:
|
|
||||||
- log_exception_info(log.info, "iscsi: Failed to get active sessions.")
|
|
||||||
+ except safe_dbus.DBusCallError as e:
|
|
||||||
+ log.info("iscsi: Failed to get active sessions: %s", str(e))
|
|
||||||
return []
|
|
||||||
|
|
||||||
sessions = (obj for obj in objects.keys() if re.match(r'.*/iscsi/session[0-9]+$', obj))
|
|
||||||
@@ -302,8 +301,8 @@ class iSCSI(object):
|
|
||||||
args = GLib.Variant("(a{sv})", ([], ))
|
|
||||||
try:
|
|
||||||
found_nodes, _n_nodes = self._call_initiator_method("DiscoverFirmware", args)
|
|
||||||
- except safe_dbus.DBusCallError:
|
|
||||||
- log_exception_info(log.info, "iscsi: No IBFT info found.")
|
|
||||||
+ except safe_dbus.DBusCallError as e:
|
|
||||||
+ log.info("iscsi: No IBFT info found: %s", str(e))
|
|
||||||
# an exception here means there is no ibft firmware, just return
|
|
||||||
return
|
|
||||||
|
|
||||||
--
|
|
||||||
2.31.1
|
|
||||||
|
|
File diff suppressed because it is too large
Load Diff
@ -1,29 +0,0 @@
|
|||||||
From a15c65a5e71f6fd53624bd657ab95b38d37c6f1b Mon Sep 17 00:00:00 2001
|
|
||||||
From: Vojtech Trefny <vtrefny@redhat.com>
|
|
||||||
Date: Mon, 16 Aug 2021 09:50:34 +0200
|
|
||||||
Subject: [PATCH] Fix getting PV info in LVMPhysicalVolume from the cache
|
|
||||||
|
|
||||||
"self.device" is string for formats so accessing "self.device.path"
|
|
||||||
results in an AttributeError.
|
|
||||||
|
|
||||||
Resolves: rhbz#2079220
|
|
||||||
---
|
|
||||||
blivet/formats/lvmpv.py | 2 +-
|
|
||||||
1 file changed, 1 insertion(+), 1 deletion(-)
|
|
||||||
|
|
||||||
diff --git a/blivet/formats/lvmpv.py b/blivet/formats/lvmpv.py
|
|
||||||
index e4182adb..8cfade9f 100644
|
|
||||||
--- a/blivet/formats/lvmpv.py
|
|
||||||
+++ b/blivet/formats/lvmpv.py
|
|
||||||
@@ -170,7 +170,7 @@ class LVMPhysicalVolume(DeviceFormat):
|
|
||||||
if self.exists:
|
|
||||||
# we don't have any actual value, but the PV exists and is
|
|
||||||
# active, we should try to determine it
|
|
||||||
- pv_info = pvs_info.cache.get(self.device.path)
|
|
||||||
+ pv_info = pvs_info.cache.get(self.device)
|
|
||||||
if pv_info is None:
|
|
||||||
log.error("Failed to get free space information for the PV '%s'", self.device)
|
|
||||||
self._free = Size(0)
|
|
||||||
--
|
|
||||||
2.35.3
|
|
||||||
|
|
@ -1,41 +0,0 @@
|
|||||||
From 78eda3d74110dbf9669c3271f7d2fddf962d0381 Mon Sep 17 00:00:00 2001
|
|
||||||
From: Vojtech Trefny <vtrefny@redhat.com>
|
|
||||||
Date: Thu, 28 Apr 2022 14:13:04 +0200
|
|
||||||
Subject: [PATCH] Do not crash when changing disklabel on disks with active
|
|
||||||
devices
|
|
||||||
|
|
||||||
The _find_active_devices_on_action_disks function originally
|
|
||||||
prevented from making any changes on disks with active devices
|
|
||||||
(active LVs, mounted partitions etc.) This was changed in
|
|
||||||
b72e957d2b23444824316331ae21d1c594371e9c and the check currently
|
|
||||||
prevents only reformatting the disklabel on such disks which
|
|
||||||
should be already impossible on disks with an existing partition.
|
|
||||||
|
|
||||||
This change for the 3.4 stable branch keeps the current behaviour
|
|
||||||
where the active devices are teared down when running in installer
|
|
||||||
mode to avoid potential issues with the installer.
|
|
||||||
|
|
||||||
Resolves: rhbz#2078801
|
|
||||||
---
|
|
||||||
blivet/actionlist.py | 5 ++---
|
|
||||||
1 file changed, 2 insertions(+), 3 deletions(-)
|
|
||||||
|
|
||||||
diff --git a/blivet/actionlist.py b/blivet/actionlist.py
|
|
||||||
index d03e32b9..4ec2dbf8 100644
|
|
||||||
--- a/blivet/actionlist.py
|
|
||||||
+++ b/blivet/actionlist.py
|
|
||||||
@@ -211,9 +211,8 @@ class ActionList(object):
|
|
||||||
except StorageError as e:
|
|
||||||
log.info("teardown of %s failed: %s", device.name, e)
|
|
||||||
else:
|
|
||||||
- raise RuntimeError("partitions in use on disks with changes "
|
|
||||||
- "pending: %s" %
|
|
||||||
- ",".join(problematic))
|
|
||||||
+ log.debug("ignoring devices in use on disks with changes: %s",
|
|
||||||
+ ",".join(problematic))
|
|
||||||
|
|
||||||
log.info("resetting parted disks...")
|
|
||||||
for device in devices:
|
|
||||||
--
|
|
||||||
2.35.3
|
|
||||||
|
|
@ -1,65 +0,0 @@
|
|||||||
From 950f51a4cc041fe1b8a98b17e4828857b7423e55 Mon Sep 17 00:00:00 2001
|
|
||||||
From: Vojtech Trefny <vtrefny@redhat.com>
|
|
||||||
Date: Wed, 13 Apr 2022 15:43:45 +0200
|
|
||||||
Subject: [PATCH] ActionDestroyDevice should not obsolete ActionRemoveMember
|
|
||||||
|
|
||||||
If we want to remove a PV from a VG and then remove the PV device,
|
|
||||||
the ActionDestroyDevice must not obsolete the ActionRemoveMember
|
|
||||||
action. Eventhough we are going to remove the device, we still
|
|
||||||
need to call "vgreduce" first.
|
|
||||||
|
|
||||||
Resolves: rhbz#2076958
|
|
||||||
---
|
|
||||||
blivet/deviceaction.py | 10 +++++-----
|
|
||||||
tests/action_test.py | 7 +++++++
|
|
||||||
2 files changed, 12 insertions(+), 5 deletions(-)
|
|
||||||
|
|
||||||
diff --git a/blivet/deviceaction.py b/blivet/deviceaction.py
|
|
||||||
index 0458e4be..78e113bf 100644
|
|
||||||
--- a/blivet/deviceaction.py
|
|
||||||
+++ b/blivet/deviceaction.py
|
|
||||||
@@ -463,8 +463,8 @@ class ActionDestroyDevice(DeviceAction):
|
|
||||||
- obsoletes all actions w/ lower id that act on the same device,
|
|
||||||
including self, if device does not exist
|
|
||||||
|
|
||||||
- - obsoletes all but ActionDestroyFormat actions w/ lower id on the
|
|
||||||
- same device if device exists
|
|
||||||
+ - obsoletes all but ActionDestroyFormat and ActionRemoveMember actions
|
|
||||||
+ w/ lower id on the same device if device exists
|
|
||||||
|
|
||||||
- obsoletes all actions that add a member to this action's
|
|
||||||
(container) device
|
|
||||||
@@ -474,9 +474,9 @@ class ActionDestroyDevice(DeviceAction):
|
|
||||||
if action.device.id == self.device.id:
|
|
||||||
if self.id >= action.id and not self.device.exists:
|
|
||||||
rc = True
|
|
||||||
- elif self.id > action.id and \
|
|
||||||
- self.device.exists and \
|
|
||||||
- not (action.is_destroy and action.is_format):
|
|
||||||
+ elif self.id > action.id and self.device.exists and \
|
|
||||||
+ not ((action.is_destroy and action.is_format) or
|
|
||||||
+ action.is_remove):
|
|
||||||
rc = True
|
|
||||||
elif action.is_add and (action.device == self.device):
|
|
||||||
rc = True
|
|
||||||
diff --git a/tests/action_test.py b/tests/action_test.py
|
|
||||||
index 1e84c20b..b3608047 100644
|
|
||||||
--- a/tests/action_test.py
|
|
||||||
+++ b/tests/action_test.py
|
|
||||||
@@ -1198,6 +1198,13 @@ class DeviceActionTestCase(StorageTestCase):
|
|
||||||
self.assertEqual(create_sdc2.requires(remove_sdc1), False)
|
|
||||||
self.assertEqual(remove_sdc1.requires(create_sdc2), False)
|
|
||||||
|
|
||||||
+ # destroy sdc1, the ActionRemoveMember should not be obsoleted
|
|
||||||
+ sdc1.exists = True
|
|
||||||
+ destroy_sdc1 = ActionDestroyDevice(sdc1)
|
|
||||||
+ destroy_sdc1.apply()
|
|
||||||
+ self.assertFalse(destroy_sdc1.obsoletes(remove_sdc1))
|
|
||||||
+ self.assertTrue(destroy_sdc1.requires(remove_sdc1))
|
|
||||||
+
|
|
||||||
def test_action_sorting(self, *args, **kwargs):
|
|
||||||
""" Verify correct functioning of action sorting. """
|
|
||||||
|
|
||||||
--
|
|
||||||
2.35.3
|
|
||||||
|
|
@ -1,63 +0,0 @@
|
|||||||
From a9cb01f948fa5371b3e6f9282e7af81aec5cb1a8 Mon Sep 17 00:00:00 2001
|
|
||||||
From: Vojtech Trefny <vtrefny@redhat.com>
|
|
||||||
Date: Mon, 2 May 2022 15:30:16 +0200
|
|
||||||
Subject: [PATCH] Correctly set vg_name after adding/removing a PV from a VG
|
|
||||||
|
|
||||||
Without setting the LVMPhysicalVolume.vg_name argument to None
|
|
||||||
after removing the PV from its VG, the PV is still considered
|
|
||||||
active and cannot be removed.
|
|
||||||
|
|
||||||
Resolves: rhbz#2081276
|
|
||||||
---
|
|
||||||
blivet/devices/lvm.py | 3 +++
|
|
||||||
tests/devices_test/lvm_test.py | 13 +++++++++++++
|
|
||||||
2 files changed, 16 insertions(+)
|
|
||||||
|
|
||||||
diff --git a/blivet/devices/lvm.py b/blivet/devices/lvm.py
|
|
||||||
index c61eeb4b..7c78c813 100644
|
|
||||||
--- a/blivet/devices/lvm.py
|
|
||||||
+++ b/blivet/devices/lvm.py
|
|
||||||
@@ -385,6 +385,8 @@ class LVMVolumeGroupDevice(ContainerDevice):
|
|
||||||
if not parent.format.exists:
|
|
||||||
parent.format.free = self._get_pv_usable_space(parent)
|
|
||||||
|
|
||||||
+ parent.format.vg_name = self.name
|
|
||||||
+
|
|
||||||
def _remove_parent(self, parent):
|
|
||||||
# XXX It would be nice to raise an exception if removing this member
|
|
||||||
# would not leave enough space, but the devicefactory relies on it
|
|
||||||
@@ -395,6 +397,7 @@ class LVMVolumeGroupDevice(ContainerDevice):
|
|
||||||
super(LVMVolumeGroupDevice, self)._remove_parent(parent)
|
|
||||||
parent.format.free = None
|
|
||||||
parent.format.container_uuid = None
|
|
||||||
+ parent.format.vg_name = None
|
|
||||||
|
|
||||||
# We can't rely on lvm to tell us about our size, free space, &c
|
|
||||||
# since we could have modifications queued, unless the VG and all of
|
|
||||||
diff --git a/tests/devices_test/lvm_test.py b/tests/devices_test/lvm_test.py
|
|
||||||
index 336c5b99..c349f003 100644
|
|
||||||
--- a/tests/devices_test/lvm_test.py
|
|
||||||
+++ b/tests/devices_test/lvm_test.py
|
|
||||||
@@ -453,6 +453,19 @@ class LVMDeviceTest(unittest.TestCase):
|
|
||||||
pool.autoset_md_size(enforced=True)
|
|
||||||
self.assertEqual(pool.chunk_size, Size("128 KiB"))
|
|
||||||
|
|
||||||
+ def test_add_remove_pv(self):
|
|
||||||
+ pv1 = StorageDevice("pv1", fmt=blivet.formats.get_format("lvmpv"),
|
|
||||||
+ size=Size("1024 MiB"))
|
|
||||||
+ pv2 = StorageDevice("pv2", fmt=blivet.formats.get_format("lvmpv"),
|
|
||||||
+ size=Size("1024 MiB"))
|
|
||||||
+ vg = LVMVolumeGroupDevice("testvg", parents=[pv1])
|
|
||||||
+
|
|
||||||
+ vg._add_parent(pv2)
|
|
||||||
+ self.assertEqual(pv2.format.vg_name, vg.name)
|
|
||||||
+
|
|
||||||
+ vg._remove_parent(pv2)
|
|
||||||
+ self.assertEqual(pv2.format.vg_name, None)
|
|
||||||
+
|
|
||||||
|
|
||||||
class TypeSpecificCallsTest(unittest.TestCase):
|
|
||||||
def test_type_specific_calls(self):
|
|
||||||
--
|
|
||||||
2.35.3
|
|
||||||
|
|
@ -1,29 +0,0 @@
|
|||||||
From 6d1bc8ae0cee4ee837d5dc8ad7f1a525208f3eec Mon Sep 17 00:00:00 2001
|
|
||||||
From: Vojtech Trefny <vtrefny@redhat.com>
|
|
||||||
Date: Thu, 5 May 2022 16:35:37 +0200
|
|
||||||
Subject: [PATCH] Use LVM PV format current_size in
|
|
||||||
LVMVolumeGroupDevice._remove
|
|
||||||
|
|
||||||
The member format size is 0 when target size is not set.
|
|
||||||
|
|
||||||
Related: rhbz#2081276
|
|
||||||
---
|
|
||||||
blivet/devices/lvm.py | 2 +-
|
|
||||||
1 file changed, 1 insertion(+), 1 deletion(-)
|
|
||||||
|
|
||||||
diff --git a/blivet/devices/lvm.py b/blivet/devices/lvm.py
|
|
||||||
index 7c78c813..4700d141 100644
|
|
||||||
--- a/blivet/devices/lvm.py
|
|
||||||
+++ b/blivet/devices/lvm.py
|
|
||||||
@@ -293,7 +293,7 @@ class LVMVolumeGroupDevice(ContainerDevice):
|
|
||||||
|
|
||||||
# do not run pvmove on empty PVs
|
|
||||||
member.format.update_size_info()
|
|
||||||
- if member.format.free < member.format.size:
|
|
||||||
+ if member.format.free < member.format.current_size:
|
|
||||||
blockdev.lvm.pvmove(member.path)
|
|
||||||
blockdev.lvm.vgreduce(self.name, member.path)
|
|
||||||
|
|
||||||
--
|
|
||||||
2.35.3
|
|
||||||
|
|
@ -1,588 +0,0 @@
|
|||||||
From 91e443af7b9f6b8d7f845f353a3897e3c91015b3 Mon Sep 17 00:00:00 2001
|
|
||||||
From: Vojtech Trefny <vtrefny@redhat.com>
|
|
||||||
Date: Thu, 30 Dec 2021 16:08:43 +0100
|
|
||||||
Subject: [PATCH 1/4] Add support for creating LVM cache pools
|
|
||||||
|
|
||||||
Resolves: rhbz#2055198
|
|
||||||
---
|
|
||||||
blivet/blivet.py | 9 +-
|
|
||||||
blivet/devicelibs/lvm.py | 9 ++
|
|
||||||
blivet/devices/lvm.py | 160 +++++++++++++++++++++++++++++++--
|
|
||||||
tests/devices_test/lvm_test.py | 26 ++++++
|
|
||||||
4 files changed, 196 insertions(+), 8 deletions(-)
|
|
||||||
|
|
||||||
diff --git a/blivet/blivet.py b/blivet/blivet.py
|
|
||||||
index c6908eb0..d29fadd0 100644
|
|
||||||
--- a/blivet/blivet.py
|
|
||||||
+++ b/blivet/blivet.py
|
|
||||||
@@ -576,6 +576,8 @@ class Blivet(object):
|
|
||||||
:type vdo_pool: bool
|
|
||||||
:keyword vdo_lv: whether to create a vdo lv
|
|
||||||
:type vdo_lv: bool
|
|
||||||
+ :keyword cache_pool: whether to create a cache pool
|
|
||||||
+ :type cache_pool: bool
|
|
||||||
:returns: the new device
|
|
||||||
:rtype: :class:`~.devices.LVMLogicalVolumeDevice`
|
|
||||||
|
|
||||||
@@ -594,6 +596,7 @@ class Blivet(object):
|
|
||||||
thin_pool = kwargs.pop("thin_pool", False)
|
|
||||||
vdo_pool = kwargs.pop("vdo_pool", False)
|
|
||||||
vdo_lv = kwargs.pop("vdo_lv", False)
|
|
||||||
+ cache_pool = kwargs.pop("cache_pool", False)
|
|
||||||
parent = kwargs.get("parents", [None])[0]
|
|
||||||
if (thin_volume or vdo_lv) and parent:
|
|
||||||
# kwargs["parents"] will contain the pool device, so...
|
|
||||||
@@ -609,6 +612,8 @@ class Blivet(object):
|
|
||||||
kwargs["seg_type"] = "vdo-pool"
|
|
||||||
if vdo_lv:
|
|
||||||
kwargs["seg_type"] = "vdo"
|
|
||||||
+ if cache_pool:
|
|
||||||
+ kwargs["seg_type"] = "cache-pool"
|
|
||||||
|
|
||||||
mountpoint = kwargs.pop("mountpoint", None)
|
|
||||||
if 'fmt_type' in kwargs:
|
|
||||||
@@ -640,7 +645,7 @@ class Blivet(object):
|
|
||||||
swap = False
|
|
||||||
|
|
||||||
prefix = ""
|
|
||||||
- if thin_pool or vdo_pool:
|
|
||||||
+ if thin_pool or vdo_pool or cache_pool:
|
|
||||||
prefix = "pool"
|
|
||||||
|
|
||||||
name = self.suggest_device_name(parent=vg,
|
|
||||||
@@ -651,7 +656,7 @@ class Blivet(object):
|
|
||||||
if "%s-%s" % (vg.name, name) in self.names:
|
|
||||||
raise ValueError("name '%s' is already in use" % name)
|
|
||||||
|
|
||||||
- if thin_pool or thin_volume or vdo_pool or vdo_lv:
|
|
||||||
+ if thin_pool or thin_volume or vdo_pool or vdo_lv or cache_pool:
|
|
||||||
cache_req = kwargs.pop("cache_request", None)
|
|
||||||
if cache_req:
|
|
||||||
raise ValueError("Creating cached thin and VDO volumes and pools is not supported")
|
|
||||||
diff --git a/blivet/devicelibs/lvm.py b/blivet/devicelibs/lvm.py
|
|
||||||
index cb6f655e..724aaff4 100644
|
|
||||||
--- a/blivet/devicelibs/lvm.py
|
|
||||||
+++ b/blivet/devicelibs/lvm.py
|
|
||||||
@@ -54,6 +54,11 @@ LVM_THINP_MIN_CHUNK_SIZE = Size("64 KiB")
|
|
||||||
LVM_THINP_MAX_CHUNK_SIZE = Size("1 GiB")
|
|
||||||
LVM_THINP_ADDRESSABLE_CHUNK_SIZE = Size("17455015526400 B") # 15.88 TiB
|
|
||||||
|
|
||||||
+# cache constants
|
|
||||||
+LVM_CACHE_MIN_METADATA_SIZE = Size("8 MiB")
|
|
||||||
+LVM_CACHE_MAX_METADATA_SIZE = Size("16 GiB")
|
|
||||||
+LVM_CACHE_DEFAULT_MODE = blockdev.LVMCacheMode.WRITETHROUGH
|
|
||||||
+
|
|
||||||
raid_levels = raid.RAIDLevels(["linear", "striped", "raid1", "raid4", "raid5", "raid6", "raid10"])
|
|
||||||
raid_seg_types = list(itertools.chain.from_iterable([level.names for level in raid_levels if level.name != "linear"]))
|
|
||||||
|
|
||||||
@@ -236,3 +241,7 @@ def recommend_thpool_chunk_size(thpool_size):
|
|
||||||
# for every ~15.88 TiB of thinpool data size
|
|
||||||
return min(math.ceil(thpool_size / LVM_THINP_ADDRESSABLE_CHUNK_SIZE) * LVM_THINP_MIN_CHUNK_SIZE,
|
|
||||||
LVM_THINP_MAX_CHUNK_SIZE)
|
|
||||||
+
|
|
||||||
+
|
|
||||||
+def is_valid_cache_md_size(md_size):
|
|
||||||
+ return md_size >= LVM_CACHE_MIN_METADATA_SIZE and md_size <= LVM_CACHE_MAX_METADATA_SIZE
|
|
||||||
diff --git a/blivet/devices/lvm.py b/blivet/devices/lvm.py
|
|
||||||
index 4700d141..7d374c3b 100644
|
|
||||||
--- a/blivet/devices/lvm.py
|
|
||||||
+++ b/blivet/devices/lvm.py
|
|
||||||
@@ -43,6 +43,7 @@ from .. import util
|
|
||||||
from ..storage_log import log_method_call
|
|
||||||
from .. import udev
|
|
||||||
from ..size import Size, KiB, MiB, ROUND_UP, ROUND_DOWN
|
|
||||||
+from ..static_data.lvm_info import lvs_info
|
|
||||||
from ..tasks import availability
|
|
||||||
|
|
||||||
import logging
|
|
||||||
@@ -646,7 +647,7 @@ class LVMLogicalVolumeBase(DMDevice, RaidDevice):
|
|
||||||
percent=None, cache_request=None, pvs=None, from_lvs=None):
|
|
||||||
|
|
||||||
if not exists:
|
|
||||||
- if seg_type not in [None, "linear", "thin", "thin-pool", "cache", "vdo-pool", "vdo"] + lvm.raid_seg_types:
|
|
||||||
+ if seg_type not in [None, "linear", "thin", "thin-pool", "cache", "vdo-pool", "vdo", "cache-pool"] + lvm.raid_seg_types:
|
|
||||||
raise ValueError("Invalid or unsupported segment type: %s" % seg_type)
|
|
||||||
if seg_type and seg_type in lvm.raid_seg_types and not pvs:
|
|
||||||
raise ValueError("List of PVs has to be given for every non-linear LV")
|
|
||||||
@@ -690,8 +691,8 @@ class LVMLogicalVolumeBase(DMDevice, RaidDevice):
|
|
||||||
# we reserve space for it
|
|
||||||
self._metadata_size = self.vg.pe_size
|
|
||||||
self._size -= self._metadata_size
|
|
||||||
- elif self.seg_type == "thin-pool":
|
|
||||||
- # LVMThinPoolMixin sets self._metadata_size on its own
|
|
||||||
+ elif self.seg_type in ("thin-pool", "cache_pool"):
|
|
||||||
+ # LVMThinPoolMixin and LVMCachePoolMixin set self._metadata_size on their own
|
|
||||||
if not self.exists and not from_lvs and not grow:
|
|
||||||
# a thin pool we are not going to grow -> lets calculate metadata
|
|
||||||
# size now if not given explicitly
|
|
||||||
@@ -1619,7 +1620,6 @@ class LVMThinPoolMixin(object):
|
|
||||||
""" A list of this pool's LVs """
|
|
||||||
return self._lvs[:] # we don't want folks changing our list
|
|
||||||
|
|
||||||
- @util.requires_property("is_thin_pool")
|
|
||||||
def autoset_md_size(self, enforced=False):
|
|
||||||
""" If self._metadata_size not set already, it calculates the recommended value
|
|
||||||
and sets it while subtracting the size from self.size.
|
|
||||||
@@ -2032,9 +2032,142 @@ class LVMVDOLogicalVolumeMixin(object):
|
|
||||||
self.pool._add_log_vol(self)
|
|
||||||
|
|
||||||
|
|
||||||
+class LVMCachePoolMixin(object):
|
|
||||||
+ def __init__(self, metadata_size, cache_mode=None):
|
|
||||||
+ self._metadata_size = metadata_size or Size(0)
|
|
||||||
+ self._cache_mode = cache_mode
|
|
||||||
+
|
|
||||||
+ def _init_check(self):
|
|
||||||
+ if not self.is_cache_pool:
|
|
||||||
+ return
|
|
||||||
+
|
|
||||||
+ if self._metadata_size and not lvm.is_valid_cache_md_size(self._metadata_size):
|
|
||||||
+ raise ValueError("invalid metadatasize value")
|
|
||||||
+
|
|
||||||
+ if not self.exists and not self._pv_specs:
|
|
||||||
+ raise ValueError("at least one fast PV must be specified to create a cache pool")
|
|
||||||
+
|
|
||||||
+ def _check_from_lvs(self):
|
|
||||||
+ if self._from_lvs:
|
|
||||||
+ if len(self._from_lvs) != 2:
|
|
||||||
+ raise errors.DeviceError("two LVs required to create a cache pool")
|
|
||||||
+
|
|
||||||
+ def _convert_from_lvs(self):
|
|
||||||
+ data_lv, metadata_lv = self._from_lvs
|
|
||||||
+
|
|
||||||
+ data_lv.parent_lv = self # also adds the LV to self._internal_lvs
|
|
||||||
+ data_lv.int_lv_type = LVMInternalLVtype.data
|
|
||||||
+ metadata_lv.parent_lv = self
|
|
||||||
+ metadata_lv.int_lv_type = LVMInternalLVtype.meta
|
|
||||||
+
|
|
||||||
+ self.size = data_lv.size
|
|
||||||
+
|
|
||||||
+ @property
|
|
||||||
+ def is_cache_pool(self):
|
|
||||||
+ return self.seg_type == "cache-pool"
|
|
||||||
+
|
|
||||||
+ @property
|
|
||||||
+ def profile(self):
|
|
||||||
+ return self._profile
|
|
||||||
+
|
|
||||||
+ @property
|
|
||||||
+ def type(self):
|
|
||||||
+ return "lvmcachepool"
|
|
||||||
+
|
|
||||||
+ @property
|
|
||||||
+ def resizable(self):
|
|
||||||
+ return False
|
|
||||||
+
|
|
||||||
+ def read_current_size(self):
|
|
||||||
+ log_method_call(self, exists=self.exists, path=self.path,
|
|
||||||
+ sysfs_path=self.sysfs_path)
|
|
||||||
+ if self.size != Size(0):
|
|
||||||
+ return self.size
|
|
||||||
+
|
|
||||||
+ if self.exists:
|
|
||||||
+ # cache pools are not active and don't have th device mapper mapping
|
|
||||||
+ # so we can't get this from sysfs
|
|
||||||
+ lv_info = lvs_info.cache.get(self.name)
|
|
||||||
+ if lv_info is None:
|
|
||||||
+ log.error("Failed to get size for existing cache pool '%s'", self.name)
|
|
||||||
+ return Size(0)
|
|
||||||
+ else:
|
|
||||||
+ return Size(lv_info.size)
|
|
||||||
+
|
|
||||||
+ return Size(0)
|
|
||||||
+
|
|
||||||
+ def autoset_md_size(self, enforced=False):
|
|
||||||
+ """ If self._metadata_size not set already, it calculates the recommended value
|
|
||||||
+ and sets it while subtracting the size from self.size.
|
|
||||||
+
|
|
||||||
+ """
|
|
||||||
+
|
|
||||||
+ log.debug("Auto-setting cache pool metadata size")
|
|
||||||
+
|
|
||||||
+ if self._size <= Size(0):
|
|
||||||
+ log.debug("Cache pool size not bigger than 0, just setting metadata size to 0")
|
|
||||||
+ self._metadata_size = 0
|
|
||||||
+ return
|
|
||||||
+
|
|
||||||
+ old_md_size = self._metadata_size
|
|
||||||
+ if self._metadata_size == 0 or enforced:
|
|
||||||
+ self._metadata_size = blockdev.lvm.cache_get_default_md_size(self._size)
|
|
||||||
+ log.debug("Using recommended metadata size: %s", self._metadata_size)
|
|
||||||
+
|
|
||||||
+ self._metadata_size = self.vg.align(self._metadata_size, roundup=True)
|
|
||||||
+ log.debug("Rounded metadata size to extents: %s MiB", self._metadata_size.convert_to("MiB"))
|
|
||||||
+
|
|
||||||
+ if self._metadata_size == old_md_size:
|
|
||||||
+ log.debug("Rounded metadata size unchanged")
|
|
||||||
+ else:
|
|
||||||
+ new_size = self.size - (self._metadata_size - old_md_size)
|
|
||||||
+ log.debug("Adjusting size from %s MiB to %s MiB",
|
|
||||||
+ self.size.convert_to("MiB"), new_size.convert_to("MiB"))
|
|
||||||
+ self.size = new_size
|
|
||||||
+
|
|
||||||
+ def _pre_create(self):
|
|
||||||
+ # make sure all the LVs this LV should be created from exist (if any)
|
|
||||||
+ if self._from_lvs and any(not lv.exists for lv in self._from_lvs):
|
|
||||||
+ raise errors.DeviceError("Component LVs need to be created first")
|
|
||||||
+
|
|
||||||
+ def _create(self):
|
|
||||||
+ """ Create the device. """
|
|
||||||
+ log_method_call(self, self.name, status=self.status)
|
|
||||||
+ if self._cache_mode:
|
|
||||||
+ try:
|
|
||||||
+ cache_mode = blockdev.lvm.cache_get_mode_from_str(self._cache_mode)
|
|
||||||
+ except blockdev.LVMError as e:
|
|
||||||
+ raise errors.DeviceError from e
|
|
||||||
+ else:
|
|
||||||
+ cache_mode = lvm.LVM_CACHE_DEFAULT_MODE
|
|
||||||
+
|
|
||||||
+ if self._from_lvs:
|
|
||||||
+ extra = dict()
|
|
||||||
+ if self.mode:
|
|
||||||
+ # we need the string here, it will be passed directly to he lvm command
|
|
||||||
+ extra["cachemode"] = self._cache_mode
|
|
||||||
+ data_lv = six.next(lv for lv in self._internal_lvs if lv.int_lv_type == LVMInternalLVtype.data)
|
|
||||||
+ meta_lv = six.next(lv for lv in self._internal_lvs if lv.int_lv_type == LVMInternalLVtype.meta)
|
|
||||||
+ blockdev.lvm.cache_pool_convert(self.vg.name, data_lv.lvname, meta_lv.lvname, self.lvname, **extra)
|
|
||||||
+ else:
|
|
||||||
+ blockdev.lvm.cache_create_pool(self.vg.name, self.lvname, self.size,
|
|
||||||
+ self.metadata_size,
|
|
||||||
+ cache_mode,
|
|
||||||
+ 0,
|
|
||||||
+ [spec.pv.path for spec in self._pv_specs])
|
|
||||||
+
|
|
||||||
+ def dracut_setup_args(self):
|
|
||||||
+ return set()
|
|
||||||
+
|
|
||||||
+ @property
|
|
||||||
+ def direct(self):
|
|
||||||
+ """ Is this device directly accessible? """
|
|
||||||
+ return False
|
|
||||||
+
|
|
||||||
+
|
|
||||||
class LVMLogicalVolumeDevice(LVMLogicalVolumeBase, LVMInternalLogicalVolumeMixin, LVMSnapshotMixin,
|
|
||||||
LVMThinPoolMixin, LVMThinLogicalVolumeMixin, LVMVDOPoolMixin,
|
|
||||||
- LVMVDOLogicalVolumeMixin):
|
|
||||||
+ LVMVDOLogicalVolumeMixin, LVMCachePoolMixin):
|
|
||||||
""" An LVM Logical Volume """
|
|
||||||
|
|
||||||
# generally resizable, see :property:`resizable` for details
|
|
||||||
@@ -2046,7 +2179,7 @@ class LVMLogicalVolumeDevice(LVMLogicalVolumeBase, LVMInternalLogicalVolumeMixin
|
|
||||||
parent_lv=None, int_type=None, origin=None, vorigin=False,
|
|
||||||
metadata_size=None, chunk_size=None, profile=None, from_lvs=None,
|
|
||||||
compression=False, deduplication=False, index_memory=0,
|
|
||||||
- write_policy=None):
|
|
||||||
+ write_policy=None, cache_mode=None):
|
|
||||||
"""
|
|
||||||
:param name: the device name (generally a device node's basename)
|
|
||||||
:type name: str
|
|
||||||
@@ -2116,6 +2249,13 @@ class LVMLogicalVolumeDevice(LVMLogicalVolumeBase, LVMInternalLogicalVolumeMixin
|
|
||||||
:keyword write_policy: write policy for the volume or None for default
|
|
||||||
:type write_policy: str
|
|
||||||
|
|
||||||
+ For cache pools only:
|
|
||||||
+
|
|
||||||
+ :keyword metadata_size: the size of the metadata LV
|
|
||||||
+ :type metadata_size: :class:`~.size.Size`
|
|
||||||
+ :keyword cache_mode: mode for the cache or None for default (writethrough)
|
|
||||||
+ :type cache_mode: str
|
|
||||||
+
|
|
||||||
"""
|
|
||||||
|
|
||||||
if isinstance(parents, (list, ParentList)):
|
|
||||||
@@ -2133,6 +2273,7 @@ class LVMLogicalVolumeDevice(LVMLogicalVolumeBase, LVMInternalLogicalVolumeMixin
|
|
||||||
LVMSnapshotMixin.__init__(self, origin, vorigin)
|
|
||||||
LVMThinPoolMixin.__init__(self, metadata_size, chunk_size, profile)
|
|
||||||
LVMThinLogicalVolumeMixin.__init__(self)
|
|
||||||
+ LVMCachePoolMixin.__init__(self, metadata_size, cache_mode)
|
|
||||||
LVMLogicalVolumeBase.__init__(self, name, parents, size, uuid, seg_type,
|
|
||||||
fmt, exists, sysfs_path, grow, maxsize,
|
|
||||||
percent, cache_request, pvs, from_lvs)
|
|
||||||
@@ -2144,6 +2285,7 @@ class LVMLogicalVolumeDevice(LVMLogicalVolumeBase, LVMInternalLogicalVolumeMixin
|
|
||||||
LVMSnapshotMixin._init_check(self)
|
|
||||||
LVMThinPoolMixin._init_check(self)
|
|
||||||
LVMThinLogicalVolumeMixin._init_check(self)
|
|
||||||
+ LVMCachePoolMixin._init_check(self)
|
|
||||||
|
|
||||||
if self._from_lvs:
|
|
||||||
self._check_from_lvs()
|
|
||||||
@@ -2169,6 +2311,8 @@ class LVMLogicalVolumeDevice(LVMLogicalVolumeBase, LVMInternalLogicalVolumeMixin
|
|
||||||
ret.append(LVMVDOPoolMixin)
|
|
||||||
if self.is_vdo_lv:
|
|
||||||
ret.append(LVMVDOLogicalVolumeMixin)
|
|
||||||
+ if self.is_cache_pool:
|
|
||||||
+ ret.append(LVMCachePoolMixin)
|
|
||||||
return ret
|
|
||||||
|
|
||||||
def _try_specific_call(self, name, *args, **kwargs):
|
|
||||||
@@ -2552,6 +2696,10 @@ class LVMLogicalVolumeDevice(LVMLogicalVolumeBase, LVMInternalLogicalVolumeMixin
|
|
||||||
|
|
||||||
return True
|
|
||||||
|
|
||||||
+ @type_specific
|
|
||||||
+ def autoset_md_size(self, enforced=False):
|
|
||||||
+ pass
|
|
||||||
+
|
|
||||||
def attach_cache(self, cache_pool_lv):
|
|
||||||
if self.is_thin_lv or self.is_snapshot_lv or self.is_internal_lv:
|
|
||||||
raise errors.DeviceError("Cannot attach a cache pool to the '%s' LV" % self.name)
|
|
||||||
diff --git a/tests/devices_test/lvm_test.py b/tests/devices_test/lvm_test.py
|
|
||||||
index c349f003..a1ddaf2d 100644
|
|
||||||
--- a/tests/devices_test/lvm_test.py
|
|
||||||
+++ b/tests/devices_test/lvm_test.py
|
|
||||||
@@ -867,3 +867,29 @@ class BlivetLVMVDODependenciesTest(unittest.TestCase):
|
|
||||||
|
|
||||||
vdo_supported = devicefactory.is_supported_device_type(devicefactory.DEVICE_TYPE_LVM_VDO)
|
|
||||||
self.assertFalse(vdo_supported)
|
|
||||||
+
|
|
||||||
+
|
|
||||||
+@unittest.skipUnless(not any(x.unavailable_type_dependencies() for x in DEVICE_CLASSES), "some unsupported device classes required for this test")
|
|
||||||
+class BlivetNewLVMCachePoolDeviceTest(unittest.TestCase):
|
|
||||||
+
|
|
||||||
+ def test_new_cache_pool(self):
|
|
||||||
+ b = blivet.Blivet()
|
|
||||||
+ pv = StorageDevice("pv1", fmt=blivet.formats.get_format("lvmpv"),
|
|
||||||
+ size=Size("10 GiB"), exists=True)
|
|
||||||
+ vg = LVMVolumeGroupDevice("testvg", parents=[pv], exists=True)
|
|
||||||
+
|
|
||||||
+ for dev in (pv, vg):
|
|
||||||
+ b.devicetree._add_device(dev)
|
|
||||||
+
|
|
||||||
+ # check that all the above devices are in the expected places
|
|
||||||
+ self.assertEqual(set(b.devices), {pv, vg})
|
|
||||||
+ self.assertEqual(set(b.vgs), {vg})
|
|
||||||
+
|
|
||||||
+ self.assertEqual(vg.size, Size("10236 MiB"))
|
|
||||||
+
|
|
||||||
+ cachepool = b.new_lv(name="cachepool", cache_pool=True,
|
|
||||||
+ parents=[vg], pvs=[pv])
|
|
||||||
+
|
|
||||||
+ b.create_device(cachepool)
|
|
||||||
+
|
|
||||||
+ self.assertEqual(cachepool.type, "lvmcachepool")
|
|
||||||
--
|
|
||||||
2.35.3
|
|
||||||
|
|
||||||
|
|
||||||
From d25d52e146559d226369afdb4b102e516bd9e332 Mon Sep 17 00:00:00 2001
|
|
||||||
From: Vojtech Trefny <vtrefny@redhat.com>
|
|
||||||
Date: Thu, 30 Dec 2021 16:09:04 +0100
|
|
||||||
Subject: [PATCH 2/4] examples: Add LVM cache pool example
|
|
||||||
|
|
||||||
Related: rhbz#2055198
|
|
||||||
---
|
|
||||||
examples/lvm_cachepool.py | 59 +++++++++++++++++++++++++++++++++++++++
|
|
||||||
1 file changed, 59 insertions(+)
|
|
||||||
create mode 100644 examples/lvm_cachepool.py
|
|
||||||
|
|
||||||
diff --git a/examples/lvm_cachepool.py b/examples/lvm_cachepool.py
|
|
||||||
new file mode 100644
|
|
||||||
index 00000000..ab2e8a72
|
|
||||||
--- /dev/null
|
|
||||||
+++ b/examples/lvm_cachepool.py
|
|
||||||
@@ -0,0 +1,59 @@
|
|
||||||
+import os
|
|
||||||
+
|
|
||||||
+import blivet
|
|
||||||
+from blivet.size import Size
|
|
||||||
+from blivet.util import set_up_logging, create_sparse_tempfile
|
|
||||||
+
|
|
||||||
+
|
|
||||||
+set_up_logging()
|
|
||||||
+b = blivet.Blivet() # create an instance of Blivet (don't add system devices)
|
|
||||||
+
|
|
||||||
+# create a disk image file on which to create new devices
|
|
||||||
+disk1_file = create_sparse_tempfile("disk1", Size("100GiB"))
|
|
||||||
+b.disk_images["disk1"] = disk1_file
|
|
||||||
+disk2_file = create_sparse_tempfile("disk2", Size("100GiB"))
|
|
||||||
+b.disk_images["disk2"] = disk2_file
|
|
||||||
+
|
|
||||||
+b.reset()
|
|
||||||
+
|
|
||||||
+try:
|
|
||||||
+ disk1 = b.devicetree.get_device_by_name("disk1")
|
|
||||||
+ disk2 = b.devicetree.get_device_by_name("disk2")
|
|
||||||
+
|
|
||||||
+ b.initialize_disk(disk1)
|
|
||||||
+ b.initialize_disk(disk2)
|
|
||||||
+
|
|
||||||
+ pv = b.new_partition(size=Size("50GiB"), fmt_type="lvmpv", parents=[disk1])
|
|
||||||
+ b.create_device(pv)
|
|
||||||
+ pv2 = b.new_partition(size=Size("50GiB"), fmt_type="lvmpv", parents=[disk2])
|
|
||||||
+ b.create_device(pv2)
|
|
||||||
+
|
|
||||||
+ # allocate the partitions (decide where and on which disks they'll reside)
|
|
||||||
+ blivet.partitioning.do_partitioning(b)
|
|
||||||
+
|
|
||||||
+ vg = b.new_vg(parents=[pv, pv2])
|
|
||||||
+ b.create_device(vg)
|
|
||||||
+
|
|
||||||
+ # new lv with base size 5GiB and growth up to 15GiB and an ext4 filesystem
|
|
||||||
+ lv = b.new_lv(fmt_type="ext4", size=Size("5GiB"), parents=[vg], name="cached")
|
|
||||||
+ b.create_device(lv)
|
|
||||||
+
|
|
||||||
+ # new cache pool
|
|
||||||
+ cpool = b.new_lv(size=Size("1 GiB"), parents=[vg], pvs=[pv2], cache_pool=True, name="fastlv")
|
|
||||||
+ b.create_device(cpool)
|
|
||||||
+
|
|
||||||
+ # write the new partitions to disk and format them as specified
|
|
||||||
+ b.do_it()
|
|
||||||
+ print(b.devicetree)
|
|
||||||
+
|
|
||||||
+ # attach the newly created cache pool to the "slow" LV
|
|
||||||
+ lv.attach_cache(cpool)
|
|
||||||
+
|
|
||||||
+ b.reset()
|
|
||||||
+ print(b.devicetree)
|
|
||||||
+
|
|
||||||
+ input("Check the state and hit ENTER to trigger cleanup")
|
|
||||||
+finally:
|
|
||||||
+ b.devicetree.teardown_disk_images()
|
|
||||||
+ os.unlink(disk1_file)
|
|
||||||
+ os.unlink(disk2_file)
|
|
||||||
--
|
|
||||||
2.35.3
|
|
||||||
|
|
||||||
|
|
||||||
From 2411d8aa082f6baf46f25d5f97455da983c0ee5f Mon Sep 17 00:00:00 2001
|
|
||||||
From: Vojtech Trefny <vtrefny@redhat.com>
|
|
||||||
Date: Thu, 30 Dec 2021 16:13:33 +0100
|
|
||||||
Subject: [PATCH 3/4] lvm: Use blivet static data when checking if the VG is
|
|
||||||
active
|
|
||||||
|
|
||||||
Instead of calling 'lvs' again in LVMVolumeGroupDevice.status
|
|
||||||
|
|
||||||
Related: rhbz#2055198
|
|
||||||
---
|
|
||||||
blivet/devices/lvm.py | 9 ++-------
|
|
||||||
1 file changed, 2 insertions(+), 7 deletions(-)
|
|
||||||
|
|
||||||
diff --git a/blivet/devices/lvm.py b/blivet/devices/lvm.py
|
|
||||||
index 7d374c3b..9f875e4e 100644
|
|
||||||
--- a/blivet/devices/lvm.py
|
|
||||||
+++ b/blivet/devices/lvm.py
|
|
||||||
@@ -220,13 +220,8 @@ class LVMVolumeGroupDevice(ContainerDevice):
|
|
||||||
|
|
||||||
# special handling for incomplete VGs
|
|
||||||
if not self.complete:
|
|
||||||
- try:
|
|
||||||
- lvs_info = blockdev.lvm.lvs(vg_name=self.name)
|
|
||||||
- except blockdev.LVMError:
|
|
||||||
- lvs_info = []
|
|
||||||
-
|
|
||||||
- for lv_info in lvs_info:
|
|
||||||
- if lv_info.attr and lv_info.attr[4] == 'a':
|
|
||||||
+ for lv_info in lvs_info.cache.values():
|
|
||||||
+ if lv_info.vg_name == self.name and lv_info.attr and lv_info.attr[4] == 'a':
|
|
||||||
return True
|
|
||||||
|
|
||||||
return False
|
|
||||||
--
|
|
||||||
2.35.3
|
|
||||||
|
|
||||||
|
|
||||||
From c8fda78915f31f3d5011ada3c7463f85e181983b Mon Sep 17 00:00:00 2001
|
|
||||||
From: Vojtech Trefny <vtrefny@redhat.com>
|
|
||||||
Date: Mon, 30 May 2022 17:02:43 +0200
|
|
||||||
Subject: [PATCH 4/4] Add option to attach a newly created cache pool to
|
|
||||||
existing LV
|
|
||||||
|
|
||||||
Because we do not have action for attaching the cache pool, we
|
|
||||||
cannot schedule both adding the fast PV to the VG and attaching
|
|
||||||
the cache pool to existing LV. This hack allows to schedule the
|
|
||||||
attach to happen after the cache pool is created.
|
|
||||||
|
|
||||||
Related: rhbz#2055198
|
|
||||||
---
|
|
||||||
blivet/devices/lvm.py | 38 +++++++++++++++++++++++++++++++++++---
|
|
||||||
1 file changed, 35 insertions(+), 3 deletions(-)
|
|
||||||
|
|
||||||
diff --git a/blivet/devices/lvm.py b/blivet/devices/lvm.py
|
|
||||||
index 9f875e4e..7e4fcf53 100644
|
|
||||||
--- a/blivet/devices/lvm.py
|
|
||||||
+++ b/blivet/devices/lvm.py
|
|
||||||
@@ -2028,9 +2028,10 @@ class LVMVDOLogicalVolumeMixin(object):
|
|
||||||
|
|
||||||
|
|
||||||
class LVMCachePoolMixin(object):
|
|
||||||
- def __init__(self, metadata_size, cache_mode=None):
|
|
||||||
+ def __init__(self, metadata_size, cache_mode=None, attach_to=None):
|
|
||||||
self._metadata_size = metadata_size or Size(0)
|
|
||||||
self._cache_mode = cache_mode
|
|
||||||
+ self._attach_to = attach_to
|
|
||||||
|
|
||||||
def _init_check(self):
|
|
||||||
if not self.is_cache_pool:
|
|
||||||
@@ -2042,6 +2043,9 @@ class LVMCachePoolMixin(object):
|
|
||||||
if not self.exists and not self._pv_specs:
|
|
||||||
raise ValueError("at least one fast PV must be specified to create a cache pool")
|
|
||||||
|
|
||||||
+ if self._attach_to and not self._attach_to.exists:
|
|
||||||
+ raise ValueError("cache pool can be attached only to an existing LV")
|
|
||||||
+
|
|
||||||
def _check_from_lvs(self):
|
|
||||||
if self._from_lvs:
|
|
||||||
if len(self._from_lvs) != 2:
|
|
||||||
@@ -2150,6 +2154,31 @@ class LVMCachePoolMixin(object):
|
|
||||||
cache_mode,
|
|
||||||
0,
|
|
||||||
[spec.pv.path for spec in self._pv_specs])
|
|
||||||
+ if self._attach_to:
|
|
||||||
+ self._attach_to.attach_cache(self)
|
|
||||||
+
|
|
||||||
+ def _post_create(self):
|
|
||||||
+ if self._attach_to:
|
|
||||||
+ # post_create tries to activate the LV and after attaching it no longer exists
|
|
||||||
+ return
|
|
||||||
+
|
|
||||||
+ # pylint: disable=bad-super-call
|
|
||||||
+ super(LVMLogicalVolumeBase, self)._post_create()
|
|
||||||
+
|
|
||||||
+ def add_hook(self, new=True):
|
|
||||||
+ if self._attach_to:
|
|
||||||
+ self._attach_to._cache = LVMCache(self._attach_to, size=self.size, exists=False,
|
|
||||||
+ pvs=self._pv_specs, mode=self._cache_mode)
|
|
||||||
+
|
|
||||||
+ # pylint: disable=bad-super-call
|
|
||||||
+ super(LVMLogicalVolumeBase, self).add_hook(new=new)
|
|
||||||
+
|
|
||||||
+ def remove_hook(self, modparent=True):
|
|
||||||
+ if self._attach_to:
|
|
||||||
+ self._attach_to._cache = None
|
|
||||||
+
|
|
||||||
+ # pylint: disable=bad-super-call
|
|
||||||
+ super(LVMLogicalVolumeBase, self).remove_hook(modparent=modparent)
|
|
||||||
|
|
||||||
def dracut_setup_args(self):
|
|
||||||
return set()
|
|
||||||
@@ -2174,7 +2203,7 @@ class LVMLogicalVolumeDevice(LVMLogicalVolumeBase, LVMInternalLogicalVolumeMixin
|
|
||||||
parent_lv=None, int_type=None, origin=None, vorigin=False,
|
|
||||||
metadata_size=None, chunk_size=None, profile=None, from_lvs=None,
|
|
||||||
compression=False, deduplication=False, index_memory=0,
|
|
||||||
- write_policy=None, cache_mode=None):
|
|
||||||
+ write_policy=None, cache_mode=None, attach_to=None):
|
|
||||||
"""
|
|
||||||
:param name: the device name (generally a device node's basename)
|
|
||||||
:type name: str
|
|
||||||
@@ -2250,6 +2279,9 @@ class LVMLogicalVolumeDevice(LVMLogicalVolumeBase, LVMInternalLogicalVolumeMixin
|
|
||||||
:type metadata_size: :class:`~.size.Size`
|
|
||||||
:keyword cache_mode: mode for the cache or None for default (writethrough)
|
|
||||||
:type cache_mode: str
|
|
||||||
+ :keyword attach_to: for non-existing cache pools a logical volume the pool should
|
|
||||||
+ be attached to when created
|
|
||||||
+ :type attach_to: :class:`LVMLogicalVolumeDevice`
|
|
||||||
|
|
||||||
"""
|
|
||||||
|
|
||||||
@@ -2268,7 +2300,7 @@ class LVMLogicalVolumeDevice(LVMLogicalVolumeBase, LVMInternalLogicalVolumeMixin
|
|
||||||
LVMSnapshotMixin.__init__(self, origin, vorigin)
|
|
||||||
LVMThinPoolMixin.__init__(self, metadata_size, chunk_size, profile)
|
|
||||||
LVMThinLogicalVolumeMixin.__init__(self)
|
|
||||||
- LVMCachePoolMixin.__init__(self, metadata_size, cache_mode)
|
|
||||||
+ LVMCachePoolMixin.__init__(self, metadata_size, cache_mode, attach_to)
|
|
||||||
LVMLogicalVolumeBase.__init__(self, name, parents, size, uuid, seg_type,
|
|
||||||
fmt, exists, sysfs_path, grow, maxsize,
|
|
||||||
percent, cache_request, pvs, from_lvs)
|
|
||||||
--
|
|
||||||
2.35.3
|
|
||||||
|
|
@ -1,23 +0,0 @@
|
|||||||
From d609cebba48744c97ac7e0461f8827ab63198026 Mon Sep 17 00:00:00 2001
|
|
||||||
From: Vojtech Trefny <vtrefny@redhat.com>
|
|
||||||
Date: Thu, 10 Jun 2021 16:58:42 +0200
|
|
||||||
Subject: [PATCH] Fix util.virt_detect on Xen
|
|
||||||
|
|
||||||
Xen is apparently still alive so we should return True for it too.
|
|
||||||
---
|
|
||||||
blivet/util.py | 2 +-
|
|
||||||
1 file changed, 1 insertion(+), 1 deletion(-)
|
|
||||||
|
|
||||||
diff --git a/blivet/util.py b/blivet/util.py
|
|
||||||
index 3bebb003..af60210b 100644
|
|
||||||
--- a/blivet/util.py
|
|
||||||
+++ b/blivet/util.py
|
|
||||||
@@ -1130,4 +1130,4 @@ def detect_virt():
|
|
||||||
except (safe_dbus.DBusCallError, safe_dbus.DBusPropertyError):
|
|
||||||
return False
|
|
||||||
else:
|
|
||||||
- return vm[0] in ('qemu', 'kvm')
|
|
||||||
+ return vm[0] in ('qemu', 'kvm', 'xen')
|
|
||||||
--
|
|
||||||
2.31.1
|
|
||||||
|
|
@ -1,776 +0,0 @@
|
|||||||
From a03be3924318788e42bcdb3ed6a5334aed771c43 Mon Sep 17 00:00:00 2001
|
|
||||||
From: Jan Stodola <jstodola@redhat.com>
|
|
||||||
Date: Thu, 28 Oct 2021 21:17:25 +0200
|
|
||||||
Subject: [PATCH 1/8] Fix removing zFCP SCSI devices
|
|
||||||
|
|
||||||
Values parsed from /proc/scsi/scsi were not correctly used to assemble
|
|
||||||
paths to SCSI devices.
|
|
||||||
|
|
||||||
For example:
|
|
||||||
/sys/bus/scsi/devices/0:0:00:00/
|
|
||||||
was incorrectly accessed instead of:
|
|
||||||
/sys/bus/scsi/devices/0:0:0:0/
|
|
||||||
|
|
||||||
Switch to a more reliable way of listing the available SCSI devices.
|
|
||||||
---
|
|
||||||
blivet/zfcp.py | 17 ++++-------------
|
|
||||||
1 file changed, 4 insertions(+), 13 deletions(-)
|
|
||||||
|
|
||||||
diff --git a/blivet/zfcp.py b/blivet/zfcp.py
|
|
||||||
index 93af5419..3747290e 100644
|
|
||||||
--- a/blivet/zfcp.py
|
|
||||||
+++ b/blivet/zfcp.py
|
|
||||||
@@ -20,6 +20,7 @@
|
|
||||||
#
|
|
||||||
|
|
||||||
import os
|
|
||||||
+import re
|
|
||||||
from . import udev
|
|
||||||
from . import util
|
|
||||||
from .i18n import _
|
|
||||||
@@ -167,20 +168,10 @@ class ZFCPDevice:
|
|
||||||
return True
|
|
||||||
|
|
||||||
def offline_scsi_device(self):
|
|
||||||
- f = open("/proc/scsi/scsi", "r")
|
|
||||||
- lines = f.readlines()
|
|
||||||
- f.close()
|
|
||||||
- # alternatively iterate over /sys/bus/scsi/devices/*:0:*:*/
|
|
||||||
+ # A list of existing SCSI devices in format Host:Bus:Target:Lun
|
|
||||||
+ scsi_devices = [f for f in os.listdir(scsidevsysfs) if re.search(r'^[0-9]+:[0-9]+:[0-9]+:[0-9]+$', f)]
|
|
||||||
|
|
||||||
- for line in lines:
|
|
||||||
- if not line.startswith("Host"):
|
|
||||||
- continue
|
|
||||||
- scsihost = line.split()
|
|
||||||
- host = scsihost[1]
|
|
||||||
- channel = "0"
|
|
||||||
- devid = scsihost[5]
|
|
||||||
- lun = scsihost[7]
|
|
||||||
- scsidev = "%s:%s:%s:%s" % (host[4:], channel, devid, lun)
|
|
||||||
+ for scsidev in scsi_devices:
|
|
||||||
fcpsysfs = "%s/%s" % (scsidevsysfs, scsidev)
|
|
||||||
scsidel = "%s/%s/delete" % (scsidevsysfs, scsidev)
|
|
||||||
|
|
||||||
--
|
|
||||||
2.36.1
|
|
||||||
|
|
||||||
|
|
||||||
From 82bd018fdc47c64f30d8422eb90bc76564072a26 Mon Sep 17 00:00:00 2001
|
|
||||||
From: Jan Stodola <jstodola@redhat.com>
|
|
||||||
Date: Sun, 21 Nov 2021 02:47:45 +0100
|
|
||||||
Subject: [PATCH 2/8] Refactor the ZFCPDevice class
|
|
||||||
|
|
||||||
Add a new base class for zFCP devices.
|
|
||||||
Move code to the new base class.
|
|
||||||
Improve documentation.
|
|
||||||
---
|
|
||||||
blivet/zfcp.py | 131 +++++++++++++++++++++++++++++++++++--------------
|
|
||||||
1 file changed, 95 insertions(+), 36 deletions(-)
|
|
||||||
|
|
||||||
diff --git a/blivet/zfcp.py b/blivet/zfcp.py
|
|
||||||
index 3747290e..4a50f65f 100644
|
|
||||||
--- a/blivet/zfcp.py
|
|
||||||
+++ b/blivet/zfcp.py
|
|
||||||
@@ -21,6 +21,7 @@
|
|
||||||
|
|
||||||
import os
|
|
||||||
import re
|
|
||||||
+from abc import ABC
|
|
||||||
from . import udev
|
|
||||||
from . import util
|
|
||||||
from .i18n import _
|
|
||||||
@@ -46,29 +47,19 @@ zfcpsysfs = "/sys/bus/ccw/drivers/zfcp"
|
|
||||||
scsidevsysfs = "/sys/bus/scsi/devices"
|
|
||||||
zfcpconf = "/etc/zfcp.conf"
|
|
||||||
|
|
||||||
+class ZFCPDeviceBase(ABC):
|
|
||||||
+ """An abstract base class for zFCP storage devices."""
|
|
||||||
|
|
||||||
-class ZFCPDevice:
|
|
||||||
- """
|
|
||||||
- .. warning::
|
|
||||||
- Since this is a singleton class, calling deepcopy() on the instance
|
|
||||||
- just returns ``self`` with no copy being created.
|
|
||||||
- """
|
|
||||||
-
|
|
||||||
- def __init__(self, devnum, wwpn, fcplun):
|
|
||||||
+ def __init__(self, devnum):
|
|
||||||
self.devnum = blockdev.s390.sanitize_dev_input(devnum)
|
|
||||||
- self.wwpn = blockdev.s390.zfcp_sanitize_wwpn_input(wwpn)
|
|
||||||
- self.fcplun = blockdev.s390.zfcp_sanitize_lun_input(fcplun)
|
|
||||||
-
|
|
||||||
if not self.devnum:
|
|
||||||
raise ValueError(_("You have not specified a device number or the number is invalid"))
|
|
||||||
- if not self.wwpn:
|
|
||||||
- raise ValueError(_("You have not specified a worldwide port name or the name is invalid."))
|
|
||||||
- if not self.fcplun:
|
|
||||||
- raise ValueError(_("You have not specified a FCP LUN or the number is invalid."))
|
|
||||||
+
|
|
||||||
+ self._device_online_path = os.path.join(zfcpsysfs, self.devnum, "online")
|
|
||||||
|
|
||||||
# Force str and unicode types in case any of the properties are unicode
|
|
||||||
def _to_string(self):
|
|
||||||
- return "%s %s %s" % (self.devnum, self.wwpn, self.fcplun)
|
|
||||||
+ return str(self.devnum)
|
|
||||||
|
|
||||||
def __str__(self):
|
|
||||||
return stringize(self._to_string())
|
|
||||||
@@ -76,33 +67,97 @@ class ZFCPDevice:
|
|
||||||
def __unicode__(self):
|
|
||||||
return unicodeize(self._to_string())
|
|
||||||
|
|
||||||
- def online_device(self):
|
|
||||||
- online = "%s/%s/online" % (zfcpsysfs, self.devnum)
|
|
||||||
- portadd = "%s/%s/port_add" % (zfcpsysfs, self.devnum)
|
|
||||||
- portdir = "%s/%s/%s" % (zfcpsysfs, self.devnum, self.wwpn)
|
|
||||||
- unitadd = "%s/unit_add" % (portdir)
|
|
||||||
- unitdir = "%s/%s" % (portdir, self.fcplun)
|
|
||||||
- failed = "%s/failed" % (unitdir)
|
|
||||||
+ def _free_device(self):
|
|
||||||
+ """Remove the device from the I/O ignore list to make it visible to the system.
|
|
||||||
+
|
|
||||||
+ :raises: ValueError if the device cannot be removed from the I/O ignore list
|
|
||||||
+ """
|
|
||||||
|
|
||||||
- if not os.path.exists(online):
|
|
||||||
+ if not os.path.exists(self._device_online_path):
|
|
||||||
log.info("Freeing zFCP device %s", self.devnum)
|
|
||||||
util.run_program(["zfcp_cio_free", "-d", self.devnum])
|
|
||||||
|
|
||||||
- if not os.path.exists(online):
|
|
||||||
+ if not os.path.exists(self._device_online_path):
|
|
||||||
raise ValueError(_("zFCP device %s not found, not even in device ignore list.") %
|
|
||||||
(self.devnum,))
|
|
||||||
|
|
||||||
+ def _set_zfcp_device_online(self):
|
|
||||||
+ """Set the zFCP device online.
|
|
||||||
+
|
|
||||||
+ :raises: ValueError if the device cannot be set online
|
|
||||||
+ """
|
|
||||||
+
|
|
||||||
try:
|
|
||||||
- f = open(online, "r")
|
|
||||||
- devonline = f.readline().strip()
|
|
||||||
- f.close()
|
|
||||||
+ with open(self._device_online_path) as f:
|
|
||||||
+ devonline = f.readline().strip()
|
|
||||||
if devonline != "1":
|
|
||||||
- logged_write_line_to_file(online, "1")
|
|
||||||
+ logged_write_line_to_file(self._device_online_path, "1")
|
|
||||||
except OSError as e:
|
|
||||||
raise ValueError(_("Could not set zFCP device %(devnum)s "
|
|
||||||
"online (%(e)s).")
|
|
||||||
% {'devnum': self.devnum, 'e': e})
|
|
||||||
|
|
||||||
+ def _set_zfcp_device_offline(self):
|
|
||||||
+ """Set the zFCP device offline.
|
|
||||||
+
|
|
||||||
+ :raises: ValueError if the device cannot be set offline
|
|
||||||
+ """
|
|
||||||
+
|
|
||||||
+ try:
|
|
||||||
+ logged_write_line_to_file(self._device_online_path, "0")
|
|
||||||
+ except OSError as e:
|
|
||||||
+ raise ValueError(_("Could not set zFCP device %(devnum)s "
|
|
||||||
+ "offline (%(e)s).")
|
|
||||||
+ % {'devnum': self.devnum, 'e': e})
|
|
||||||
+
|
|
||||||
+ def online_device(self):
|
|
||||||
+ """Initialize the device and make its storage block device(s) ready to use.
|
|
||||||
+
|
|
||||||
+ :returns: True if success
|
|
||||||
+ :raises: ValueError if the device cannot be initialized
|
|
||||||
+ """
|
|
||||||
+
|
|
||||||
+ self._free_device()
|
|
||||||
+ self._set_zfcp_device_online()
|
|
||||||
+ return True
|
|
||||||
+
|
|
||||||
+
|
|
||||||
+class ZFCPDevice(ZFCPDeviceBase):
|
|
||||||
+ """A class for zFCP devices that are not configured in NPIV mode. Such
|
|
||||||
+ devices have to be specified by a device number, WWPN and LUN.
|
|
||||||
+ """
|
|
||||||
+
|
|
||||||
+ def __init__(self, devnum, wwpn, fcplun):
|
|
||||||
+ super().__init__(devnum)
|
|
||||||
+
|
|
||||||
+ self.wwpn = blockdev.s390.zfcp_sanitize_wwpn_input(wwpn)
|
|
||||||
+ if not self.wwpn:
|
|
||||||
+ raise ValueError(_("You have not specified a worldwide port name or the name is invalid."))
|
|
||||||
+
|
|
||||||
+ self.fcplun = blockdev.s390.zfcp_sanitize_lun_input(fcplun)
|
|
||||||
+ if not self.fcplun:
|
|
||||||
+ raise ValueError(_("You have not specified a FCP LUN or the number is invalid."))
|
|
||||||
+
|
|
||||||
+ # Force str and unicode types in case any of the properties are unicode
|
|
||||||
+ def _to_string(self):
|
|
||||||
+ return "{} {} {}".format(self.devnum, self.wwpn, self.fcplun)
|
|
||||||
+
|
|
||||||
+ def online_device(self):
|
|
||||||
+ """Initialize the device and make its storage block device(s) ready to use.
|
|
||||||
+
|
|
||||||
+ :returns: True if success
|
|
||||||
+ :raises: ValueError if the device cannot be initialized
|
|
||||||
+ """
|
|
||||||
+
|
|
||||||
+ super().online_device()
|
|
||||||
+
|
|
||||||
+ portadd = "%s/%s/port_add" % (zfcpsysfs, self.devnum)
|
|
||||||
+ portdir = "%s/%s/%s" % (zfcpsysfs, self.devnum, self.wwpn)
|
|
||||||
+ unitadd = "%s/unit_add" % (portdir)
|
|
||||||
+ unitdir = "%s/%s" % (portdir, self.fcplun)
|
|
||||||
+ failed = "%s/failed" % (unitdir)
|
|
||||||
+
|
|
||||||
+ # create the sysfs directory for the WWPN/port
|
|
||||||
if not os.path.exists(portdir):
|
|
||||||
if os.path.exists(portadd):
|
|
||||||
# older zfcp sysfs interface
|
|
||||||
@@ -127,6 +182,7 @@ class ZFCPDevice:
|
|
||||||
"there.", {'wwpn': self.wwpn,
|
|
||||||
'devnum': self.devnum})
|
|
||||||
|
|
||||||
+ # create the sysfs directory for the LUN/unit
|
|
||||||
if not os.path.exists(unitdir):
|
|
||||||
try:
|
|
||||||
logged_write_line_to_file(unitadd, self.fcplun)
|
|
||||||
@@ -144,6 +200,7 @@ class ZFCPDevice:
|
|
||||||
'wwpn': self.wwpn,
|
|
||||||
'devnum': self.devnum})
|
|
||||||
|
|
||||||
+ # check the state of the LUN
|
|
||||||
fail = "0"
|
|
||||||
try:
|
|
||||||
f = open(failed, "r")
|
|
||||||
@@ -168,6 +225,8 @@ class ZFCPDevice:
|
|
||||||
return True
|
|
||||||
|
|
||||||
def offline_scsi_device(self):
|
|
||||||
+ """Find SCSI devices associated to the zFCP device and remove them from the system."""
|
|
||||||
+
|
|
||||||
# A list of existing SCSI devices in format Host:Bus:Target:Lun
|
|
||||||
scsi_devices = [f for f in os.listdir(scsidevsysfs) if re.search(r'^[0-9]+:[0-9]+:[0-9]+:[0-9]+$', f)]
|
|
||||||
|
|
||||||
@@ -196,7 +255,8 @@ class ZFCPDevice:
|
|
||||||
self.devnum, self.wwpn, self.fcplun)
|
|
||||||
|
|
||||||
def offline_device(self):
|
|
||||||
- offline = "%s/%s/online" % (zfcpsysfs, self.devnum)
|
|
||||||
+ """Remove the zFCP device from the system."""
|
|
||||||
+
|
|
||||||
portadd = "%s/%s/port_add" % (zfcpsysfs, self.devnum)
|
|
||||||
portremove = "%s/%s/port_remove" % (zfcpsysfs, self.devnum)
|
|
||||||
unitremove = "%s/%s/%s/unit_remove" % (zfcpsysfs, self.devnum, self.wwpn)
|
|
||||||
@@ -212,6 +272,7 @@ class ZFCPDevice:
|
|
||||||
% {'devnum': self.devnum, 'wwpn': self.wwpn,
|
|
||||||
'fcplun': self.fcplun, 'e': e})
|
|
||||||
|
|
||||||
+ # remove the LUN
|
|
||||||
try:
|
|
||||||
logged_write_line_to_file(unitremove, self.fcplun)
|
|
||||||
except OSError as e:
|
|
||||||
@@ -221,6 +282,7 @@ class ZFCPDevice:
|
|
||||||
% {'fcplun': self.fcplun, 'wwpn': self.wwpn,
|
|
||||||
'devnum': self.devnum, 'e': e})
|
|
||||||
|
|
||||||
+ # remove the WWPN only if there are no other LUNs attached
|
|
||||||
if os.path.exists(portadd):
|
|
||||||
# only try to remove ports with older zfcp sysfs interface
|
|
||||||
for lun in os.listdir(portdir):
|
|
||||||
@@ -238,6 +300,7 @@ class ZFCPDevice:
|
|
||||||
% {'wwpn': self.wwpn,
|
|
||||||
'devnum': self.devnum, 'e': e})
|
|
||||||
|
|
||||||
+ # check if there are other WWPNs existing for the zFCP device number
|
|
||||||
if os.path.exists(portadd):
|
|
||||||
# older zfcp sysfs interface
|
|
||||||
for port in os.listdir(devdir):
|
|
||||||
@@ -256,12 +319,8 @@ class ZFCPDevice:
|
|
||||||
self.devnum, luns[0])
|
|
||||||
return True
|
|
||||||
|
|
||||||
- try:
|
|
||||||
- logged_write_line_to_file(offline, "0")
|
|
||||||
- except OSError as e:
|
|
||||||
- raise ValueError(_("Could not set zFCP device %(devnum)s "
|
|
||||||
- "offline (%(e)s).")
|
|
||||||
- % {'devnum': self.devnum, 'e': e})
|
|
||||||
+ # no other WWPNs/LUNs exists for this device number, it's safe to bring it offline
|
|
||||||
+ self._set_zfcp_device_offline()
|
|
||||||
|
|
||||||
return True
|
|
||||||
|
|
||||||
--
|
|
||||||
2.36.1
|
|
||||||
|
|
||||||
|
|
||||||
From a9b9fe124dbc23104c0b60c8e0326cab3eb7a28d Mon Sep 17 00:00:00 2001
|
|
||||||
From: Jan Stodola <jstodola@redhat.com>
|
|
||||||
Date: Sun, 21 Nov 2021 02:35:05 +0100
|
|
||||||
Subject: [PATCH 3/8] Move offline_scsi_device() to the base class
|
|
||||||
|
|
||||||
---
|
|
||||||
blivet/zfcp.py | 74 ++++++++++++++++++++++++++++++--------------------
|
|
||||||
1 file changed, 44 insertions(+), 30 deletions(-)
|
|
||||||
|
|
||||||
diff --git a/blivet/zfcp.py b/blivet/zfcp.py
|
|
||||||
index 4a50f65f..af8f841d 100644
|
|
||||||
--- a/blivet/zfcp.py
|
|
||||||
+++ b/blivet/zfcp.py
|
|
||||||
@@ -110,6 +110,15 @@ class ZFCPDeviceBase(ABC):
|
|
||||||
"offline (%(e)s).")
|
|
||||||
% {'devnum': self.devnum, 'e': e})
|
|
||||||
|
|
||||||
+ def _is_scsi_associated_with_fcp(self, fcphbasysfs, _fcpwwpnsysfs, _fcplunsysfs):
|
|
||||||
+ """Decide if the SCSI device with the provided SCSI attributes
|
|
||||||
+ corresponds to the zFCP device.
|
|
||||||
+
|
|
||||||
+ :returns: True or False
|
|
||||||
+ """
|
|
||||||
+
|
|
||||||
+ return fcphbasysfs == self.devnum
|
|
||||||
+
|
|
||||||
def online_device(self):
|
|
||||||
"""Initialize the device and make its storage block device(s) ready to use.
|
|
||||||
|
|
||||||
@@ -121,6 +130,30 @@ class ZFCPDeviceBase(ABC):
|
|
||||||
self._set_zfcp_device_online()
|
|
||||||
return True
|
|
||||||
|
|
||||||
+ def offline_scsi_device(self):
|
|
||||||
+ """Find SCSI devices associated to the zFCP device and remove them from the system."""
|
|
||||||
+
|
|
||||||
+ # A list of existing SCSI devices in format Host:Bus:Target:Lun
|
|
||||||
+ scsi_devices = [f for f in os.listdir(scsidevsysfs) if re.search(r'^[0-9]+:[0-9]+:[0-9]+:[0-9]+$', f)]
|
|
||||||
+
|
|
||||||
+ for scsidev in scsi_devices:
|
|
||||||
+ fcpsysfs = os.path.join(scsidevsysfs, scsidev)
|
|
||||||
+
|
|
||||||
+ with open(os.path.join(fcpsysfs, "hba_id")) as f:
|
|
||||||
+ fcphbasysfs = f.readline().strip()
|
|
||||||
+ with open(os.path.join(fcpsysfs, "wwpn")) as f:
|
|
||||||
+ fcpwwpnsysfs = f.readline().strip()
|
|
||||||
+ with open(os.path.join(fcpsysfs, "fcp_lun")) as f:
|
|
||||||
+ fcplunsysfs = f.readline().strip()
|
|
||||||
+
|
|
||||||
+ if self._is_scsi_associated_with_fcp(fcphbasysfs, fcpwwpnsysfs, fcplunsysfs):
|
|
||||||
+ scsidel = os.path.join(scsidevsysfs, scsidev, "delete")
|
|
||||||
+ logged_write_line_to_file(scsidel, "1")
|
|
||||||
+ udev.settle()
|
|
||||||
+ return
|
|
||||||
+
|
|
||||||
+ log.warning("No scsi device found to delete for zfcp %s", self)
|
|
||||||
+
|
|
||||||
|
|
||||||
class ZFCPDevice(ZFCPDeviceBase):
|
|
||||||
"""A class for zFCP devices that are not configured in NPIV mode. Such
|
|
||||||
@@ -142,6 +175,17 @@ class ZFCPDevice(ZFCPDeviceBase):
|
|
||||||
def _to_string(self):
|
|
||||||
return "{} {} {}".format(self.devnum, self.wwpn, self.fcplun)
|
|
||||||
|
|
||||||
+ def _is_scsi_associated_with_fcp(self, fcphbasysfs, fcpwwpnsysfs, fcplunsysfs):
|
|
||||||
+ """Decide if the SCSI device with the provided SCSI attributes
|
|
||||||
+ corresponds to the zFCP device.
|
|
||||||
+
|
|
||||||
+ :returns: True or False
|
|
||||||
+ """
|
|
||||||
+
|
|
||||||
+ return (fcphbasysfs == self.devnum and
|
|
||||||
+ fcpwwpnsysfs == self.wwpn and
|
|
||||||
+ fcplunsysfs == self.fcplun)
|
|
||||||
+
|
|
||||||
def online_device(self):
|
|
||||||
"""Initialize the device and make its storage block device(s) ready to use.
|
|
||||||
|
|
||||||
@@ -224,36 +268,6 @@ class ZFCPDevice(ZFCPDeviceBase):
|
|
||||||
|
|
||||||
return True
|
|
||||||
|
|
||||||
- def offline_scsi_device(self):
|
|
||||||
- """Find SCSI devices associated to the zFCP device and remove them from the system."""
|
|
||||||
-
|
|
||||||
- # A list of existing SCSI devices in format Host:Bus:Target:Lun
|
|
||||||
- scsi_devices = [f for f in os.listdir(scsidevsysfs) if re.search(r'^[0-9]+:[0-9]+:[0-9]+:[0-9]+$', f)]
|
|
||||||
-
|
|
||||||
- for scsidev in scsi_devices:
|
|
||||||
- fcpsysfs = "%s/%s" % (scsidevsysfs, scsidev)
|
|
||||||
- scsidel = "%s/%s/delete" % (scsidevsysfs, scsidev)
|
|
||||||
-
|
|
||||||
- f = open("%s/hba_id" % (fcpsysfs), "r")
|
|
||||||
- fcphbasysfs = f.readline().strip()
|
|
||||||
- f.close()
|
|
||||||
- f = open("%s/wwpn" % (fcpsysfs), "r")
|
|
||||||
- fcpwwpnsysfs = f.readline().strip()
|
|
||||||
- f.close()
|
|
||||||
- f = open("%s/fcp_lun" % (fcpsysfs), "r")
|
|
||||||
- fcplunsysfs = f.readline().strip()
|
|
||||||
- f.close()
|
|
||||||
-
|
|
||||||
- if fcphbasysfs == self.devnum \
|
|
||||||
- and fcpwwpnsysfs == self.wwpn \
|
|
||||||
- and fcplunsysfs == self.fcplun:
|
|
||||||
- logged_write_line_to_file(scsidel, "1")
|
|
||||||
- udev.settle()
|
|
||||||
- return
|
|
||||||
-
|
|
||||||
- log.warning("no scsi device found to delete for zfcp %s %s %s",
|
|
||||||
- self.devnum, self.wwpn, self.fcplun)
|
|
||||||
-
|
|
||||||
def offline_device(self):
|
|
||||||
"""Remove the zFCP device from the system."""
|
|
||||||
|
|
||||||
--
|
|
||||||
2.36.1
|
|
||||||
|
|
||||||
|
|
||||||
From 47997255cf12505d743d6e01a40a51b23ed64a6d Mon Sep 17 00:00:00 2001
|
|
||||||
From: Jan Stodola <jstodola@redhat.com>
|
|
||||||
Date: Sat, 6 Nov 2021 21:27:52 +0100
|
|
||||||
Subject: [PATCH 4/8] Allow to delete more than one SCSI device
|
|
||||||
|
|
||||||
NPIV zFCP devices can attach more than one SCSI device, so allow to
|
|
||||||
delete them all. For non-NPIV devices it means possible slowdown, since
|
|
||||||
all SCSI devices would now be checked.
|
|
||||||
---
|
|
||||||
blivet/zfcp.py | 6 ++++--
|
|
||||||
1 file changed, 4 insertions(+), 2 deletions(-)
|
|
||||||
|
|
||||||
diff --git a/blivet/zfcp.py b/blivet/zfcp.py
|
|
||||||
index af8f841d..3b3f623b 100644
|
|
||||||
--- a/blivet/zfcp.py
|
|
||||||
+++ b/blivet/zfcp.py
|
|
||||||
@@ -136,6 +136,7 @@ class ZFCPDeviceBase(ABC):
|
|
||||||
# A list of existing SCSI devices in format Host:Bus:Target:Lun
|
|
||||||
scsi_devices = [f for f in os.listdir(scsidevsysfs) if re.search(r'^[0-9]+:[0-9]+:[0-9]+:[0-9]+$', f)]
|
|
||||||
|
|
||||||
+ scsi_device_found = False
|
|
||||||
for scsidev in scsi_devices:
|
|
||||||
fcpsysfs = os.path.join(scsidevsysfs, scsidev)
|
|
||||||
|
|
||||||
@@ -147,12 +148,13 @@ class ZFCPDeviceBase(ABC):
|
|
||||||
fcplunsysfs = f.readline().strip()
|
|
||||||
|
|
||||||
if self._is_scsi_associated_with_fcp(fcphbasysfs, fcpwwpnsysfs, fcplunsysfs):
|
|
||||||
+ scsi_device_found = True
|
|
||||||
scsidel = os.path.join(scsidevsysfs, scsidev, "delete")
|
|
||||||
logged_write_line_to_file(scsidel, "1")
|
|
||||||
udev.settle()
|
|
||||||
- return
|
|
||||||
|
|
||||||
- log.warning("No scsi device found to delete for zfcp %s", self)
|
|
||||||
+ if not scsi_device_found:
|
|
||||||
+ log.warning("No scsi device found to delete for zfcp %s", self)
|
|
||||||
|
|
||||||
|
|
||||||
class ZFCPDevice(ZFCPDeviceBase):
|
|
||||||
--
|
|
||||||
2.36.1
|
|
||||||
|
|
||||||
|
|
||||||
From feace41093c97dc88aa20b07a5ff6049df4bd01d Mon Sep 17 00:00:00 2001
|
|
||||||
From: Jan Stodola <jstodola@redhat.com>
|
|
||||||
Date: Sun, 21 Nov 2021 03:01:02 +0100
|
|
||||||
Subject: [PATCH 5/8] Add a function for reading the value of a kernel module
|
|
||||||
parameter
|
|
||||||
|
|
||||||
---
|
|
||||||
blivet/util.py | 33 +++++++++++++++++++++++++++++++++
|
|
||||||
tests/util_test.py | 11 +++++++++++
|
|
||||||
2 files changed, 44 insertions(+)
|
|
||||||
|
|
||||||
diff --git a/blivet/util.py b/blivet/util.py
|
|
||||||
index af60210b..cbef65e0 100644
|
|
||||||
--- a/blivet/util.py
|
|
||||||
+++ b/blivet/util.py
|
|
||||||
@@ -1131,3 +1131,36 @@ def detect_virt():
|
|
||||||
return False
|
|
||||||
else:
|
|
||||||
return vm[0] in ('qemu', 'kvm', 'xen')
|
|
||||||
+
|
|
||||||
+
|
|
||||||
+def natural_sort_key(device):
|
|
||||||
+ """ Sorting key for devices which makes sure partitions are sorted in natural
|
|
||||||
+ way, e.g. 'sda1, sda2, ..., sda10' and not like 'sda1, sda10, sda2, ...'
|
|
||||||
+ """
|
|
||||||
+ if device.type == "partition" and device.parted_partition and device.disk:
|
|
||||||
+ part_num = getattr(device.parted_partition, "number", -1)
|
|
||||||
+ return [device.disk.name, part_num]
|
|
||||||
+ else:
|
|
||||||
+ return [device.name, 0]
|
|
||||||
+
|
|
||||||
+
|
|
||||||
+def get_kernel_module_parameter(module, parameter):
|
|
||||||
+ """ Return the value of a given kernel module parameter
|
|
||||||
+
|
|
||||||
+ :param str module: a kernel module
|
|
||||||
+ :param str parameter: a module parameter
|
|
||||||
+ :returns: the value of the given kernel module parameter or None
|
|
||||||
+ :rtype: str
|
|
||||||
+ """
|
|
||||||
+
|
|
||||||
+ value = None
|
|
||||||
+
|
|
||||||
+ parameter_path = os.path.join("/sys/module", module, "parameters", parameter)
|
|
||||||
+ try:
|
|
||||||
+ with open(parameter_path) as f:
|
|
||||||
+ value = f.read().strip()
|
|
||||||
+ except IOError as e:
|
|
||||||
+ log.warning("Couldn't get the value of the parameter '%s' from the kernel module '%s': %s",
|
|
||||||
+ parameter, module, str(e))
|
|
||||||
+
|
|
||||||
+ return value
|
|
||||||
diff --git a/tests/util_test.py b/tests/util_test.py
|
|
||||||
index 853b6166..ed2549ad 100644
|
|
||||||
--- a/tests/util_test.py
|
|
||||||
+++ b/tests/util_test.py
|
|
||||||
@@ -180,3 +180,14 @@ class GetSysfsAttrTestCase(unittest.TestCase):
|
|
||||||
# the unicode replacement character (U+FFFD) should be used instead
|
|
||||||
model = util.get_sysfs_attr(sysfs, "model")
|
|
||||||
self.assertEqual(model, "test model\ufffd")
|
|
||||||
+
|
|
||||||
+
|
|
||||||
+class GetKernelModuleParameterTestCase(unittest.TestCase):
|
|
||||||
+
|
|
||||||
+ def test_nonexisting_kernel_module(self):
|
|
||||||
+ self.assertIsNone(util.get_kernel_module_parameter("unknown_module", "unknown_parameter"))
|
|
||||||
+
|
|
||||||
+ def test_get_kernel_module_parameter_value(self):
|
|
||||||
+ with mock.patch('blivet.util.open', mock.mock_open(read_data='value\n')):
|
|
||||||
+ value = util.get_kernel_module_parameter("module", "parameter")
|
|
||||||
+ self.assertEqual(value, "value")
|
|
||||||
--
|
|
||||||
2.36.1
|
|
||||||
|
|
||||||
|
|
||||||
From cea53c0f95793d8041391dd8e1edc58aa0f7868c Mon Sep 17 00:00:00 2001
|
|
||||||
From: Jan Stodola <jstodola@redhat.com>
|
|
||||||
Date: Sun, 21 Nov 2021 03:01:46 +0100
|
|
||||||
Subject: [PATCH 6/8] LUN and WWPN should not be used for NPIV zFCP devices
|
|
||||||
|
|
||||||
Log a warning if activating a zFCP device in NPIV mode and WWPN or
|
|
||||||
LUN have been provided. They are superfluous for NPIV devices.
|
|
||||||
---
|
|
||||||
blivet/zfcp.py | 58 +++++++++++++++++++++++++++++++++++++++++++++++++-
|
|
||||||
1 file changed, 57 insertions(+), 1 deletion(-)
|
|
||||||
|
|
||||||
diff --git a/blivet/zfcp.py b/blivet/zfcp.py
|
|
||||||
index 3b3f623b..726e9364 100644
|
|
||||||
--- a/blivet/zfcp.py
|
|
||||||
+++ b/blivet/zfcp.py
|
|
||||||
@@ -22,6 +22,7 @@
|
|
||||||
import os
|
|
||||||
import re
|
|
||||||
from abc import ABC
|
|
||||||
+import glob
|
|
||||||
from . import udev
|
|
||||||
from . import util
|
|
||||||
from .i18n import _
|
|
||||||
@@ -47,6 +48,55 @@ zfcpsysfs = "/sys/bus/ccw/drivers/zfcp"
|
|
||||||
scsidevsysfs = "/sys/bus/scsi/devices"
|
|
||||||
zfcpconf = "/etc/zfcp.conf"
|
|
||||||
|
|
||||||
+
|
|
||||||
+def _is_lun_scan_allowed():
|
|
||||||
+ """Return True if automatic LUN scanning is enabled by the kernel."""
|
|
||||||
+
|
|
||||||
+ allow_lun_scan = util.get_kernel_module_parameter("zfcp", "allow_lun_scan")
|
|
||||||
+ return allow_lun_scan == "Y"
|
|
||||||
+
|
|
||||||
+
|
|
||||||
+def _is_port_in_npiv_mode(device_id):
|
|
||||||
+ """Return True if the device ID is configured in NPIV mode. See
|
|
||||||
+ https://www.ibm.com/docs/en/linux-on-systems?topic=devices-use-npiv
|
|
||||||
+ """
|
|
||||||
+
|
|
||||||
+ port_in_npiv_mode = False
|
|
||||||
+ port_type_path = "/sys/bus/ccw/devices/{}/host*/fc_host/host*/port_type".format(device_id)
|
|
||||||
+ port_type_paths = glob.glob(port_type_path)
|
|
||||||
+ try:
|
|
||||||
+ for filename in port_type_paths:
|
|
||||||
+ with open(filename) as f:
|
|
||||||
+ port_type = f.read()
|
|
||||||
+ if re.search(r"(^|\s)NPIV(\s|$)", port_type):
|
|
||||||
+ port_in_npiv_mode = True
|
|
||||||
+ except OSError as e:
|
|
||||||
+ log.warning("Couldn't read the port_type attribute of the %s device: %s", device_id, str(e))
|
|
||||||
+ port_in_npiv_mode = False
|
|
||||||
+
|
|
||||||
+ return port_in_npiv_mode
|
|
||||||
+
|
|
||||||
+
|
|
||||||
+def is_npiv_enabled(device_id):
|
|
||||||
+ """Return True if the given zFCP device ID is configured and usable in
|
|
||||||
+ NPIV (N_Port ID Virtualization) mode.
|
|
||||||
+
|
|
||||||
+ :returns: True or False
|
|
||||||
+ """
|
|
||||||
+
|
|
||||||
+ # LUN scanning disabled by the kernel module prevents using the device in NPIV mode
|
|
||||||
+ if not _is_lun_scan_allowed():
|
|
||||||
+ log.warning("Automatic LUN scanning is disabled by the zfcp kernel module.")
|
|
||||||
+ return False
|
|
||||||
+
|
|
||||||
+ # The port itself has to be configured in NPIV mode
|
|
||||||
+ if not _is_port_in_npiv_mode(device_id):
|
|
||||||
+ log.warning("The zFCP device %s is not configured in NPIV mode.", device_id)
|
|
||||||
+ return False
|
|
||||||
+
|
|
||||||
+ return True
|
|
||||||
+
|
|
||||||
+
|
|
||||||
class ZFCPDeviceBase(ABC):
|
|
||||||
"""An abstract base class for zFCP storage devices."""
|
|
||||||
|
|
||||||
@@ -203,6 +253,13 @@ class ZFCPDevice(ZFCPDeviceBase):
|
|
||||||
unitdir = "%s/%s" % (portdir, self.fcplun)
|
|
||||||
failed = "%s/failed" % (unitdir)
|
|
||||||
|
|
||||||
+ # Activating an NPIV enabled device using devnum, WWPN and LUN should still be possible
|
|
||||||
+ # as this method was used as a workaround until the support for NPIV enabled devices has
|
|
||||||
+ # been implemented. Just log a warning message and continue.
|
|
||||||
+ if is_npiv_enabled(self.devnum):
|
|
||||||
+ log.warning("zFCP device %s in NPIV mode brought online. All LUNs will be activated "
|
|
||||||
+ "automatically although WWPN and LUN have been provided.", self.devnum)
|
|
||||||
+
|
|
||||||
# create the sysfs directory for the WWPN/port
|
|
||||||
if not os.path.exists(portdir):
|
|
||||||
if os.path.exists(portadd):
|
|
||||||
@@ -327,7 +384,6 @@ class ZFCPDevice(ZFCPDeviceBase):
|
|
||||||
return True
|
|
||||||
else:
|
|
||||||
# newer zfcp sysfs interface with auto port scan
|
|
||||||
- import glob
|
|
||||||
luns = glob.glob("%s/0x????????????????/0x????????????????"
|
|
||||||
% (devdir,))
|
|
||||||
if len(luns) != 0:
|
|
||||||
--
|
|
||||||
2.36.1
|
|
||||||
|
|
||||||
|
|
||||||
From ff01832941a62fc3113983a51a22369566b3f900 Mon Sep 17 00:00:00 2001
|
|
||||||
From: Jan Stodola <jstodola@redhat.com>
|
|
||||||
Date: Sat, 6 Nov 2021 21:27:52 +0100
|
|
||||||
Subject: [PATCH 7/8] Add new class for NPIV-enabled devices
|
|
||||||
|
|
||||||
---
|
|
||||||
blivet/zfcp.py | 53 +++++++++++++++++++++++++++++++++++++++++++++++---
|
|
||||||
1 file changed, 50 insertions(+), 3 deletions(-)
|
|
||||||
|
|
||||||
diff --git a/blivet/zfcp.py b/blivet/zfcp.py
|
|
||||||
index 726e9364..e6c0e48a 100644
|
|
||||||
--- a/blivet/zfcp.py
|
|
||||||
+++ b/blivet/zfcp.py
|
|
||||||
@@ -397,6 +397,44 @@ class ZFCPDevice(ZFCPDeviceBase):
|
|
||||||
return True
|
|
||||||
|
|
||||||
|
|
||||||
+class ZFCPNPIVDevice(ZFCPDeviceBase):
|
|
||||||
+ """Class for zFCP devices configured in NPIV mode. Only a zFCP device number is
|
|
||||||
+ needed for such devices.
|
|
||||||
+ """
|
|
||||||
+
|
|
||||||
+ def online_device(self):
|
|
||||||
+ """Initialize the device and make its storage block device(s) ready to use.
|
|
||||||
+
|
|
||||||
+ :returns: True if success
|
|
||||||
+ :raises: ValueError if the device cannot be initialized
|
|
||||||
+ """
|
|
||||||
+
|
|
||||||
+ super().online_device()
|
|
||||||
+
|
|
||||||
+ if not is_npiv_enabled(self.devnum):
|
|
||||||
+ raise ValueError(_("zFCP device %s cannot be used in NPIV mode.") % self)
|
|
||||||
+
|
|
||||||
+ return True
|
|
||||||
+
|
|
||||||
+ def offline_device(self):
|
|
||||||
+ """Remove the zFCP device from the system.
|
|
||||||
+
|
|
||||||
+ :returns: True if success
|
|
||||||
+ :raises: ValueError if the device cannot be brought offline
|
|
||||||
+ """
|
|
||||||
+
|
|
||||||
+ try:
|
|
||||||
+ self.offline_scsi_device()
|
|
||||||
+ except OSError as e:
|
|
||||||
+ raise ValueError(_("Could not correctly delete SCSI device of "
|
|
||||||
+ "zFCP %(zfcpdev)s (%(e)s).")
|
|
||||||
+ % {'zfcpdev': self, 'e': e})
|
|
||||||
+
|
|
||||||
+ self._set_zfcp_device_offline()
|
|
||||||
+
|
|
||||||
+ return True
|
|
||||||
+
|
|
||||||
+
|
|
||||||
class zFCP:
|
|
||||||
|
|
||||||
""" ZFCP utility class.
|
|
||||||
@@ -439,7 +477,12 @@ class zFCP:
|
|
||||||
|
|
||||||
fields = line.split()
|
|
||||||
|
|
||||||
- if len(fields) == 3:
|
|
||||||
+ # NPIV enabled device
|
|
||||||
+ if len(fields) == 1:
|
|
||||||
+ devnum = fields[0]
|
|
||||||
+ wwpn = None
|
|
||||||
+ fcplun = None
|
|
||||||
+ elif len(fields) == 3:
|
|
||||||
devnum = fields[0]
|
|
||||||
wwpn = fields[1]
|
|
||||||
fcplun = fields[2]
|
|
||||||
@@ -458,8 +501,12 @@ class zFCP:
|
|
||||||
except ValueError as e:
|
|
||||||
log.warning("%s", str(e))
|
|
||||||
|
|
||||||
- def add_fcp(self, devnum, wwpn, fcplun):
|
|
||||||
- d = ZFCPDevice(devnum, wwpn, fcplun)
|
|
||||||
+ def add_fcp(self, devnum, wwpn=None, fcplun=None):
|
|
||||||
+ if wwpn and fcplun:
|
|
||||||
+ d = ZFCPDevice(devnum, wwpn, fcplun)
|
|
||||||
+ else:
|
|
||||||
+ d = ZFCPNPIVDevice(devnum)
|
|
||||||
+
|
|
||||||
if d.online_device():
|
|
||||||
self.fcpdevs.add(d)
|
|
||||||
|
|
||||||
--
|
|
||||||
2.36.1
|
|
||||||
|
|
||||||
|
|
||||||
From ee5b0cdc2393775925fbd9d32caed16eee33fcb0 Mon Sep 17 00:00:00 2001
|
|
||||||
From: Jan Stodola <jstodola@redhat.com>
|
|
||||||
Date: Sat, 20 Nov 2021 23:12:43 +0100
|
|
||||||
Subject: [PATCH 8/8] Generate correct dracut boot arguments for NPIV devices
|
|
||||||
|
|
||||||
NPIV enabled devices need only the device ID. WWPNs/LUNs are discovered
|
|
||||||
automatically by the kernel module.
|
|
||||||
---
|
|
||||||
blivet/devices/disk.py | 10 +++++++++-
|
|
||||||
1 file changed, 9 insertions(+), 1 deletion(-)
|
|
||||||
|
|
||||||
diff --git a/blivet/devices/disk.py b/blivet/devices/disk.py
|
|
||||||
index 67a01ba6..36278507 100644
|
|
||||||
--- a/blivet/devices/disk.py
|
|
||||||
+++ b/blivet/devices/disk.py
|
|
||||||
@@ -577,7 +577,15 @@ class ZFCPDiskDevice(DiskDevice):
|
|
||||||
'lun': self.fcp_lun}
|
|
||||||
|
|
||||||
def dracut_setup_args(self):
|
|
||||||
- return set(["rd.zfcp=%s,%s,%s" % (self.hba_id, self.wwpn, self.fcp_lun,)])
|
|
||||||
+ from ..zfcp import is_npiv_enabled
|
|
||||||
+
|
|
||||||
+ # zFCP devices in NPIV mode need only the device ID
|
|
||||||
+ if is_npiv_enabled(self.hba_id):
|
|
||||||
+ dracut_args = set(["rd.zfcp=%s" % self.hba_id])
|
|
||||||
+ else:
|
|
||||||
+ dracut_args = set(["rd.zfcp=%s,%s,%s" % (self.hba_id, self.wwpn, self.fcp_lun,)])
|
|
||||||
+
|
|
||||||
+ return dracut_args
|
|
||||||
|
|
||||||
|
|
||||||
class DASDDevice(DiskDevice):
|
|
||||||
--
|
|
||||||
2.36.1
|
|
||||||
|
|
@ -19,11 +19,11 @@
|
|||||||
Summary: A python module for system storage configuration
|
Summary: A python module for system storage configuration
|
||||||
Name: python-blivet
|
Name: python-blivet
|
||||||
Url: https://storageapis.wordpress.com/projects/blivet
|
Url: https://storageapis.wordpress.com/projects/blivet
|
||||||
Version: 3.4.0
|
Version: 3.6.0
|
||||||
|
|
||||||
#%%global prerelease .b2
|
#%%global prerelease .b2
|
||||||
# prerelease, if defined, should be something like .a1, .b1, .b2.dev1, or .c2
|
# prerelease, if defined, should be something like .a1, .b1, .b2.dev1, or .c2
|
||||||
Release: 13%{?prerelease}%{?dist}
|
Release: 4%{?prerelease}%{?dist}
|
||||||
Epoch: 1
|
Epoch: 1
|
||||||
License: LGPLv2+
|
License: LGPLv2+
|
||||||
%global realname blivet
|
%global realname blivet
|
||||||
@ -34,23 +34,13 @@ Patch0: 0001-force-lvm-cli.plugin
|
|||||||
Patch1: 0002-remove-btrfs-plugin.patch
|
Patch1: 0002-remove-btrfs-plugin.patch
|
||||||
Patch2: 0003-Revert-More-consistent-lvm-errors.patch
|
Patch2: 0003-Revert-More-consistent-lvm-errors.patch
|
||||||
Patch3: 0004-Revert-Terminology-cleanups.patch
|
Patch3: 0004-Revert-Terminology-cleanups.patch
|
||||||
Patch4: 0005-Fix-activating-old-style-LVM-snapshots.patch
|
Patch4: 0005-DDF-RAID-support-using-mdadm.patch
|
||||||
Patch5: 0006-Fix-resolving-devices-with-names-that-look-like-BIOS.patch
|
Patch5: 0006-Revert-Remove-the-Blivet.roots-attribute.patch
|
||||||
Patch6: 0007-Do-not-set-chunk-size-for-RAID1.patch
|
Patch6: 0007-Fix-potential-AttributeError-when-getting-stratis-bl.patch
|
||||||
Patch7: 0008-Fix-running-tests-in-gating.patch
|
Patch7: 0008-tests-Skip-XFS-resize-test-on-CentOS-RHEL-8.patch
|
||||||
Patch8: 0009-Improve-error-message-printed-for-missing-dependecie.patch
|
Patch8: 0009-Revert-Adjust-to-new-XFS-min-size.patch
|
||||||
Patch9: 0010-Use-bigger-chunk-size-for-thinpools-bigger-than-15.8.patch
|
Patch9: 0010-Catch-BlockDevNotImplementedError-for-btrfs-plugin-c.patch
|
||||||
Patch10: 0011-iscsi-Replace-all-log_exception_info-calls-with-log.patch
|
Patch10: 0011-Default-to-encryption-sector-size-512-for-LUKS-devic.patch
|
||||||
Patch11: 0012-PO-update.patch
|
|
||||||
Patch12: 0013-Fix-getting-PV-info-in-LVMPhysicalVolume-from-the-ca.patch
|
|
||||||
Patch13: 0014-Do-not-crash-when-changing-disklabel-on-disks-with-a.patch
|
|
||||||
Patch14: 0015-ActionDestroyDevice-should-not-obsolete-ActionRemove.patch
|
|
||||||
Patch15: 0016-Correctly-set-vg_name-after-adding-removing-a-PV-fro.patch
|
|
||||||
Patch16: 0017-Use-LVM-PV-format-current_size-in-LVMVolumeGroupDevi.patch
|
|
||||||
Patch17: 0018-Add-support-for-creating-LVM-cache-pools.patch
|
|
||||||
Patch18: 0019-Fix-util.virt_detect-on-Xen.patch
|
|
||||||
Patch19: 0020-Add-support-for-NPIV-enabled-zFCP-devices.patch
|
|
||||||
Patch20: 0021-DDF-RAID-support-using-mdadm.patch
|
|
||||||
|
|
||||||
# Versions of required components (done so we make sure the buildrequires
|
# Versions of required components (done so we make sure the buildrequires
|
||||||
# match the requires versions of things).
|
# match the requires versions of things).
|
||||||
@ -213,6 +203,24 @@ configuration.
|
|||||||
%endif
|
%endif
|
||||||
|
|
||||||
%changelog
|
%changelog
|
||||||
|
* Thu Jan 19 2023 Vojtech Trefny <vtrefny@redhat.com> - 3.6.0-4
|
||||||
|
- Default to encryption sector size 512 for LUKS devices
|
||||||
|
Resolves: rhbz#2160465
|
||||||
|
|
||||||
|
* Thu Nov 03 2022 Vojtech Trefny <vtrefny@redhat.com> - 3.6.0-3
|
||||||
|
- Catch BlockDevNotImplementedError for btrfs plugin calls
|
||||||
|
Resolves: rhbz#2139169
|
||||||
|
- Revert "Adjust to new XFS min size"
|
||||||
|
Resolves: rhbz#2139187
|
||||||
|
|
||||||
|
* Fri Oct 21 2022 Vojtech Trefny <vtrefny@redhat.com> - 3.6.0-2
|
||||||
|
- Skip XFS resize test on CentOS/RHEL 8
|
||||||
|
Related: rhbz#2123712
|
||||||
|
|
||||||
|
* Fri Oct 21 2022 Vojtech Trefny <vtrefny@redhat.com> - 3.6.0-1
|
||||||
|
- Rebase to the latest upstream release 3.6.0
|
||||||
|
Resolves: rhbz#2123712
|
||||||
|
|
||||||
* Thu Aug 18 2022 Vojtech Trefny <vtrefny@redhat.com> - 3.4.0-13
|
* Thu Aug 18 2022 Vojtech Trefny <vtrefny@redhat.com> - 3.4.0-13
|
||||||
- DDF RAID support using mdadm
|
- DDF RAID support using mdadm
|
||||||
Resolves: rhbz#2063791
|
Resolves: rhbz#2063791
|
||||||
@ -239,7 +247,7 @@ configuration.
|
|||||||
- Add support for creating LVM cache pools
|
- Add support for creating LVM cache pools
|
||||||
Resolves: rhbz#2055198
|
Resolves: rhbz#2055198
|
||||||
|
|
||||||
* Thu Jan 10 2022 Vojtech Trefny <vtrefny@redhat.com> - 3.4.0-9
|
* Mon Jan 10 2022 Vojtech Trefny <vtrefny@redhat.com> - 3.4.0-9
|
||||||
- Translation update
|
- Translation update
|
||||||
Resolves: rhbz#2003050
|
Resolves: rhbz#2003050
|
||||||
|
|
||||||
|
Loading…
Reference in New Issue
Block a user