rhel-system-roles/SOURCES/storage-safemode.diff
2021-09-10 03:54:22 +00:00

1022 lines
35 KiB
Diff

diff --git a/README.md b/README.md
index c2debc9..f808adc 100644
--- a/README.md
+++ b/README.md
@@ -73,6 +73,9 @@ The `mount_point` specifies the directory on which the file system will be mount
##### `mount_options`
The `mount_options` specifies custom mount options as a string, e.g.: 'ro'.
+#### `storage_safe_mode`
+When true (the default), an error will occur instead of automatically removing existing devices and/or formatting.
+
Example Playbook
----------------
diff --git a/defaults/main.yml b/defaults/main.yml
index 7b500e5..476616b 100644
--- a/defaults/main.yml
+++ b/defaults/main.yml
@@ -3,6 +3,7 @@
storage_provider: "blivet"
storage_use_partitions: null
storage_disklabel_type: null # leave unset to allow the role to select an appropriate label type
+storage_safe_mode: true # fail instead of implicitly/automatically removing devices or formatting
storage_pool_defaults:
state: "present"
diff --git a/library/blivet.py b/library/blivet.py
index d416944..1d8cd36 100644
--- a/library/blivet.py
+++ b/library/blivet.py
@@ -31,6 +31,9 @@ options:
disklabel_type:
description:
- disklabel type string (eg: 'gpt') to use, overriding the built-in logic in blivet
+ safe_mode:
+ description:
+ - boolean indicating that we should fail rather than implicitly/automatically removing devices or formatting
author:
- David Lehman (dlehman@redhat.com)
@@ -112,13 +115,15 @@ if BLIVET_PACKAGE:
use_partitions = None # create partitions on pool backing device disks?
disklabel_type = None # user-specified disklabel type
+safe_mode = None # do not remove any existing devices or formatting
+packages_only = None # only set things up enough to get a list of required packages
class BlivetAnsibleError(Exception):
pass
-class BlivetVolume:
+class BlivetVolume(object):
def __init__(self, blivet_obj, volume, bpool=None):
self._blivet = blivet_obj
self._volume = volume
@@ -206,11 +211,16 @@ class BlivetVolume:
def _reformat(self):
""" Schedule actions as needed to ensure the volume is formatted as specified. """
+ global packages_only
+
fmt = self._get_format()
if self._device.format.type == fmt.type:
return
- if self._device.format.status:
+ if safe_mode and (self._device.format.type is not None or self._device.format.name != get_format(None).name):
+ raise BlivetAnsibleError("cannot remove existing formatting on volume '%s' in safe mode" % self._volume['name'])
+
+ if self._device.format.status and not packages_only:
self._device.format.teardown()
self._blivet.format_device(self._device, fmt)
@@ -251,6 +261,19 @@ class BlivetDiskVolume(BlivetVolume):
def _type_check(self):
return self._device.is_disk
+ def _look_up_device(self):
+ super(BlivetDiskVolume, self)._look_up_device()
+ if not self._get_device_id():
+ # FAIL: no disks specified for volume
+ raise BlivetAnsibleError("no disks specified for volume '%s'" % self._volume['name']) # sure about this one?
+ elif not isinstance(self._volume['disks'], list):
+ raise BlivetAnsibleError("volume disks must be specified as a list")
+
+ if self._device is None:
+ # FAIL: failed to find the disk
+ raise BlivetAnsibleError("unable to resolve disk specified for volume '%s' (%s)" % (self._volume['name'], self._volume['disks']))
+
+
class BlivetPartitionVolume(BlivetVolume):
def _type_check(self):
@@ -342,7 +365,7 @@ def _get_blivet_volume(blivet_obj, volume, bpool=None):
return _BLIVET_VOLUME_TYPES[volume_type](blivet_obj, volume, bpool=bpool)
-class BlivetPool:
+class BlivetPool(object):
def __init__(self, blivet_obj, pool):
self._blivet = blivet_obj
self._pool = pool
@@ -424,8 +447,11 @@ class BlivetPool:
""" Schedule actions as needed to ensure pool member devices exist. """
members = list()
for disk in self._disks:
- if not disk.isleaf:
- self._blivet.devicetree.recursive_remove(disk)
+ if not disk.isleaf or disk.format.type is not None:
+ if not safe_mode:
+ self._blivet.devicetree.recursive_remove(disk)
+ else:
+ raise BlivetAnsibleError("cannot remove existing formatting and/or devices on disk '%s' (pool '%s') in safe mode" % (disk.name, self._pool['name']))
if use_partitions:
label = get_format("disklabel", device=disk.path)
@@ -486,7 +512,10 @@ class BlivetPartitionPool(BlivetPool):
def _create(self):
if self._device.format.type != "disklabel" or \
self._device.format.label_type != disklabel_type:
- self._blivet.devicetree.recursive_remove(self._device, remove_device=False)
+ if not safe_mode:
+ self._blivet.devicetree.recursive_remove(self._device, remove_device=False)
+ else:
+ raise BlivetAnsibleError("cannot remove existing formatting and/or devices on disk '%s' (pool '%s') in safe mode" % (self._device.name, self._pool['name']))
label = get_format("disklabel", device=self._device.path, label_type=disklabel_type)
self._blivet.format_device(self._device, label)
@@ -520,7 +549,7 @@ class BlivetLVMPool(BlivetPool):
_BLIVET_POOL_TYPES = {
- "disk": BlivetPartitionPool,
+ "partition": BlivetPartitionPool,
"lvm": BlivetLVMPool
}
@@ -550,7 +579,7 @@ def manage_pool(b, pool):
volume['_mount_id'] = bvolume._volume.get('_mount_id', '')
-class FSTab:
+class FSTab(object):
def __init__(self, blivet_obj):
self._blivet = blivet_obj
self._entries = list()
@@ -656,6 +685,7 @@ def run_module():
volumes=dict(type='list'),
packages_only=dict(type='bool', required=False, default=False),
disklabel_type=dict(type='str', required=False, default=None),
+ safe_mode=dict(type='bool', required=False, default=False),
use_partitions=dict(type='bool', required=False, default=True))
# seed the result dict in the object
@@ -684,6 +714,12 @@ def run_module():
global use_partitions
use_partitions = module.params['use_partitions']
+ global safe_mode
+ safe_mode = module.params['safe_mode']
+
+ global packages_only
+ packages_only = module.params['packages_only']
+
b = Blivet()
b.reset()
fstab = FSTab(b)
diff --git a/tasks/main-blivet.yml b/tasks/main-blivet.yml
index 061195c..65b8580 100644
--- a/tasks/main-blivet.yml
+++ b/tasks/main-blivet.yml
@@ -38,7 +38,7 @@
_storage_vols_no_defaults: "{{ _storage_vols_no_defaults|default([]) }} + [{{ item.1 }}]"
_storage_vol_defaults: "{{ _storage_vol_defaults|default([]) }} + [{{ storage_volume_defaults }}]"
_storage_vol_pools: "{{ _storage_vol_pools|default([]) }} + ['{{ item.0.name }}']"
- loop: "{{ _storage_pools|subelements('volumes') }}"
+ loop: "{{ _storage_pools|subelements('volumes', skip_missing=true) }}"
when: storage_pools is defined
- name: Apply defaults to pools and volumes [3/6]
@@ -85,6 +85,15 @@
- debug:
var: _storage_volumes
+- name: load mount facts
+ setup:
+ gather_subset: '!all,!min,mounts'
+ register: __storage_mounts_before_packages
+
+# - name: show mounts before get required packages
+# debug:
+# var: __storage_mounts_before_packages
+
- name: get required packages
blivet:
pools: "{{ _storage_pools }}"
@@ -94,6 +103,30 @@
packages_only: true
register: package_info
+- name: load mount facts
+ setup:
+ gather_subset: '!all,!min,mounts'
+ register: __storage_mounts_after_packages
+
+- name: detect mount alteration by 'get required packages'
+ block:
+ - name: show mounts before manage the pools and volumes
+ debug:
+ var: __storage_mounts_before_packages.ansible_facts.ansible_mounts
+
+ - name: show mounts after manage the pools and volumes
+ debug:
+ var: __storage_mounts_after_packages.ansible_facts.ansible_mounts
+
+ - name: fail if mounts changed
+ fail:
+ msg: "get required packages changed mounts. Changed status is
+ {{ package_info.changed }}"
+ when:
+ - __storage_mounts_before_packages.ansible_facts.ansible_mounts |
+ count !=
+ __storage_mounts_after_packages.ansible_facts.ansible_mounts | count
+
- name: make sure required packages are installed
package:
name: "{{ package_info.packages }}"
@@ -105,6 +138,7 @@
volumes: "{{ _storage_volumes }}"
use_partitions: "{{ storage_use_partitions }}"
disklabel_type: "{{ storage_disklabel_type }}"
+ safe_mode: "{{ storage_safe_mode }}"
register: blivet_output
- debug:
diff --git a/tests/get_unused_disk.yml b/tests/get_unused_disk.yml
index 9f4c5d2..79e952a 100644
--- a/tests/get_unused_disk.yml
+++ b/tests/get_unused_disk.yml
@@ -9,12 +9,10 @@
unused_disks: "{{ unused_disks_return.disks }}"
when: "'Unable to find unused disk' not in unused_disks_return.disks"
-- block:
- - name: Exit playbook when there's no unused disks in the system
- debug:
- msg: "Unable to find unused disks. Exiting playbook."
- - meta: end_play
- when: unused_disks is undefined
+- name: Exit playbook when there's not enough unused disks in the system
+ fail:
+ msg: "Unable to find enough unused disks. Exiting playbook."
+ when: unused_disks is undefined or unused_disks|length < disks_needed|default(1)
- name: Print unused disks
debug:
diff --git a/tests/tests_change_disk_fs.yml b/tests/tests_change_disk_fs.yml
index b6aa80b..f7962c6 100644
--- a/tests/tests_change_disk_fs.yml
+++ b/tests/tests_change_disk_fs.yml
@@ -2,6 +2,7 @@
- hosts: all
become: true
vars:
+ storage_safe_mode: false
mount_location: '/opt/test'
volume_size: '5g'
fs_type_after: "{{ 'ext3' if (ansible_distribution == 'RedHat' and ansible_distribution_major_version == '6') else 'ext4' }}"
diff --git a/tests/tests_change_fs.yml b/tests/tests_change_fs.yml
index cca23eb..b88e768 100644
--- a/tests/tests_change_fs.yml
+++ b/tests/tests_change_fs.yml
@@ -2,6 +2,7 @@
- hosts: all
become: true
vars:
+ storage_safe_mode: false
mount_location: '/opt/test1'
volume_size: '5g'
fs_after: "{{ (ansible_distribution == 'RedHat' and ansible_distribution_major_version == '6') | ternary('ext4', 'xfs') }}"
diff --git a/tests/tests_change_fs_use_partitions.yml b/tests/tests_change_fs_use_partitions.yml
index e4aa76c..eb93c11 100644
--- a/tests/tests_change_fs_use_partitions.yml
+++ b/tests/tests_change_fs_use_partitions.yml
@@ -2,6 +2,7 @@
- hosts: all
become: true
vars:
+ storage_safe_mode: false
storage_use_partitions: true
mount_location: '/opt/test1'
volume_size: '5g'
diff --git a/tests/tests_create_disk_then_remove.yml b/tests/tests_create_disk_then_remove.yml
index b19ae35..c5290eb 100644
--- a/tests/tests_create_disk_then_remove.yml
+++ b/tests/tests_create_disk_then_remove.yml
@@ -2,6 +2,7 @@
- hosts: all
become: true
vars:
+ storage_safe_mode: false
mount_location: '/opt/test1'
tasks:
diff --git a/tests/tests_create_lvm_pool_then_remove.yml b/tests/tests_create_lvm_pool_then_remove.yml
index 6b25939..f2c06fb 100644
--- a/tests/tests_create_lvm_pool_then_remove.yml
+++ b/tests/tests_create_lvm_pool_then_remove.yml
@@ -2,6 +2,7 @@
- hosts: all
become: true
vars:
+ storage_safe_mode: false
mount_location1: '/opt/test1'
mount_location2: '/opt/test2'
volume_group_size: '10g'
diff --git a/tests/tests_create_partition_volume_then_remove.yml b/tests/tests_create_partition_volume_then_remove.yml
index 40b3e62..ae589d3 100644
--- a/tests/tests_create_partition_volume_then_remove.yml
+++ b/tests/tests_create_partition_volume_then_remove.yml
@@ -2,6 +2,7 @@
- hosts: all
become: true
vars:
+ storage_safe_mode: false
mount_location: '/opt/test1'
tasks:
@@ -18,7 +19,7 @@
vars:
storage_pools:
- name: "{{ unused_disks[0] }}"
- type: disk
+ type: partition
disks: "{{ unused_disks }}"
volumes:
- name: test1
@@ -33,7 +34,7 @@
vars:
storage_pools:
- name: "{{ unused_disks[0] }}"
- type: disk
+ type: partition
disks: "{{ unused_disks }}"
volumes:
- name: test1
@@ -48,7 +49,7 @@
vars:
storage_pools:
- name: "{{ unused_disks[0] }}"
- type: disk
+ type: partition
disks: "{{ unused_disks }}"
state: absent
volumes:
@@ -65,7 +66,7 @@
vars:
storage_pools:
- name: "{{ unused_disks[0] }}"
- type: disk
+ type: partition
disks: "{{ unused_disks }}"
state: absent
volumes:
diff --git a/tests/tests_disk_errors.yml b/tests/tests_disk_errors.yml
index 36eec41..7112f6e 100644
--- a/tests/tests_disk_errors.yml
+++ b/tests/tests_disk_errors.yml
@@ -3,8 +3,17 @@
become: true
vars:
mount_location: '/opt/test1'
+ testfile: "{{ mount_location }}/quux"
tasks:
+ - include_role:
+ name: storage
+
+ - include_tasks: get_unused_disk.yml
+ vars:
+ min_size: "10g"
+ max_return: 1
+
- name: Verify that the play fails with the expected error message
block:
- name: Create a disk volume mounted at "{{ mount_location }}"
@@ -14,11 +23,246 @@
storage_volumes:
- name: test1
type: disk
- disks: "['/dev/surelyidonotexist']"
+ disks: ['/dev/surelyidonotexist']
mount_point: "{{ mount_location }}"
- - name: Check the error output
+ - name: UNREACH
+ fail:
+ msg: "this should be unreachable"
+
+ rescue:
+ - name: Check that we failed in the role
+ assert:
+ that:
+ - ansible_failed_task.name != 'UNREACH'
+ msg: "Role has not failed when it should have"
+ vars:
+ # Ugh! needed to expand ansible_failed_task
+ storage_provider: blivet
+
+ # the following does not work properly,
+ # blivet_output.failed is false.
+ # - name: Show the error output
+ # debug:
+ # msg: "{{ blivet_output.failed }}"
+
+ # - name: Check the error output
+ # assert:
+ # that: blivet_output.failed | bool
+ # msg: "Expected error message not found for missing disk"
+
+ - name: Create a file system on disk
+ include_role:
+ name: storage
+ vars:
+ storage_volumes:
+ - name: test1
+ type: disk
+ fs_type: 'ext4'
+ disks: "{{ unused_disks }}"
+ mount_point: "{{ mount_location }}"
+
+ - name: create a file
+ file:
+ path: "{{ testfile }}"
+ state: touch
+
+ - name: Test for correct handling of safe_mode
+ block:
+ - name: Try to replace the file system on disk in safe mode
+ include_role:
+ name: storage
+ vars:
+ storage_volumes:
+ - name: test1
+ type: disk
+ fs_type: 'ext3'
+ disks: "{{ unused_disks }}"
+
+ - name: UNREACH
+ fail:
+ msg: "this should be unreachable"
+
+ rescue:
+ - name: Check that we failed in the role
+ assert:
+ that:
+ - ansible_failed_task.name != 'UNREACH'
+ msg: "Role has not failed when it should have"
+ vars:
+ # Ugh! needed to expand ansible_failed_task
+ storage_provider: blivet
+
+ - name: Verify the output
+ assert:
+ that: "blivet_output.failed and
+ blivet_output.msg|regex_search('cannot remove existing formatting on volume.*in safe mode') and
+ not blivet_output.changed"
+ msg: "Unexpected behavior w/ existing data on specified disks"
+
+ - name: Unmount file system
+ include_role:
+ name: storage
+ vars:
+ storage_volumes:
+ - name: test1
+ type: disk
+ fs_type: 'ext4'
+ disks: "{{ unused_disks }}"
+ mount_point: none
+
+ - name: Test for correct handling of safe_mode with unmounted filesystem
+ block:
+ - name: Try to replace the file system on disk in safe mode
+ include_role:
+ name: storage
+ vars:
+ storage_volumes:
+ - name: test1
+ type: disk
+ fs_type: 'ext3'
+ disks: "{{ unused_disks }}"
+
+ - name: UNREACH
+ fail:
+ msg: "this should be unreachable"
+
+ rescue:
+ - name: Check that we failed in the role
assert:
- that: "{{ blivet_output.failed }}"
- msg: "Expected error message not found for missing disk"
- ignore_errors: yes
+ that:
+ - ansible_failed_task.name != 'UNREACH'
+ msg: "Role has not failed when it should have"
+ vars:
+ # Ugh! needed to expand ansible_failed_task
+ storage_provider: blivet
+
+ - name: Verify the output
+ assert:
+ that: "blivet_output.failed and
+ blivet_output.msg|regex_search('cannot remove existing formatting on volume.*in safe mode') and
+ not blivet_output.changed"
+ msg: "Unexpected behavior w/ existing data on specified disks"
+
+ - name: Remount file system
+ include_role:
+ name: storage
+ vars:
+ storage_volumes:
+ - name: test1
+ type: disk
+ fs_type: 'ext4'
+ disks: "{{ unused_disks }}"
+ mount_point: "{{ mount_location }}"
+
+ - name: stat the file
+ stat:
+ path: "{{ testfile }}"
+ register: stat_r
+
+ - name: assert file presence
+ assert:
+ that:
+ stat_r.stat.isreg is defined and stat_r.stat.isreg
+ msg: "data lost!"
+
+ - name: Test for correct handling of safe_mode
+ block:
+ - name: Try to create a partition pool on the disk already containing a file system in safe_mode
+ include_role:
+ name: storage
+ vars:
+ storage_pools:
+ - name: foo
+ disks: "{{ unused_disks }}"
+ type: partition
+
+ - name: UNREACH
+ fail:
+ msg: "this should be unreachable"
+
+ rescue:
+ - name: Check that we failed in the role
+ assert:
+ that:
+ - ansible_failed_task.name != 'UNREACH'
+ msg: "Role has not failed when it should have"
+ vars:
+ # Ugh! needed to expand ansible_failed_task
+ storage_provider: blivet
+
+ - name: Verify the output
+ assert:
+ that: "blivet_output.failed and
+ blivet_output.msg|regex_search('cannot remove existing formatting and/or devices on disk.*in safe mode') and
+ not blivet_output.changed"
+ msg: "Unexpected behavior w/ existing data on specified disks"
+
+ - name: Test for correct handling of safe_mode with existing filesystem
+ block:
+ - name: Try to create LVM pool on disk that already belongs to an existing filesystem
+ include_role:
+ name: storage
+ vars:
+ storage_pools:
+ - name: foo
+ disks: "{{ unused_disks }}"
+ type: lvm
+
+ - name: UNREACH
+ fail:
+ msg: "this should be unreachable"
+
+ rescue:
+ - name: Check that we failed in the role
+ assert:
+ that:
+ - ansible_failed_task.name != 'UNREACH'
+ msg: "Role has not failed when it should have"
+ vars:
+ # Ugh! needed to expand ansible_failed_task
+ storage_provider: blivet
+
+ - name: Verify the output
+ assert:
+ that: "{{ blivet_output.failed and
+ blivet_output.msg|regex_search('cannot remove existing formatting and/or devices on disk.*in safe mode') and
+ not blivet_output.changed }}"
+ msg: "Unexpected behavior w/ existing data on specified disks"
+
+ - name: stat the file
+ stat:
+ path: "{{ testfile }}"
+ register: stat_r
+
+ - name: assert file presence
+ assert:
+ that:
+ stat_r.stat.isreg is defined and stat_r.stat.isreg
+ msg: "data lost!"
+
+ - name: Create a partition pool on the disk already containing a file system w/o safe_mode
+ include_role:
+ name: storage
+ vars:
+ storage_safe_mode: false
+ storage_pools:
+ - name: foo
+ disks: "{{ unused_disks }}"
+ type: partition
+
+ - name: Verify the output
+ assert:
+ that: not blivet_output.failed
+ msg: "failed to create partition pool over existing file system w/o safe_mode"
+
+ - name: Clean up
+ include_role:
+ name: storage
+ vars:
+ storage_safe_mode: false
+ storage_pools:
+ - name: foo
+ type: partition
+ disks: "{{ unused_disks }}"
+ state: absent
diff --git a/tests/tests_lvm_errors.yml b/tests/tests_lvm_errors.yml
index ab23674..e8be153 100644
--- a/tests/tests_lvm_errors.yml
+++ b/tests/tests_lvm_errors.yml
@@ -33,13 +33,32 @@
size: "{{ volume1_size }}"
mount_point: "{{ mount_location1 }}"
- - name: Verify the output
+ - name: UNREACH
+ fail:
+ msg: "this should be unreachable"
+
+ rescue:
+ - name: Check that we failed in the role
assert:
- that: "{{ blivet_output.failed and
- blivet_output.msg|regex_search('unable to resolve.+disk')|length>0 and
- not blivet_output.changed }}"
- msg: "Unexpected behavior w/ non-existent pool disk"
- ignore_errors: yes
+ that:
+ - ansible_failed_task.name != 'UNREACH'
+ msg: "Role has not failed when it should have"
+ vars:
+ # Ugh! needed to expand ansible_failed_task
+ storage_provider: blivet
+
+ # the following does not work properly
+ # - debug:
+ # msg: "{{ 'failed: ' + blivet_output.failed | string +
+ # 'msg: ' + blivet_output.msg +
+ # 'changed: ' + blivet_output.changed | string }}"
+
+ # - name: Verify the output
+ # assert:
+ # that: "{{ blivet_output.failed and
+ # blivet_output.msg|regex_search('unable to resolve.+disk')|length>0 and
+ # not blivet_output.changed }}"
+ # msg: "Unexpected behavior w/ non-existent pool disk"
- name: Test for correct handling of invalid size specification.
block:
@@ -55,13 +74,27 @@
size: "{{ invalid_size }}"
mount_point: "{{ mount_location1 }}"
- - name: Verify the output
+ - name: UNREACH
+ fail:
+ msg: "this should be unreachable"
+
+ rescue:
+ - name: Check that we failed in the role
assert:
- that: "{{ blivet_output.failed and
- blivet_output.msg|regex_search('invalid size.+for volume') and
- not blivet_output.changed }}"
- msg: "Unexpected behavior w/ invalid volume size"
- ignore_errors: yes
+ that:
+ - ansible_failed_task.name != 'UNREACH'
+ msg: "Role has not failed when it should have"
+ vars:
+ # Ugh! needed to expand ansible_failed_task
+ storage_provider: blivet
+
+ # the following does not work properly
+ # - name: Verify the output
+ # assert:
+ # that: "{{ blivet_output.failed and
+ # blivet_output.msg|regex_search('invalid size.+for volume') and
+ # not blivet_output.changed }}"
+ # msg: "Unexpected behavior w/ invalid volume size"
- name: Test for correct handling of too-large volume size.
block:
@@ -77,13 +110,27 @@
size: "{{ too_large_size }}"
mount_point: "{{ mount_location1 }}"
- - name: Verify the output
+ - name: UNREACH
+ fail:
+ msg: "this should be unreachable"
+
+ rescue:
+ - name: Check that we failed in the role
assert:
- that: "{{ blivet_output.failed and
- blivet_output.msg|regex_search('size.+exceeds.+space in pool') and
- not blivet_output.changed }}"
- msg: "Unexpected behavior w/ too-large volume size"
- ignore_errors: yes
+ that:
+ - ansible_failed_task.name != 'UNREACH'
+ msg: "Role has not failed when it should have"
+ vars:
+ # Ugh! needed to expand ansible_failed_task
+ storage_provider: blivet
+
+ # the following does not work properly
+ # - name: Verify the output
+ # assert:
+ # that: "{{ blivet_output.failed and
+ # blivet_output.msg|regex_search('size.+exceeds.+space in pool') and
+ # not blivet_output.changed }}"
+ # msg: "Unexpected behavior w/ too-large volume size"
- name: Test for correct handling of non-list disk specification.
block:
@@ -99,13 +146,27 @@
size: "{{ too_large_size }}"
mount_point: "{{ mount_location1 }}"
- - name: Verify the output
+ - name: UNREACH
+ fail:
+ msg: "this should be unreachable"
+
+ rescue:
+ - name: Check that we failed in the role
assert:
- that: "{{ blivet_output.failed and
- blivet_output.msg|regex_search('disk.+list') and
- not blivet_output.changed }}"
- msg: "Unexpected behavior w/ disks not in list form"
- ignore_errors: yes
+ that:
+ - ansible_failed_task.name != 'UNREACH'
+ msg: "Role has not failed when it should have"
+ vars:
+ # Ugh! needed to expand ansible_failed_task
+ storage_provider: blivet
+
+ # the following does not work properly
+ # - name: Verify the output
+ # assert:
+ # that: "{{ blivet_output.failed and
+ # blivet_output.msg|regex_search('disk.+list') and
+ # not blivet_output.changed }}"
+ # msg: "Unexpected behavior w/ disks not in list form"
- name: Test for correct handling of missing disk specification.
block:
@@ -121,13 +182,27 @@
size: "{{ too_large_size }}"
mount_point: "{{ mount_location1 }}"
- - name: Verify the output
+ - name: UNREACH
+ fail:
+ msg: "this should be unreachable"
+
+ rescue:
+ - name: Check that we failed in the role
assert:
- that: "{{ blivet_output.failed and
- blivet_output.msg|regex_search('no disks.+pool') and
- not blivet_output.changed }}"
- msg: "Unexpected behavior w/ no disks specified"
- ignore_errors: yes
+ that:
+ - ansible_failed_task.name != 'UNREACH'
+ msg: "Role has not failed when it should have"
+ vars:
+ # Ugh! needed to expand ansible_failed_task
+ storage_provider: blivet
+
+ # the following does not work properly
+ # - name: Verify the output
+ # assert:
+ # that: "{{ blivet_output.failed and
+ # blivet_output.msg|regex_search('no disks.+pool') and
+ # not blivet_output.changed }}"
+ # msg: "Unexpected behavior w/ no disks specified"
- name: Test for correct handling of LVM volume not defined within a pool.
block:
@@ -142,10 +217,179 @@
size: "{{ volume1_size }}"
mount_point: "{{ mount_location1 }}"
+ - name: UNREACH
+ fail:
+ msg: "this should be unreachable"
+
+ rescue:
+ - name: Check that we failed in the role
+ assert:
+ that:
+ - ansible_failed_task.name != 'UNREACH'
+ msg: "Role has not failed when it should have"
+ vars:
+ # Ugh! needed to expand ansible_failed_task
+ storage_provider: blivet
+
+ # the following does not work properly
+ # - name: Verify the output
+ # assert:
+ # that: "{{ blivet_output.failed and
+ # blivet_output.msg|regex_search('failed to find pool .+ for volume') and
+ # not blivet_output.changed }}"
+ # msg: "Unexpected behavior w/ LVM volume defined outside of any pool"
+
+ - name: Create a pool
+ include_role:
+ name: storage
+ vars:
+ storage_pools:
+ - name: testpool1
+ type: lvm
+ disks: "{{ unused_disks }}"
+ volumes:
+ - name: testvol1
+ fs_type: 'ext4'
+ size: '1g'
+
+ - name: Test for correct handling of safe_mode
+ block:
+ - name: Try to replace file system in safe mode
+ include_role:
+ name: storage
+ vars:
+ storage_pools:
+ - name: testpool1
+ type: lvm
+ disks: "{{ unused_disks }}"
+ volumes:
+ - name: testvol1
+ fs_type: 'ext3'
+ size: '1g'
+
+ - name: UNREACH
+ fail:
+ msg: "this should be unreachable"
+
+ rescue:
+ - name: Check that we failed in the role
+ assert:
+ that:
+ - ansible_failed_task.name != 'UNREACH'
+ msg: "Role has not failed when it should have"
+ vars:
+ # Ugh! needed to expand ansible_failed_task
+ storage_provider: blivet
+
- name: Verify the output
assert:
that: "{{ blivet_output.failed and
- blivet_output.msg|regex_search('failed to find pool .+ for volume') and
+ blivet_output.msg|regex_search('cannot remove existing formatting on volume.*in safe mode') and
not blivet_output.changed }}"
- msg: "Unexpected behavior w/ LVM volume defined outside of any pool"
- ignore_errors: yes
+ msg: "Unexpected behavior w/ existing data on specified disks"
+
+ - name: Test for correct handling of safe_mode with resize
+ block:
+ - name: Try to resize in safe mode
+ include_role:
+ name: storage
+ vars:
+ storage_pools:
+ - name: testpool1
+ type: lvm
+ disks: "{{ unused_disks }}"
+ volumes:
+ - name: testvol1
+ fs_type: 'ext4'
+ size: '2g'
+
+ - name: Verify the output
+ assert:
+ that: "{{ not blivet_output.failed and blivet_output.changed }}"
+ msg: "Unexpected behavior w/ existing data on specified disks"
+
+ when: false
+
+ - name: Test for correct handling of safe_mode with existing pool
+ block:
+ - name: Try to create LVM pool on disks that already belong to an existing pool
+ include_role:
+ name: storage
+ vars:
+ storage_pools:
+ - name: foo
+ disks: "{{ unused_disks }}"
+ type: lvm
+
+ - name: UNREACH
+ fail:
+ msg: "this should be unreachable"
+
+ rescue:
+ - name: Check that we failed in the role
+ assert:
+ that:
+ - ansible_failed_task.name != 'UNREACH'
+ msg: "Role has not failed when it should have"
+ vars:
+ # Ugh! needed to expand ansible_failed_task
+ storage_provider: blivet
+
+ - name: Verify the output
+ assert:
+ that: "{{ blivet_output.failed and
+ blivet_output.msg|regex_search('cannot remove existing formatting and/or devices on disk.*in safe mode') and
+ not blivet_output.changed }}"
+ msg: "Unexpected behavior w/ existing data on specified disks"
+
+ - name: Test for correct handling of safe_mode
+ block:
+ - name: Try to replace a pool by a file system on disk in safe mode
+ include_role:
+ name: storage
+ vars:
+ storage_volumes:
+ - name: test1
+ type: disk
+ fs_type: 'ext3'
+ disks:
+ - "{{ unused_disks[0] }}"
+
+ - name: UNREACH
+ fail:
+ msg: "this should be unreachable"
+
+ rescue:
+ - name: Check that we failed in the role
+ assert:
+ that:
+ - ansible_failed_task.name != 'UNREACH'
+ msg: "Role has not failed when it should have"
+ vars:
+ # Ugh! needed to expand ansible_failed_task
+ storage_provider: blivet
+
+ - name: Verify the output
+ assert:
+ that: "blivet_output.failed and
+ blivet_output.msg|regex_search('cannot remove existing formatting on volume.*in safe mode') and
+ not blivet_output.changed"
+ msg: "Unexpected behavior w/ existing data on specified disks"
+
+ - name: Verify the output
+ assert:
+ that: "blivet_output.failed and
+ blivet_output.msg|regex_search('cannot remove existing formatting on volume.*in safe mode') and
+ not blivet_output.changed"
+ msg: "Unexpected behavior w/ existing data on specified disks"
+
+ - name: Clean up
+ include_role:
+ name: storage
+ vars:
+ storage_safe_mode: false
+ storage_pools:
+ - name: testpool1
+ type: lvm
+ disks: "{{ unused_disks }}"
+ state: absent
diff --git a/tests/tests_lvm_multiple_disks_multiple_volumes.yml b/tests/tests_lvm_multiple_disks_multiple_volumes.yml
index bbc7bb0..ca3968f 100644
--- a/tests/tests_lvm_multiple_disks_multiple_volumes.yml
+++ b/tests/tests_lvm_multiple_disks_multiple_volumes.yml
@@ -15,13 +15,7 @@
vars:
min_size: "{{ volume_group_size }}"
max_return: 2
-
- - block:
- - debug:
- msg: "There needs to be two unused disks in the system to run this playbook."
- - name: End playbook if there isn't two disks available
- meta: end_play
- when: unused_disks|length < 2
+ disks_needed: 2
- name: Create a logical volume spanning two physical volumes that changes its mount location
include_role: