
Having no luck here. I've had a read on the LVM config usage and there were no filters enabled in lvm.conf. I enabled debug logging and can see the default global filter being applied. I then manually forced the 'all' fiter and 'pvcreate /dev/sdb' still tells me it is excluded by a filter. The device section from 'lvmconfig' follows. What's wrong? devices { dir="/dev" scan="/dev" obtain_device_list_from_udev=1 external_device_info_source="none" preferred_names=["^/dev/mpath/","^/dev/mapper/mpath","^/dev/[hs]d"] filter="a|.*/|" cache_dir="/etc/lvm/cache" cache_file_prefix="" write_cache_state=1 sysfs_scan=1 scan_lvs=0 multipath_component_detection=1 md_component_detection=1 fw_raid_component_detection=0 md_chunk_alignment=1 data_alignment_detection=1 data_alignment=0 data_alignment_offset_detection=1 ignore_suspended_devices=0 ignore_lvm_mirrors=1 disable_after_error_count=0 require_restorefile_with_uuid=1 pv_min_size=2048 issue_discards=0 allow_changes_with_duplicate_pvs=1 } On Wed, Apr 29, 2020 at 6:21 PM Shareef Jalloq <shareef@jalloq.co.uk> wrote:
Actually, now I've fixed that, indeed, the deployment now fails with an lvm filter error. I'm not familiar with filters but there aren't any uncommented instances of 'filter' in /etc/lvm/lvm.conf.
On Wed, Apr 29, 2020 at 5:42 PM Shareef Jalloq <shareef@jalloq.co.uk> wrote:
Ah of course. I was assuming something had gone wrong with the deployment and it couldn't clean up its own mess. I'll raise a bug on the documentation.
Strahil, what are the other options to using /dev/sdxxx?
On Wed, Apr 29, 2020 at 10:17 AM Strahil Nikolov <hunter86_bg@yahoo.com> wrote:
On April 29, 2020 2:39:05 AM GMT+03:00, Jayme <jaymef@gmail.com> wrote:
Has the drive been used before, it might have existing partition/filesystem on it? If you are sure it's fine to overwrite try running wipefs -a /dev/sdb on all hosts. Also make sure there aren't any filters setup in lvm.conf (there shouldn't be on fresh install, but worth checking).
On Tue, Apr 28, 2020 at 8:22 PM Shareef Jalloq <shareef@jalloq.co.uk> wrote:
Hi,
I'm running the gluster deployment flow and am trying to use a second drive as the gluster volume. It's /dev/sdb on each node and I'm using the JBOD mode.
I'm seeing the following gluster ansible task fail and a google search doesn't bring up much.
TASK [gluster.infra/roles/backend_setup : Create volume groups] ****************
failed: [ovirt-gluster-01.jalloq.co.uk] (item={u'vgname': u'gluster_vg_sdb', u'pvname': u'/dev/sdb'}) => {"ansible_loop_var": "item", "changed": false, "err": " Couldn't find device with uuid Y8FVs8-LP6w-R6CR-Yosh-c40j-17XP-ttP3Np.\n Couldn't find device with uuid tA4lpO-hM9f-S8ci-BdPh-lTve-0Rh1-3Bcsfy.\n Couldn't find device with uuid RG3w6j-yrxn-2iMw-ngd0-HgMS-i5dP-CGjaRk.\n Couldn't find device with uuid lQV02e-TUZE-PXCd-GWEd-eGqe-c2xC-pauHG7.\n Device /dev/sdb excluded by a filter.\n", "item": {"pvname": "/dev/sdb", "vgname": "gluster_vg_sdb"}, "msg": "Creating physical volume '/dev/sdb' failed", "rc": 5} _______________________________________________ Users mailing list -- users@ovirt.org To unsubscribe send an email to users-leave@ovirt.org Privacy Statement: https://www.ovirt.org/privacy-policy.html oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/ List Archives:
https://lists.ovirt.org/archives/list/users@ovirt.org/message/5U3K3IPYCFOLUF...
Actually best practice is not to use /dev/sdxxx as they can change.
In your case most peobably the LUN is not fresh, so wipe it with dd/blktrim so any remnants of old FS signature is gone.
Best Regards, Strahil Nikolov