
----- Original Message -----
From: "Stefano Danzi" <s.danzi@hawai.it> To: "Nir Soffer" <nsoffer@redhat.com> Sent: Thursday, February 5, 2015 1:39:41 PM Subject: Re: [ovirt-users] Self hosted engine iusses
Here
In vdsm log I see that you installed hosted engine on NFS storage domain; not having multipath devices and ovirt vgs/lvs is expected. Your multipath errors may be related to ovirt, but only because vdsm requires and starts multipathd, and install new multipath.conf. I suggest to try to get help about it in device-mapper channels (e.g. #lvm on freenode). If you cannot resolve this, please open a bug. Nir
Il 05/02/2015 12.31, Nir Soffer ha scritto:
----- Original Message -----
From: "Stefano Danzi" <s.danzi@hawai.it> To: "Nir Soffer" <nsoffer@redhat.com> Cc: users@ovirt.org Sent: Thursday, February 5, 2015 1:17:01 PM Subject: Re: [ovirt-users] Self hosted engine iusses
Il 05/02/2015 12.08, Nir Soffer ha scritto:
----- Original Message -----
From: "Stefano Danzi" <s.danzi@hawai.it> To: "Nir Soffer" <nsoffer@redhat.com> Cc: users@ovirt.org Sent: Thursday, February 5, 2015 12:58:35 PM Subject: Re: [ovirt-users] Self hosted engine iusses
Il 05/02/2015 11.52, Nir Soffer ha scritto:
----- Original Message -----
After ovirt installation on host console I see this error every 5 minutes:
[ 1823.837020] device-mapper: table: 253:4: multipath: error getting device [ 1823.837228] device-mapper: ioctl: error adding target to table This may be caused by the fact that vdsm does not cleanup properly after deactivating storage domains. We have an open bugs on this.
You may have an active lv using non-existent multipath device.
Can you share with us the output of:
lsblk multipath -ll dmsetup table cat /etc/multipath.conf pvscan --cache > /dev/null && lvs
Nir
See above:
[root@ovirt01 etc]# lsblk NAME MAJ:MIN RM SIZE RO TYPE MOUNTPOINT sda 8:0 0 931,5G 0 disk ├─sda1 8:1 0 500M 0 part │ └─md0 9:0 0 500M 0 raid1 /boot └─sda2 8:2 0 931G 0 part └─md1 9:1 0 930,9G 0 raid1 ├─centos_ovirt01-swap 253:0 0 7,9G 0 lvm [SWAP] ├─centos_ovirt01-root 253:1 0 50G 0 lvm / ├─centos_ovirt01-home 253:2 0 10G 0 lvm /home └─centos_ovirt01-glusterOVEngine 253:3 0 50G 0 lvm /home/glusterfs/engine sdb 8:16 0 931,5G 0 disk ├─sdb1 8:17 0 500M 0 part │ └─md0 9:0 0 500M 0 raid1 /boot └─sdb2 8:18 0 931G 0 part └─md1 9:1 0 930,9G 0 raid1 ├─centos_ovirt01-swap 253:0 0 7,9G 0 lvm [SWAP] ├─centos_ovirt01-root 253:1 0 50G 0 lvm / ├─centos_ovirt01-home 253:2 0 10G 0 lvm /home └─centos_ovirt01-glusterOVEngine 253:3 0 50G 0 lvm /home/glusterfs/engine
[root@ovirt01 etc]# multipath -ll Feb 05 11:56:25 | multipath.conf +5, invalid keyword: getuid_callout Feb 05 11:56:25 | multipath.conf +18, invalid keyword: getuid_callout Feb 05 11:56:25 | multipath.conf +37, invalid keyword: getuid_callout
[root@ovirt01 etc]# dmsetup table centos_ovirt01-home: 0 20971520 linear 9:1 121391104 centos_ovirt01-swap: 0 16531456 linear 9:1 2048 centos_ovirt01-root: 0 104857600 linear 9:1 16533504 centos_ovirt01-glusterOVEngine: 0 104857600 linear 9:1 142362624
[root@ovirt01 etc]# cat /etc/multipath.conf # RHEV REVISION 1.1
defaults { polling_interval 5 getuid_callout "/usr/lib/udev/scsi_id --whitelisted --replace-whitespace --device=/dev/%n" no_path_retry fail user_friendly_names no flush_on_last_del yes fast_io_fail_tmo 5 dev_loss_tmo 30 max_fds 4096 }
devices { device { vendor "HITACHI" product "DF.*" getuid_callout "/usr/lib/udev/scsi_id --whitelisted --replace-whitespace --device=/dev/%n" } device { vendor "COMPELNT" product "Compellent Vol" no_path_retry fail } device { # multipath.conf.default vendor "DGC" product ".*" product_blacklist "LUNZ" path_grouping_policy "group_by_prio" path_checker "emc_clariion" hardware_handler "1 emc" prio "emc" failback immediate rr_weight "uniform" # vdsm required configuration getuid_callout "/usr/lib/udev/scsi_id --whitelisted --replace-whitespace --device=/dev/%n" features "0" no_path_retry fail } }
[root@ovirt01 etc]# pvscan --cache > /dev/null && lvs Incorrect metadata area header checksum on /dev/sda2 at offset 4096 Incorrect metadata area header checksum on /dev/sda2 at offset 4096 LV VG Attr LSize Pool Origin Data% Move Log Cpy%Sync Convert glusterOVEngine centos_ovirt01 -wi-ao---- 50,00g home centos_ovirt01 -wi-ao---- 10,00g root centos_ovirt01 -wi-ao---- 50,00g swap centos_ovirt01 -wi-ao---- 7,88g Are you sure this is the correct host that the multipath error came from?
There are no multipath devices in this host and no ovirt storage domains lvs.
Nir
Yes this is the host. I'm sure. I've not yet configured ovirt storage domains (only installed ovirt on host and self hosted engine VM)
Here a part of /var/log/messages:
Feb 5 10:04:43 ovirt01 kernel: device-mapper: table: 253:4: multipath: error getting device Feb 5 10:04:43 ovirt01 kernel: device-mapper: ioctl: error adding target to table Feb 5 10:04:43 ovirt01 multipathd: dm-4: remove map (uevent) Feb 5 10:04:43 ovirt01 multipathd: dm-4: remove map (uevent) Feb 5 10:04:43 ovirt01 kernel: device-mapper: table: 253:4: multipath: error getting device Feb 5 10:04:43 ovirt01 kernel: device-mapper: ioctl: error adding target to table Feb 5 10:04:43 ovirt01 multipathd: dm-4: remove map (uevent) Feb 5 10:04:43 ovirt01 multipathd: dm-4: remove map (uevent)
and kernel version:
[root@ovirt01 etc]# uname -a Linux ovirt01.hawai.lan 3.10.0-123.20.1.el7.x86_64 #1 SMP Thu Jan 29 18:05:33 UTC 2015 x86_64 x86_64 x86_64 GNU/Linux Can you attach vdsm logs? (/var/log/vdsm/vdsm.log*)
Nir