[ovirt-users] can not use iscsi storage type on ovirtandGlusterfshyper-converged environment
胡茂荣
maorong.hu at horebdata.cn
Fri Nov 25 09:44:03 UTC 2016
===================================---
###vdsm or supervdsm log report :
MainProcess|jsonrpc.Executor/7::ERROR::2016-11-01 11:07:00,178::supervdsmServer::96::SuperVdsm.ServerCallback::(wrapper) Error in getPathsStatus
MainProcess|jsonrpc.Executor/4::ERROR::2016-11-01 11:07:20,964::supervdsmServer::96::SuperVdsm.ServerCallback::(wrapper) Error in getPathsStatus
============================ some code info ------------>
[root at horeba storage]# pwd
/usr/share/vdsm/storage
[root at horeba storage]# grep "getPathsStatus" -R ./
./devicemapper.py:def _getPathsStatus():
./devicemapper.py:def getPathsStatus():
./devicemapper.py: return getProxy().getPathsStatus()
./multipath.py: pathStatuses = devicemapper.getPathsStatus()
def _getPathsStatus():
cmd = [EXT_DMSETUP, "status"]
rc, out, err = misc.execCmd(cmd)
if rc != 0:
raise Exception("Could not get device statuses")
res = {}
for statusLine in out:
try:
devName, statusLine = statusLine.split(":", 1)
except ValueError:
if len(out) == 1:
# return an empty dict when status output is: No devices found
return res
else:
raise
for m in PATH_STATUS_RE.finditer(statusLine):
devNum, status = m.groups()
physdevName = findDev(*[int(i) for i in devNum.split(":")])
res[physdevName] = {"A": "active", "F": "failed"}[status]
return res
def getPathsStatus():
return getProxy().getPathsStatus()
=============================================
and flashcache dm device will error when use getPathsStatus() function . could change code not check flashcache dm device ?
========================================dmsetup info ----------->
[root at horebc ~]# dmsetup status
flash_sdb: 0 976771072 flashcache stats:
reads(1388761), writes(15548965)
read hits(1235671), read hit percent(88)
write hits(6539144) write hit percent(42)
dirty write hits(21372) dirty write hit percent(0)
replacement(147711), write replacement(524881)
write invalidates(0), read invalidates(1)
pending enqueues(810), pending inval(810)
metadata dirties(15196370), metadata cleans(15196322)
metadata batch(30087377) metadata ssd writes(305315)
cleanings(15196322) fallow cleanings(48187)
no room(337139) front merge(716153) back merge(14391395)
force_clean_block(0)
disk reads(153093), disk writes(15530535) ssd reads(16431974) ssd writes(15672221)
uncached reads(3714), uncached writes(334235), uncached IO requeue(0)
disk read errors(0), disk write errors(0) ssd read errors(0) ssd write errors(0)
uncached sequential reads(0), uncached sequential writes(0)
pid_adds(0), pid_dels(0), pid_drops(0) pid_expiry(0)
lru hot blocks(12158976), lru warm blocks(12158976)
lru promotions(0), lru demotions(0)
dedbd337--ca66--43ff--b78c--4e9347682a9c-metadata: 0 1048576 linear
dedbd337--ca66--43ff--b78c--4e9347682a9c-leases: 0 4194304 linear
23137643634356633: 0 2147483648 multipath 2 0 0 0 1 1 A 0 1 2 8:128 A 0 0 1
dedbd337--ca66--43ff--b78c--4e9347682a9c-master: 0 2097152 linear
dedbd337--ca66--43ff--b78c--4e9347682a9c-inbox: 0 262144 linear
dedbd337--ca66--43ff--b78c--4e9347682a9c-outbox: 0 262144 linear
[root at horebc ~]# dmsetup info -C
Name Maj Min Stat Open Targ Event UUID
flash_sdb 253 0 L--w 1 1 0
dedbd337--ca66--43ff--b78c--4e9347682a9c-metadata 253 4 L--w 0 1 0 LVM-zSZKhQXbs2Ms5ceM2Wlt3p46BgsS8gGIGwifAPyYj9GmjFzCmJkIf9vFFFHn9n7V
dedbd337--ca66--43ff--b78c--4e9347682a9c-leases 253 6 L--w 0 1 0 LVM-zSZKhQXbs2Ms5ceM2Wlt3p46BgsS8gGIVSCllWEYYKziY1bSeiTL0dAKAd27JqDT
23137643634356633 253 3 L--w 6 1 0 mpath-23137643634356633
dedbd337--ca66--43ff--b78c--4e9347682a9c-master 253 9 L--w 0 1 0 LVM-zSZKhQXbs2Ms5ceM2Wlt3p46BgsS8gGIiEnFZklRhZfFZ4YRdYWFImKWsUGr5pHg
dedbd337--ca66--43ff--b78c--4e9347682a9c-inbox 253 8 L--w 0 1 0 LVM-zSZKhQXbs2Ms5ceM2Wlt3p46BgsS8gGILobNK8KRD4SzDWyg50aG7jGdcNAi3KNw
dedbd337--ca66--43ff--b78c--4e9347682a9c-outbox 253 5 L--w 0 1 0 LVM-zSZKhQXbs2Ms5ceM2Wlt3p46BgsS8gGIAvbT8CLegbVL802bG3QgLtH7I7llmS6R
flash_sdf 253 2 L--w 1 1 0
dedbd337--ca66--43ff--b78c--4e9347682a9c-ids 253 7 L--w 1 1 0 LVM-zSZKhQXbs2Ms5ceM2Wlt3p46BgsS8gGIIkeaFaWvxa9wbHy7WrfiXNsP4F2J3gg0
flash_sdd 253 1 L--w 1 1 0
====================================================================
and flashcache dm have no UUID ,I think it can export it before checking .
humaorong
2016-11-25
------------------ Original ------------------
From: "胡茂荣"<maorong.hu at horebdata.cn>;
Date: Fri, Nov 25, 2016 01:18 PM
To: "Sahina Bose"<sabose at redhat.com>;
Cc: "Maor Lipchuk"<mlipchuk at redhat.com>; "Jeff Nelson"<jenelson at redhat.com>; "users"<users at ovirt.org>;
Subject: Re: [ovirt-users] can not use iscsi storage type on ovirtandGlusterfshyper-converged environment
I find this problem more info :
I use flashcache on ovirt hosts , have dm device in /dev/mapper/ :
[root at horeba init.d]# dmsetup info -C (and I set them in multipath blacklist )
Name Maj Min Stat Open Targ Event UUID
flash_sdb 253 0 L--w 0 1 0
flash_sdf 253 2 L--w 0 1 0
flash_sdd 253 1 L--w 0 1 0
[root at horeba init.d]# multipath -l
[root at horeba init.d]#
[root at horeba init.d]# ll /dev/mapper/
total 0
crw------- 1 root root 10, 236 Nov 25 10:09 control
lrwxrwxrwx 1 root root 7 Nov 25 12:51 flash_sdb -> ../dm-0
lrwxrwxrwx 1 root root 7 Nov 25 12:51 flash_sdd -> ../dm-1
lrwxrwxrwx 1 root root 7 Nov 25 12:51 flash_sdf -> ../dm-2
on this condition , ovirt UI add iscsi type storage will fail .
If I delete the flashcache device /dev/mapper/flash_* , ovirt UI add iscsi type storage have no problem .
I need flashcache for using ssd cache on my environment , how can I use iscsi type storage on this environment , please help me , Thanks!
------------------ Original ------------------
From: "Sahina Bose"<sabose at redhat.com>;
Date: Thu, Nov 3, 2016 05:54 PM
To: "胡茂荣"<maorong.hu at horebdata.cn>;
Cc: "Maor Lipchuk"<mlipchuk at redhat.com>; "Jeff Nelson"<jenelson at redhat.com>; "users"<users at ovirt.org>;
Subject: Re: [ovirt-users] can not use iscsi storage type on ovirtandGlusterfshyper-converged environment
A wild guess, not sure if it is related - can you check if multipathd service is enabled. If you set up your oVirt-Gluster hyperconverged environment via gdeploy, multipathd service is disabled and the /etc/multipath.conf is edited to blacklist all devices - this was to fix https://bugzilla.redhat.com/show_bug.cgi?id=1303940
Since you mention you're unable to add iscsi storage only in this environment, thought it's worth checking.
On Thu, Nov 3, 2016 at 6:40 AM, 胡茂荣 <maorong.hu at horebdata.cn> wrote:
my environment rpm are :
[root at horeba ~]# rpm -q vdsm
vdsm-4.18.13-1.el7.centos.x86_64
[root at horeba ~]# rpm -aq | grep ovirt
ovirt-vmconsole-host-1.0.4-1.el7.centos.noarch
ovirt-imageio-common-0.4.0-1.el7.noarch
ovirt-hosted-engine-setup-2.0.2.2-1.el7.centos.noarch
ovirt-imageio-daemon-0.4.0-1.el7.noarch
ovirt-engine-appliance-4.0-20160928.1.el7.centos.noarch
ovirt-vmconsole-1.0.4-1.el7.centos.noarch
ovirt-host-deploy-1.5.2-1.el7.centos.noarch
ovirt-hosted-engine-ha-2.0.4-1.el7.centos.noarch
ovirt-release40-4.0.4-1.noarch
ovirt-setup-lib-1.0.2-1.el7.centos.noarch
ovirt-engine-sdk-python-3.6.9.1-1.el7.centos.noarch
and I test if not on ' ovirt and Glusterfs hyper-converged environment ' , ovirt WEB UI add iscsi storage work ok .
------------------ Original ------------------
From: "Maor Lipchuk"<mlipchuk at redhat.com>;
Date: Wed, Nov 2, 2016 07:37 PM
To: "胡茂荣"<maorong.hu at horebdata.cn>;
Cc: "users"<users at ovirt.org>; "Jeff Nelson"<jenelson at redhat.com>; "Nir Soffer"<nsoffer at redhat.com>;
Subject: Re: [ovirt-users] can not use iscsi storage type on ovirtandGlusterfs hyper-converged environment
Thanks for the logs,
What kind of VDSM version are you using?
"rpm -q vdsm"
There seems to be a similar issue which was reported recently in the VDSM area
(see https://bugzilla.redhat.com/show_bug.cgi?id=1197292)
It should be fixed in later versions of VDSM vdsm-4.16.12-2.el7ev.x86_64
Adding also Nir and Jeff to the thread, if they have any insights
Regards,
Maor
On Wed, Nov 2, 2016 at 4:11 AM, 胡茂荣 <maorong.hu at horebdata.cn> wrote:
Hi Maor:
vdsm/supervdsm/engine log on attachment . I mkfs.xfs the lun block device and mount to /mnt , dd write ,dmesg not report error ,dd result is ok :
/dev/sdi 50G 33M 50G 1% /mnt
[root at horebc mnt]# for i in `seq 3`; do dd if=/dev/zero of=./file bs=1G count=1 oflag=direct ; done
1+0 records in
1+0 records out
1073741824 bytes (1.1 GB) copied, 13.3232 s, 80.6 MB/s
1+0 records in
1+0 records out
1073741824 bytes (1.1 GB) copied, 9.89988 s, 108 MB/s
1+0 records in
1+0 records out
1073741824 bytes (1.1 GB) copied, 14.0143 s, 76.6 MB/s
my envirnment have three network segments (hosts have 3 network segments ) :
engine and glusterfs mount : 192.168.11.X/24
glusterfs brick : 192.168.10.x/24
iscsi : 192.168.1.0/24
and I add 192.168.1.0/24 to engine vm , ovirt web UI report the same error .
humaorong
2016-11-2
------------------ Original ------------------
From: "Maor Lipchuk"<mlipchuk at redhat.com>;
Date: Tue, Nov 1, 2016 08:14 PM
To: "胡茂荣"<maorong.hu at horebdata.cn>;
Cc: "users"<users at ovirt.org>;
Subject: Re: [ovirt-users] can not use iscsi storage type on ovirt andGlusterfs hyper-converged environment
Hi 胡茂荣
Can u please also add the VDSM and engine logs.
If you try discover and connect to those luns directly from your Host does it work?
Regards,
Maor
On Tue, Nov 1, 2016 at 6:12 AM, 胡茂荣 <maorong.hu at horebdata.cn> wrote:
on ovirt and Glusterfs hyper-converged environment , can not use iscsi storage type , UI report error: "Could not retrieve LUNs, please check your storage." , vdsm report :"VDSM hosted_engine_3 command failed: Error block device action: ()" .
but this block device alse login on centos 7 host :
=============================================================
## lsscsi
[7:0:0:0] disk SCST_BIO DEVFOR_OVIRT_rbd 221 /dev/sdi
## dmesg :
[684521.131186] sd 7:0:0:0: [sdi] Attached SCSI disk
===================================---
###vdsm or supervdsm log report :
MainProcess|jsonrpc.Executor/7::ERROR::2016-11-01 11:07:00,178::supervdsmServer::96::SuperVdsm.ServerCallback::(wrapper) Error in getPathsStatus
MainProcess|jsonrpc.Executor/4::ERROR::2016-11-01 11:07:20,964::supervdsmServer::96::SuperVdsm.ServerCallback::(wrapper) Error in getPathsStatus
jsonrpc.Executor/4::DEBUG::2016-11-01 11:07:04,251::iscsi::434::Storage.ISCSI::(rescan) Performing SCSI scan, this will take up to 30 seconds
jsonrpc.Executor/5::INFO::2016-11-01 11:07:19,413::iscsi::567::Storage.ISCSI::(setRpFilterIfNeeded) iSCSI iface.net_ifacename not provided. Skipping.
11:09:15,753::iscsiadm::119::Storage.Misc.excCmd::(_runCmd) /usr/bin/taskset --cpu-list 0-7 /usr/bin/sudo -n /usr/sbin/iscsiadm -m session -R (cwd None)
======================================
the other info please the attachment "bug-info.doc".
this prolem on ovirt3.6 and 4.X ovirt and Glusterfs hyper-converged environment . how can I use iscsi storage type on ovirt and Glusterfs hyper-converged environment .Please help me !
humaorong
2016-11-1
_______________________________________________
Users mailing list
Users at ovirt.org
http://lists.ovirt.org/mailman/listinfo/users
_______________________________________________
Users mailing list
Users at ovirt.org
http://lists.ovirt.org/mailman/listinfo/users
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.ovirt.org/pipermail/users/attachments/20161125/d65c6a1a/attachment-0001.html>
More information about the Users
mailing list