[ovirt-users] can not use iscsi storage type onovirtandGlusterfshyper-converged environment
Fred Rolland
frolland at redhat.com
Mon Dec 5 08:44:56 UTC 2016
Hi,
After changing the code, you need to restart the VDSM service to take
effect.
Regards,
Fred
On Mon, Dec 5, 2016 at 10:28 AM, 胡茂荣 <maorong.hu at horebdata.cn> wrote:
>
> Thanks for Yaniv Kaul , change code need build vdsm source code , and
> if only change /usr/share/vdsm/storage/devicemapper.py will not really
> take effect .
>
> could this problem as a bug , and correct it to ovirt vdsm source code
> ?
>
> ------------------ Original ------------------
> *From: * "Yaniv Kaul"<ykaul at redhat.com>;
> *Date: * Sun, Dec 4, 2016 07:07 PM
> *To: * "胡茂荣"<maorong.hu at horebdata.cn>;
> *Cc: * "胡晓宇"<samuel.xhu at horebdata.cn>; "users"<users at ovirt.org>; "Sahina
> Bose"<sabose at redhat.com>; "Jeff Nelson"<jenelson at redhat.com>;
> *Subject: * Re: [ovirt-users] can not use iscsi storage type
> onovirtandGlusterfshyper-converged environment
>
>
>
> On Dec 2, 2016 11:53 AM, "胡茂荣" <maorong.hu at horebdata.cn> wrote:
>
> I find supervdsm used " /usr/sbin/dmsetup status" :
>
> MainProcess|jsonrpc.Executor/2::DEBUG::2016-12-02
> 17:12:16,372::supervdsmServer::92::SuperVdsm.ServerCallback::(wrapper)
> call getPathsStatus with () {}
> MainProcess|jsonrpc.Executor/2::DEBUG::2016-12-02
> 17:12:16,373::devicemapper::154::Storage.Misc.excCmd::(_getPathsStatus)
> /usr/bin/taskset --cpu-list 0-7 /usr/sbin/dmsetup status (cwd None)
> MainProcess|jsonrpc.Executor/2::DEBUG::2016-12-02
> 17:12:16,377::devicemapper::154::Storage.Misc.excCmd::(_getPathsStatus)
> SUCCESS: <err> = ''; <rc> = 0
> MainProcess|jsonrpc.Executor/2::ERROR::2016-12-02
> 17:12:16,378::supervdsmServer::96::SuperVdsm.ServerCallback::(wrapper) Error
> in getPathsStatus
>
> problem :
> how can I change Storage.Misc.excCmd _getPathsStatus " /usr/bin/taskset
> --cpu-list 0-7 /usr/sbin/dmsetup status" to :
>
> /usr/bin/taskset --cpu-list 0-7 /usr/sbin/dmsetup status --target
> multipah
>
> I think add iscsi type storage , if supervdsm scan mulitpah will solve
> my problem .(my environment have other dm devices, use "dmsetup status"
> will show them, and vdsm get dm path status will occur error )
>
> ============================================================
> ====================
> so I changed some as follow :
> (1)
> I define EXT_DMSETUP_STATUS in /usr/lib/python2.7/site-packages/vdsm/constants.py
> :
>
> /usr/lib/python2.7/site-packages/vdsm/constants.py:EXT_DMSETUP =
> '/usr/sbin/dmsetup'
> /usr/lib/python2.7/site-packages/vdsm/constants.py:EXT_DMSETUP_STATUS =
> "/usr/sbin/dmsetup status --target multipath"
>
> (2)
> /usr/share/vdsm/storage/devicemapper.py add :
> from vdsm.constants import EXT_DMSETUP_STATUS
>
> and changed getPathsStatus cmd to " EXT_DMSETUP_STATUS" :
>
> def _getPathsStatus():
> cmd = [EXT_DMSETUP_STATUS] ##### before : cmd=[
> EXT_DMSETUP,"status"]
>
>
> Why not change this to:
> cmd = [EXT_DMSETUP, "status", "--target", "multipath"]
>
> Y.
>
> rc, out, err = misc.execCmd(cmd)
> ============================================================
> ===========================
>
> but log in supervdsm log also not change . Please help me ,how to change
> code to let supervdsm exec "/usr/sbin/dmsetup status --target multipath"
> in function getPathsStatus() 。
>
>
>
>
> ------------------ Original ------------------
> *From: * "胡茂荣"<maorong.hu at horebdata.cn>;
> *Date: * Fri, Nov 25, 2016 05:44 PM
> *To: * "Sahina Bose"<sabose at redhat.com>;
> *Cc: * "Maor Lipchuk"<mlipchuk at redhat.com>; "Jeff Nelson"<
> jenelson at redhat.com>; "users"<users at ovirt.org>;
> *Subject: * Re: [ovirt-users] can not use iscsi storage type on
> ovirtandGlusterfshyper-converged environment
>
>
> ===================================---
>
> ###vdsm or supervdsm log report :
>
> MainProcess|jsonrpc.Executor/7::ERROR::2016-11-01
> 11:07:00,178::supervdsmServer::96::SuperVdsm.ServerCallback::(wrapper)
> Error in getPathsStatus
>
> MainProcess|jsonrpc.Executor/4::ERROR::2016-11-01
> 11:07:20,964::supervdsmServer::96::SuperVdsm.ServerCallback::(wrapper)
> Error in getPathsStatus
> ============================ some code info ------------>
> [root at horeba storage]# pwd
> /usr/share/vdsm/storage
>
> [root at horeba storage]# grep "getPathsStatus" -R ./
> ./devicemapper.py:def _getPathsStatus():
> ./devicemapper.py:def getPathsStatus():
> ./devicemapper.py: return getProxy().getPathsStatus()
> ./multipath.py: pathStatuses = devicemapper.getPathsStatus()
>
> def _getPathsStatus():
> cmd = [EXT_DMSETUP, "status"]
> rc, out, err = misc.execCmd(cmd)
> if rc != 0:
> raise Exception("Could not get device statuses")
>
> res = {}
> for statusLine in out:
> try:
> devName, statusLine = statusLine.split(":", 1)
> except ValueError:
> if len(out) == 1:
> # return an empty dict when status output is: No devices
> found
> return res
> else:
> raise
>
> for m in PATH_STATUS_RE.finditer(statusLine):
> devNum, status = m.groups()
> physdevName = findDev(*[int(i) for i in devNum.split(":")])
> res[physdevName] = {"A": "active", "F": "failed"}[status]
>
> return res
> def getPathsStatus():
> return getProxy().getPathsStatus()
> =============================================
> and flashcache dm device will error when use getPathsStatus() function
> . could change code not check flashcache dm device ?
> ========================================dmsetup info ----------->
> [root at horebc ~]# dmsetup status
> flash_sdb: 0 976771072 flashcache stats:
> reads(1388761), writes(15548965)
> read hits(1235671), read hit percent(88)
> write hits(6539144) write hit percent(42)
> dirty write hits(21372) dirty write hit percent(0)
> replacement(147711), write replacement(524881)
> write invalidates(0), read invalidates(1)
> pending enqueues(810), pending inval(810)
> metadata dirties(15196370), metadata cleans(15196322)
> metadata batch(30087377) metadata ssd writes(305315)
> cleanings(15196322) fallow cleanings(48187)
> no room(337139) front merge(716153) back merge(14391395)
> force_clean_block(0)
> disk reads(153093), disk writes(15530535) ssd reads(16431974) ssd
> writes(15672221)
> uncached reads(3714), uncached writes(334235), uncached IO
> requeue(0)
> disk read errors(0), disk write errors(0) ssd read errors(0) ssd
> write errors(0)
> uncached sequential reads(0), uncached sequential writes(0)
> pid_adds(0), pid_dels(0), pid_drops(0) pid_expiry(0)
> lru hot blocks(12158976), lru warm blocks(12158976)
> lru promotions(0), lru demotions(0)
> dedbd337--ca66--43ff--b78c--4e9347682a9c-metadata: 0 1048576 linear
> dedbd337--ca66--43ff--b78c--4e9347682a9c-leases: 0 4194304 linear
> 23137643634356633: 0 2147483648 multipath 2 0 0 0 1 1 A 0 1 2 8:128 A 0 0
> 1
> dedbd337--ca66--43ff--b78c--4e9347682a9c-master: 0 2097152 linear
> dedbd337--ca66--43ff--b78c--4e9347682a9c-inbox: 0 262144 linear
> dedbd337--ca66--43ff--b78c--4e9347682a9c-outbox: 0 262144 linear
>
> [root at horebc ~]# dmsetup info -C
> Name Maj Min Stat Open Targ
> Event UUID
> flash_sdb 253 0 L--w 1 1
> 0
> dedbd337--ca66--43ff--b78c--4e9347682a9c-metadata 253 4 L--w 0 1
> 0 LVM-zSZKhQXbs2Ms5ceM2Wlt3p46BgsS8gGIGwifAPyYj9GmjFzCmJkIf9vF
> FFHn9n7V
> dedbd337--ca66--43ff--b78c--4e9347682a9c-leases 253 6 L--w 0 1
> 0 LVM-zSZKhQXbs2Ms5ceM2Wlt3p46BgsS8gGIVSCllWEYYKziY1bSeiTL0dAK
> Ad27JqDT
> 23137643634356633 253 3 L--w 6 1
> 0 mpath-23137643634356633
> dedbd337--ca66--43ff--b78c--4e9347682a9c-master 253 9 L--w 0 1
> 0 LVM-zSZKhQXbs2Ms5ceM2Wlt3p46BgsS8gGIiEnFZklRhZfFZ4YRdYWFImKW
> sUGr5pHg
> dedbd337--ca66--43ff--b78c--4e9347682a9c-inbox 253 8 L--w 0 1
> 0 LVM-zSZKhQXbs2Ms5ceM2Wlt3p46BgsS8gGILobNK8KRD4SzDWyg50aG7jGd
> cNAi3KNw
> dedbd337--ca66--43ff--b78c--4e9347682a9c-outbox 253 5 L--w 0 1
> 0 LVM-zSZKhQXbs2Ms5ceM2Wlt3p46BgsS8gGIAvbT8CLegbVL802bG3QgLtH7
> I7llmS6R
> flash_sdf 253 2 L--w 1 1
> 0
> dedbd337--ca66--43ff--b78c--4e9347682a9c-ids 253 7 L--w 1 1
> 0 LVM-zSZKhQXbs2Ms5ceM2Wlt3p46BgsS8gGIIkeaFaWvxa9wbHy7WrfiXNsP
> 4F2J3gg0
> flash_sdd 253 1 L--w 1 1
> 0
>
> ====================================================================
> and flashcache dm have no UUID ,I think it can export it before checking
> .
>
>
> humaorong
> 2016-11-25
>
>
>
> ------------------ Original ------------------
> *From: * "胡茂荣"<maorong.hu at horebdata.cn>;
> *Date: * Fri, Nov 25, 2016 01:18 PM
> *To: * "Sahina Bose"<sabose at redhat.com>;
> *Cc: * "Maor Lipchuk"<mlipchuk at redhat.com>; "Jeff Nelson"<
> jenelson at redhat.com>; "users"<users at ovirt.org>;
> *Subject: * Re: [ovirt-users] can not use iscsi storage type on
> ovirtandGlusterfshyper-converged environment
>
>
> I find this problem more info :
> I use flashcache on ovirt hosts , have dm device in /dev/mapper/ :
>
> [root at horeba init.d]# dmsetup info -C (and I set them in multipath
> blacklist )
> Name Maj Min Stat Open Targ Event UUID
>
> flash_sdb 253 0 L--w 0 1 0
>
> flash_sdf 253 2 L--w 0 1 0
>
> flash_sdd 253 1 L--w 0 1 0
> [root at horeba init.d]# multipath -l
> [root at horeba init.d]#
>
> [root at horeba init.d]# ll /dev/mapper/
> total 0
> crw------- 1 root root 10, 236 Nov 25 10:09 control
> lrwxrwxrwx 1 root root 7 Nov 25 12:51 flash_sdb -> ../dm-0
> lrwxrwxrwx 1 root root 7 Nov 25 12:51 flash_sdd -> ../dm-1
> lrwxrwxrwx 1 root root 7 Nov 25 12:51 flash_sdf -> ../dm-2
>
> on this condition , ovirt UI add iscsi type storage will fail .
>
> If I delete the flashcache device /dev/mapper/flash_* , ovirt UI add
> iscsi type storage have no problem .
>
> I need flashcache for using ssd cache on my environment , how can I
> use iscsi type storage on this environment , please help me , Thanks!
>
>
> ------------------ Original ------------------
> *From: * "Sahina Bose"<sabose at redhat.com>;
> *Date: * Thu, Nov 3, 2016 05:54 PM
> *To: * "胡茂荣"<maorong.hu at horebdata.cn>;
> *Cc: * "Maor Lipchuk"<mlipchuk at redhat.com>; "Jeff Nelson"<
> jenelson at redhat.com>; "users"<users at ovirt.org>;
> *Subject: * Re: [ovirt-users] can not use iscsi storage type on
> ovirtandGlusterfshyper-converged environment
>
> A wild guess, not sure if it is related - can you check if multipathd
> service is enabled. If you set up your oVirt-Gluster hyperconverged
> environment via gdeploy, multipathd service is disabled and the
> /etc/multipath.conf is edited to blacklist all devices - this was to fix
> https://bugzilla.redhat.com/show_bug.cgi?id=1303940
>
> Since you mention you're unable to add iscsi storage only in this
> environment, thought it's worth checking.
>
> On Thu, Nov 3, 2016 at 6:40 AM, 胡茂荣 <maorong.hu at horebdata.cn> wrote:
>
>>
>> my environment rpm are :
>> [root at horeba ~]# rpm -q vdsm
>> vdsm-4.18.13-1.el7.centos.x86_64
>>
>> [root at horeba ~]# rpm -aq | grep ovirt
>> ovirt-vmconsole-host-1.0.4-1.el7.centos.noarch
>> ovirt-imageio-common-0.4.0-1.el7.noarch
>> ovirt-hosted-engine-setup-2.0.2.2-1.el7.centos.noarch
>> ovirt-imageio-daemon-0.4.0-1.el7.noarch
>> ovirt-engine-appliance-4.0-20160928.1.el7.centos.noarch
>> ovirt-vmconsole-1.0.4-1.el7.centos.noarch
>> ovirt-host-deploy-1.5.2-1.el7.centos.noarch
>> ovirt-hosted-engine-ha-2.0.4-1.el7.centos.noarch
>> ovirt-release40-4.0.4-1.noarch
>> ovirt-setup-lib-1.0.2-1.el7.centos.noarch
>> ovirt-engine-sdk-python-3.6.9.1-1.el7.centos.noarch
>>
>> and I test if not on ' ovirt and Glusterfs hyper-converged environment '
>> , ovirt WEB UI add iscsi storage work ok .
>>
>>
>> ------------------ Original ------------------
>> *From: * "Maor Lipchuk"<mlipchuk at redhat.com>;
>> *Date: * Wed, Nov 2, 2016 07:37 PM
>> *To: * "胡茂荣"<maorong.hu at horebdata.cn>;
>> *Cc: * "users"<users at ovirt.org>; "Jeff Nelson"<jenelson at redhat.com>;
>> "Nir Soffer"<nsoffer at redhat.com>;
>> *Subject: * Re: [ovirt-users] can not use iscsi storage type on
>> ovirtandGlusterfs hyper-converged environment
>>
>> Thanks for the logs,
>>
>> What kind of VDSM version are you using?
>> "rpm -q vdsm"
>> There seems to be a similar issue which was reported recently in the VDSM
>> area
>> (see https://bugzilla.redhat.com/show_bug.cgi?id=1197292)
>> It should be fixed in later versions of VDSM vdsm-4.16.12-2.el7ev.x86_64
>> Adding also Nir and Jeff to the thread, if they have any insights
>>
>> Regards,
>> Maor
>>
>> On Wed, Nov 2, 2016 at 4:11 AM, 胡茂荣 <maorong.hu at horebdata.cn> wrote:
>>
>>>
>>> Hi Maor:
>>> vdsm/supervdsm/engine log on attachment . I mkfs.xfs the lun
>>> block device and mount to /mnt , dd write ,dmesg not report error ,dd
>>> result is ok :
>>>
>>> /dev/sdi 50G 33M 50G 1% /mnt
>>>
>>> [root at horebc mnt]# for i in `seq 3`; do dd if=/dev/zero of=./file
>>> bs=1G count=1 oflag=direct ; done
>>> 1+0 records in
>>> 1+0 records out
>>> 1073741824 bytes (1.1 GB) copied, 13.3232 s, 80.6 MB/s
>>> 1+0 records in
>>> 1+0 records out
>>> 1073741824 bytes (1.1 GB) copied, 9.89988 s, 108 MB/s
>>> 1+0 records in
>>> 1+0 records out
>>> 1073741824 bytes (1.1 GB) copied, 14.0143 s, 76.6 MB/s
>>>
>>> my envirnment have three network segments (hosts have 3 network
>>> segments ) :
>>> engine and glusterfs mount : 192.168.11.X/24
>>> glusterfs brick : 192.168.10.x/24
>>> iscsi : 192.168.1.0/24
>>>
>>> and I add 192.168.1.0/24 to engine vm , ovirt web UI report the
>>> same error .
>>>
>>> humaorong
>>> 2016-11-2
>>>
>>> ------------------ Original ------------------
>>> *From: * "Maor Lipchuk"<mlipchuk at redhat.com>;
>>> *Date: * Tue, Nov 1, 2016 08:14 PM
>>> *To: * "胡茂荣"<maorong.hu at horebdata.cn>;
>>> *Cc: * "users"<users at ovirt.org>;
>>> *Subject: * Re: [ovirt-users] can not use iscsi storage type on ovirt
>>> andGlusterfs hyper-converged environment
>>>
>>> Hi 胡茂荣Can u please also add the VDSM and engine logs.
>>> If you try discover and connect to those luns directly from your Host
>>> does it work?
>>>
>>> Regards,
>>> Maor
>>>
>>>
>>> On Tue, Nov 1, 2016 at 6:12 AM, 胡茂荣 <maorong.hu at horebdata.cn> wrote:
>>>
>>>>
>>>>
>>>> on ovirt and Glusterfs hyper-converged environment , can not use
>>>> iscsi storage type , UI report error: "Could not retrieve LUNs, please
>>>> check your storage." , vdsm report :"VDSM hosted_engine_3 command
>>>> failed: Error block device action: ()" .
>>>> but this block device alse login on centos 7 host :
>>>> =============================================================
>>>>
>>>> ## lsscsi
>>>>
>>>> [7:0:0:0] disk SCST_BIO DEVFOR_OVIRT_rbd 221 /dev/sdi
>>>>
>>>> ## dmesg :
>>>>
>>>> [684521.131186] sd 7:0:0:0: [sdi] Attached SCSI disk
>>>>
>>>> ===================================---
>>>>
>>>> ###vdsm or supervdsm log report :
>>>>
>>>> MainProcess|jsonrpc.Executor/7::ERROR::2016-11-01
>>>> 11:07:00,178::supervdsmServer::96::SuperVdsm.ServerCallback::(wrapper)
>>>> Error in getPathsStatus
>>>>
>>>> MainProcess|jsonrpc.Executor/4::ERROR::2016-11-01
>>>> 11:07:20,964::supervdsmServer::96::SuperVdsm.ServerCallback::(wrapper)
>>>> Error in getPathsStatus
>>>>
>>>> jsonrpc.Executor/4::DEBUG::2016-11-01 11:07:04,251::iscsi::434::Storage.ISCSI::(rescan)
>>>> Performing SCSI scan, this will take up to 30 seconds
>>>>
>>>> jsonrpc.Executor/5::INFO::2016-11-01 11:07:19,413::iscsi::567::Stor
>>>> age.ISCSI::(setRpFilterIfNeeded) iSCSI iface.net_ifacename not
>>>> provided. Skipping.
>>>>
>>>> 11:09:15,753::iscsiadm::119::Storage.Misc.excCmd::(_runCmd)
>>>> /usr/bin/taskset --cpu-list 0-7 /usr/bin/sudo -n /usr/sbin/iscsiadm -m
>>>> session -R (cwd None)
>>>>
>>>> ======================================
>>>>
>>>> the other info please the attachment "bug-info.doc".
>>>>
>>>> this prolem on ovirt3.6 and 4.X ovirt and Glusterfs
>>>> hyper-converged environment . how can I use iscsi storage type on ovirt
>>>> and Glusterfs hyper-converged environment .Please help me !
>>>>
>>>> humaorong
>>>>
>>>> 2016-11-1
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>> _______________________________________________
>>>> Users mailing list
>>>> Users at ovirt.org
>>>> http://lists.ovirt.org/mailman/listinfo/users
>>>>
>>>>
>>>
>>
>> _______________________________________________
>> Users mailing list
>> Users at ovirt.org
>> http://lists.ovirt.org/mailman/listinfo/users
>>
>>
>
> _______________________________________________
> Users mailing list
> Users at ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
>
>
> _______________________________________________
> Users mailing list
> Users at ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.ovirt.org/pipermail/users/attachments/20161205/52eafd68/attachment-0001.html>
More information about the Users
mailing list