[ovirt-users] can not use iscsi storage type onovirtandGlusterfshyper-convergedenvironment
Nir Soffer
nsoffer at redhat.com
Tue Dec 6 02:16:13 UTC 2016
On Tue, Dec 6, 2016 at 3:15 AM, 胡茂荣 <maorong.hu at horebdata.cn> wrote:
> Hi Nir :
> before change code ,supervdsm report error log like those :
> =================================== ------>
>
> MainProcess|jsonrpc.Executor/2::DEBUG::2016-12-02
> 17:12:16,372::supervdsmServer::92::SuperVdsm.ServerCallback::(wrapper)
> call getPathsStatus with () {}
>
> MainProcess|jsonrpc.Executor/2::DEBUG::2016-12-02
> 17:12:16,373::devicemapper::154::Storage.Misc.excCmd::(_getPathsStatus)
> /usr/bin/taskset --cpu-list 0-7 /usr/sbin/dmsetup status (cwd None)
>
> MainProcess|jsonrpc.Executor/2::DEBUG::2016-12-02
> 17:12:16,377::devicemapper::154::Storage.Misc.excCmd::(_getPathsStatus)
> SUCCESS: <err> = ''; <rc> = 0
>
> MainProcess|jsonrpc.Executor/2::ERROR::2016-12-02
> 17:12:16,378::supervdsmServer::96::SuperVdsm.ServerCallback::(wrapper) Error
> in getPathsStatus
> ======================================= <----------------
> my environment have other dm device , dmsetup status will show like
> this as follow, so will occure above error !
> ==================================================== -------->
> flash_sdd: 0 976771072 flashcache stats:
> reads(5723182), writes(47879823)
> read hits(2938125), read hit percent(51)
> write hits(12617126) write hit percent(26)
> dirty write hits(4592227) dirty write hit percent(9)
> replacement(701168), write replacement(3189334)
> write invalidates(0), read invalidates(1)
> ----
> ==================================================== <--------------
> use your patch to test , add or import iscsi type storage domain
> have no problem , log like follow :
> ====================================================== --------->
>
> MainProcess|jsonrpc.Executor/1::DEBUG::2016-12-06
> 08:58:07,439::supervdsmServer::92::SuperVdsm.ServerCallback::(wrapper)
> call getPathsStatus with () {}
>
> MainProcess|jsonrpc.Executor/1::DEBUG::2016-12-06
> 08:58:07,439::devicemapper::155::Storage.Misc.excCmd::(_getPathsStatus)
> /usr/bin/taskset --cpu-list 0-7 /usr/sbin/dmsetup status --target
> multipath (cwd None)
> MainProcess|jsonrpc.Executor/1::DEBUG::2016-12-06
> 08:58:07,443::devicemapper::155::Storage.Misc.excCmd::(_getPathsStatus)
> SUCCESS: <err> = ''; <rc> = 0
> ============================================================
> =====================
>
> Regards,
> humaorong
>
Thanks for testing this.
Sahina, can you test if this solve the issue with hyperconverge setup?
Maybe we can enable multipath with this patch.
Cheers,
Nir
>
>
> ------------------ Original ------------------
> *From: * "Nir Soffer"<nsoffer at redhat.com>;
> *Date: * Tue, Dec 6, 2016 01:41 AM
> *To: * "胡茂荣"<maorong.hu at horebdata.cn>;
> *Cc: * "users"<users at ovirt.org>; "Jeff Nelson"<jenelson at redhat.com>;
> "胡晓宇"<samuel.xhu at horebdata.cn>;
> *Subject: * Re: [ovirt-users] can not use iscsi storage type
> onovirtandGlusterfshyper-convergedenvironment
>
> Hi 胡茂荣,
>
> Can you test this patch?
> https://gerrit.ovirt.org/67844
>
> I also need more information on your setup, to add more
> details to the commit message.
>
> Thanks for reporting this,
> Nir
>
> On Mon, Dec 5, 2016 at 10:28 AM, 胡茂荣 <maorong.hu at horebdata.cn> wrote:
>
>>
>> Thanks for Yaniv Kaul , change code need build vdsm source code , and
>> if only change /usr/share/vdsm/storage/devicemapper.py will not really
>> take effect .
>>
>> could this problem as a bug , and correct it to ovirt vdsm source code
>> ?
>>
>> ------------------ Original ------------------
>> *From: * "Yaniv Kaul"<ykaul at redhat.com>;
>> *Date: * Sun, Dec 4, 2016 07:07 PM
>> *To: * "胡茂荣"<maorong.hu at horebdata.cn>;
>> *Cc: * "胡晓宇"<samuel.xhu at horebdata.cn>; "users"<users at ovirt.org>; "Sahina
>> Bose"<sabose at redhat.com>; "Jeff Nelson"<jenelson at redhat.com>;
>> *Subject: * Re: [ovirt-users] can not use iscsi storage type
>> onovirtandGlusterfshyper-converged environment
>>
>>
>>
>> On Dec 2, 2016 11:53 AM, "胡茂荣" <maorong.hu at horebdata.cn> wrote:
>>
>> I find supervdsm used " /usr/sbin/dmsetup status" :
>>
>> MainProcess|jsonrpc.Executor/2::DEBUG::2016-12-02
>> 17:12:16,372::supervdsmServer::92::SuperVdsm.ServerCallback::(wrapper)
>> call getPathsStatus with () {}
>> MainProcess|jsonrpc.Executor/2::DEBUG::2016-12-02
>> 17:12:16,373::devicemapper::154::Storage.Misc.excCmd::(_getPathsStatus)
>> /usr/bin/taskset --cpu-list 0-7 /usr/sbin/dmsetup status (cwd None)
>> MainProcess|jsonrpc.Executor/2::DEBUG::2016-12-02
>> 17:12:16,377::devicemapper::154::Storage.Misc.excCmd::(_getPathsStatus)
>> SUCCESS: <err> = ''; <rc> = 0
>> MainProcess|jsonrpc.Executor/2::ERROR::2016-12-02
>> 17:12:16,378::supervdsmServer::96::SuperVdsm.ServerCallback::(wrapper) Error
>> in getPathsStatus
>>
>> problem :
>> how can I change Storage.Misc.excCmd _getPathsStatus "
>> /usr/bin/taskset --cpu-list 0-7 /usr/sbin/dmsetup status" to :
>>
>> /usr/bin/taskset --cpu-list 0-7 /usr/sbin/dmsetup status --target
>> multipah
>>
>> I think add iscsi type storage , if supervdsm scan mulitpah will solve
>> my problem .(my environment have other dm devices, use "dmsetup status"
>> will show them, and vdsm get dm path status will occur error )
>>
>> ============================================================
>> ====================
>> so I changed some as follow :
>> (1)
>> I define EXT_DMSETUP_STATUS in /usr/lib/python2.7/site-packages/vdsm/constants.py
>> :
>>
>> /usr/lib/python2.7/site-packages/vdsm/constants.py:EXT_DMSETUP =
>> '/usr/sbin/dmsetup'
>> /usr/lib/python2.7/site-packages/vdsm/constants.py:EXT_DMSETUP_STATUS =
>> "/usr/sbin/dmsetup status --target multipath"
>>
>> (2)
>> /usr/share/vdsm/storage/devicemapper.py add :
>> from vdsm.constants import EXT_DMSETUP_STATUS
>>
>> and changed getPathsStatus cmd to " EXT_DMSETUP_STATUS" :
>>
>> def _getPathsStatus():
>> cmd = [EXT_DMSETUP_STATUS] ##### before : cmd=[
>> EXT_DMSETUP,"status"]
>>
>>
>> Why not change this to:
>> cmd = [EXT_DMSETUP, "status", "--target", "multipath"]
>>
>> Y.
>>
>> rc, out, err = misc.execCmd(cmd)
>> ============================================================
>> ===========================
>>
>> but log in supervdsm log also not change . Please help me ,how to change
>> code to let supervdsm exec "/usr/sbin/dmsetup status --target multipath"
>> in function getPathsStatus() 。
>>
>>
>>
>>
>> ------------------ Original ------------------
>> *From: * "胡茂荣"<maorong.hu at horebdata.cn>;
>> *Date: * Fri, Nov 25, 2016 05:44 PM
>> *To: * "Sahina Bose"<sabose at redhat.com>;
>> *Cc: * "Maor Lipchuk"<mlipchuk at redhat.com>; "Jeff Nelson"<
>> jenelson at redhat.com>; "users"<users at ovirt.org>;
>> *Subject: * Re: [ovirt-users] can not use iscsi storage type on
>> ovirtandGlusterfshyper-converged environment
>>
>>
>> ===================================---
>>
>> ###vdsm or supervdsm log report :
>>
>> MainProcess|jsonrpc.Executor/7::ERROR::2016-11-01
>> 11:07:00,178::supervdsmServer::96::SuperVdsm.ServerCallback::(wrapper)
>> Error in getPathsStatus
>>
>> MainProcess|jsonrpc.Executor/4::ERROR::2016-11-01
>> 11:07:20,964::supervdsmServer::96::SuperVdsm.ServerCallback::(wrapper)
>> Error in getPathsStatus
>> ============================ some code info ------------>
>> [root at horeba storage]# pwd
>> /usr/share/vdsm/storage
>>
>> [root at horeba storage]# grep "getPathsStatus" -R ./
>> ./devicemapper.py:def _getPathsStatus():
>> ./devicemapper.py:def getPathsStatus():
>> ./devicemapper.py: return getProxy().getPathsStatus()
>> ./multipath.py: pathStatuses = devicemapper.getPathsStatus()
>>
>> def _getPathsStatus():
>> cmd = [EXT_DMSETUP, "status"]
>> rc, out, err = misc.execCmd(cmd)
>> if rc != 0:
>> raise Exception("Could not get device statuses")
>>
>> res = {}
>> for statusLine in out:
>> try:
>> devName, statusLine = statusLine.split(":", 1)
>> except ValueError:
>> if len(out) == 1:
>> # return an empty dict when status output is: No devices
>> found
>> return res
>> else:
>> raise
>>
>> for m in PATH_STATUS_RE.finditer(statusLine):
>> devNum, status = m.groups()
>> physdevName = findDev(*[int(i) for i in devNum.split(":")])
>> res[physdevName] = {"A": "active", "F": "failed"}[status]
>>
>> return res
>> def getPathsStatus():
>> return getProxy().getPathsStatus()
>> =============================================
>> and flashcache dm device will error when use getPathsStatus()
>> function . could change code not check flashcache dm device ?
>> ========================================dmsetup info ----------->
>> [root at horebc ~]# dmsetup status
>> flash_sdb: 0 976771072 flashcache stats:
>> reads(1388761), writes(15548965)
>> read hits(1235671), read hit percent(88)
>> write hits(6539144) write hit percent(42)
>> dirty write hits(21372) dirty write hit percent(0)
>> replacement(147711), write replacement(524881)
>> write invalidates(0), read invalidates(1)
>> pending enqueues(810), pending inval(810)
>> metadata dirties(15196370), metadata cleans(15196322)
>> metadata batch(30087377) metadata ssd writes(305315)
>> cleanings(15196322) fallow cleanings(48187)
>> no room(337139) front merge(716153) back merge(14391395)
>> force_clean_block(0)
>> disk reads(153093), disk writes(15530535) ssd reads(16431974) ssd
>> writes(15672221)
>> uncached reads(3714), uncached writes(334235), uncached IO
>> requeue(0)
>> disk read errors(0), disk write errors(0) ssd read errors(0) ssd
>> write errors(0)
>> uncached sequential reads(0), uncached sequential writes(0)
>> pid_adds(0), pid_dels(0), pid_drops(0) pid_expiry(0)
>> lru hot blocks(12158976), lru warm blocks(12158976)
>> lru promotions(0), lru demotions(0)
>> dedbd337--ca66--43ff--b78c--4e9347682a9c-metadata: 0 1048576 linear
>> dedbd337--ca66--43ff--b78c--4e9347682a9c-leases: 0 4194304 linear
>> 23137643634356633: 0 2147483648 multipath 2 0 0 0 1 1 A 0 1 2 8:128 A 0
>> 0 1
>> dedbd337--ca66--43ff--b78c--4e9347682a9c-master: 0 2097152 linear
>> dedbd337--ca66--43ff--b78c--4e9347682a9c-inbox: 0 262144 linear
>> dedbd337--ca66--43ff--b78c--4e9347682a9c-outbox: 0 262144 linear
>>
>> [root at horebc ~]# dmsetup info -C
>> Name Maj Min Stat Open Targ
>> Event UUID
>> flash_sdb 253 0 L--w 1 1
>> 0
>> dedbd337--ca66--43ff--b78c--4e9347682a9c-metadata 253 4 L--w 0 1
>> 0 LVM-zSZKhQXbs2Ms5ceM2Wlt3p46BgsS8gGIGwifAPyYj9GmjFzCmJkIf9vF
>> FFHn9n7V
>> dedbd337--ca66--43ff--b78c--4e9347682a9c-leases 253 6 L--w 0 1
>> 0 LVM-zSZKhQXbs2Ms5ceM2Wlt3p46BgsS8gGIVSCllWEYYKziY1bSeiTL0dAK
>> Ad27JqDT
>> 23137643634356633 253 3 L--w 6 1
>> 0 mpath-23137643634356633
>> dedbd337--ca66--43ff--b78c--4e9347682a9c-master 253 9 L--w 0 1
>> 0 LVM-zSZKhQXbs2Ms5ceM2Wlt3p46BgsS8gGIiEnFZklRhZfFZ4YRdYWFImKW
>> sUGr5pHg
>> dedbd337--ca66--43ff--b78c--4e9347682a9c-inbox 253 8 L--w 0 1
>> 0 LVM-zSZKhQXbs2Ms5ceM2Wlt3p46BgsS8gGILobNK8KRD4SzDWyg50aG7jGd
>> cNAi3KNw
>> dedbd337--ca66--43ff--b78c--4e9347682a9c-outbox 253 5 L--w 0 1
>> 0 LVM-zSZKhQXbs2Ms5ceM2Wlt3p46BgsS8gGIAvbT8CLegbVL802bG3QgLtH7
>> I7llmS6R
>> flash_sdf 253 2 L--w 1 1
>> 0
>> dedbd337--ca66--43ff--b78c--4e9347682a9c-ids 253 7 L--w 1 1
>> 0 LVM-zSZKhQXbs2Ms5ceM2Wlt3p46BgsS8gGIIkeaFaWvxa9wbHy7WrfiXNsP
>> 4F2J3gg0
>> flash_sdd 253 1 L--w 1 1
>> 0
>>
>> ====================================================================
>> and flashcache dm have no UUID ,I think it can export it before
>> checking .
>>
>>
>> humaorong
>> 2016-11-25
>>
>>
>>
>> ------------------ Original ------------------
>> *From: * "胡茂荣"<maorong.hu at horebdata.cn>;
>> *Date: * Fri, Nov 25, 2016 01:18 PM
>> *To: * "Sahina Bose"<sabose at redhat.com>;
>> *Cc: * "Maor Lipchuk"<mlipchuk at redhat.com>; "Jeff Nelson"<
>> jenelson at redhat.com>; "users"<users at ovirt.org>;
>> *Subject: * Re: [ovirt-users] can not use iscsi storage type on
>> ovirtandGlusterfshyper-converged environment
>>
>>
>> I find this problem more info :
>> I use flashcache on ovirt hosts , have dm device in /dev/mapper/ :
>>
>> [root at horeba init.d]# dmsetup info -C (and I set them in multipath
>> blacklist )
>> Name Maj Min Stat Open Targ Event UUID
>>
>> flash_sdb 253 0 L--w 0 1 0
>>
>> flash_sdf 253 2 L--w 0 1 0
>>
>> flash_sdd 253 1 L--w 0 1 0
>> [root at horeba init.d]# multipath -l
>> [root at horeba init.d]#
>>
>> [root at horeba init.d]# ll /dev/mapper/
>> total 0
>> crw------- 1 root root 10, 236 Nov 25 10:09 control
>> lrwxrwxrwx 1 root root 7 Nov 25 12:51 flash_sdb -> ../dm-0
>> lrwxrwxrwx 1 root root 7 Nov 25 12:51 flash_sdd -> ../dm-1
>> lrwxrwxrwx 1 root root 7 Nov 25 12:51 flash_sdf -> ../dm-2
>>
>> on this condition , ovirt UI add iscsi type storage will fail .
>>
>> If I delete the flashcache device /dev/mapper/flash_* , ovirt UI
>> add iscsi type storage have no problem .
>>
>> I need flashcache for using ssd cache on my environment , how can I
>> use iscsi type storage on this environment , please help me , Thanks!
>>
>>
>> ------------------ Original ------------------
>> *From: * "Sahina Bose"<sabose at redhat.com>;
>> *Date: * Thu, Nov 3, 2016 05:54 PM
>> *To: * "胡茂荣"<maorong.hu at horebdata.cn>;
>> *Cc: * "Maor Lipchuk"<mlipchuk at redhat.com>; "Jeff Nelson"<
>> jenelson at redhat.com>; "users"<users at ovirt.org>;
>> *Subject: * Re: [ovirt-users] can not use iscsi storage type on
>> ovirtandGlusterfshyper-converged environment
>>
>> A wild guess, not sure if it is related - can you check if multipathd
>> service is enabled. If you set up your oVirt-Gluster hyperconverged
>> environment via gdeploy, multipathd service is disabled and the
>> /etc/multipath.conf is edited to blacklist all devices - this was to fix
>> https://bugzilla.redhat.com/show_bug.cgi?id=1303940
>>
>> Since you mention you're unable to add iscsi storage only in this
>> environment, thought it's worth checking.
>>
>> On Thu, Nov 3, 2016 at 6:40 AM, 胡茂荣 <maorong.hu at horebdata.cn> wrote:
>>
>>>
>>> my environment rpm are :
>>> [root at horeba ~]# rpm -q vdsm
>>> vdsm-4.18.13-1.el7.centos.x86_64
>>>
>>> [root at horeba ~]# rpm -aq | grep ovirt
>>> ovirt-vmconsole-host-1.0.4-1.el7.centos.noarch
>>> ovirt-imageio-common-0.4.0-1.el7.noarch
>>> ovirt-hosted-engine-setup-2.0.2.2-1.el7.centos.noarch
>>> ovirt-imageio-daemon-0.4.0-1.el7.noarch
>>> ovirt-engine-appliance-4.0-20160928.1.el7.centos.noarch
>>> ovirt-vmconsole-1.0.4-1.el7.centos.noarch
>>> ovirt-host-deploy-1.5.2-1.el7.centos.noarch
>>> ovirt-hosted-engine-ha-2.0.4-1.el7.centos.noarch
>>> ovirt-release40-4.0.4-1.noarch
>>> ovirt-setup-lib-1.0.2-1.el7.centos.noarch
>>> ovirt-engine-sdk-python-3.6.9.1-1.el7.centos.noarch
>>>
>>> and I test if not on ' ovirt and Glusterfs hyper-converged
>>> environment ' , ovirt WEB UI add iscsi storage work ok .
>>>
>>>
>>> ------------------ Original ------------------
>>> *From: * "Maor Lipchuk"<mlipchuk at redhat.com>;
>>> *Date: * Wed, Nov 2, 2016 07:37 PM
>>> *To: * "胡茂荣"<maorong.hu at horebdata.cn>;
>>> *Cc: * "users"<users at ovirt.org>; "Jeff Nelson"<jenelson at redhat.com>;
>>> "Nir Soffer"<nsoffer at redhat.com>;
>>> *Subject: * Re: [ovirt-users] can not use iscsi storage type on
>>> ovirtandGlusterfs hyper-converged environment
>>>
>>> Thanks for the logs,
>>>
>>> What kind of VDSM version are you using?
>>> "rpm -q vdsm"
>>> There seems to be a similar issue which was reported recently in the
>>> VDSM area
>>> (see https://bugzilla.redhat.com/show_bug.cgi?id=1197292)
>>> It should be fixed in later versions of VDSM vdsm-4.16.12-2.el7ev.x86_64
>>> Adding also Nir and Jeff to the thread, if they have any insights
>>>
>>> Regards,
>>> Maor
>>>
>>> On Wed, Nov 2, 2016 at 4:11 AM, 胡茂荣 <maorong.hu at horebdata.cn> wrote:
>>>
>>>>
>>>> Hi Maor:
>>>> vdsm/supervdsm/engine log on attachment . I mkfs.xfs the lun
>>>> block device and mount to /mnt , dd write ,dmesg not report error ,dd
>>>> result is ok :
>>>>
>>>> /dev/sdi 50G 33M 50G 1% /mnt
>>>>
>>>> [root at horebc mnt]# for i in `seq 3`; do dd if=/dev/zero of=./file
>>>> bs=1G count=1 oflag=direct ; done
>>>> 1+0 records in
>>>> 1+0 records out
>>>> 1073741824 bytes (1.1 GB) copied, 13.3232 s, 80.6 MB/s
>>>> 1+0 records in
>>>> 1+0 records out
>>>> 1073741824 bytes (1.1 GB) copied, 9.89988 s, 108 MB/s
>>>> 1+0 records in
>>>> 1+0 records out
>>>> 1073741824 bytes (1.1 GB) copied, 14.0143 s, 76.6 MB/s
>>>>
>>>> my envirnment have three network segments (hosts have 3 network
>>>> segments ) :
>>>> engine and glusterfs mount : 192.168.11.X/24
>>>> glusterfs brick : 192.168.10.x/24
>>>> iscsi : 192.168.1.0/24
>>>>
>>>> and I add 192.168.1.0/24 to engine vm , ovirt web UI report the
>>>> same error .
>>>>
>>>> humaorong
>>>> 2016-11-2
>>>>
>>>> ------------------ Original ------------------
>>>> *From: * "Maor Lipchuk"<mlipchuk at redhat.com>;
>>>> *Date: * Tue, Nov 1, 2016 08:14 PM
>>>> *To: * "胡茂荣"<maorong.hu at horebdata.cn>;
>>>> *Cc: * "users"<users at ovirt.org>;
>>>> *Subject: * Re: [ovirt-users] can not use iscsi storage type on ovirt
>>>> andGlusterfs hyper-converged environment
>>>>
>>>> Hi 胡茂荣Can u please also add the VDSM and engine logs.
>>>> If you try discover and connect to those luns directly from your Host
>>>> does it work?
>>>>
>>>> Regards,
>>>> Maor
>>>>
>>>>
>>>> On Tue, Nov 1, 2016 at 6:12 AM, 胡茂荣 <maorong.hu at horebdata.cn> wrote:
>>>>
>>>>>
>>>>>
>>>>> on ovirt and Glusterfs hyper-converged environment , can not use
>>>>> iscsi storage type , UI report error: "Could not retrieve LUNs,
>>>>> please check your storage." , vdsm report :"VDSM hosted_engine_3
>>>>> command failed: Error block device action: ()" .
>>>>> but this block device alse login on centos 7 host :
>>>>> =============================================================
>>>>>
>>>>> ## lsscsi
>>>>>
>>>>> [7:0:0:0] disk SCST_BIO DEVFOR_OVIRT_rbd 221 /dev/sdi
>>>>>
>>>>> ## dmesg :
>>>>>
>>>>> [684521.131186] sd 7:0:0:0: [sdi] Attached SCSI disk
>>>>>
>>>>> ===================================---
>>>>>
>>>>> ###vdsm or supervdsm log report :
>>>>>
>>>>> MainProcess|jsonrpc.Executor/7::ERROR::2016-11-01
>>>>> 11:07:00,178::supervdsmServer::96::SuperVdsm.ServerCallback::(wrapper)
>>>>> Error in getPathsStatus
>>>>>
>>>>> MainProcess|jsonrpc.Executor/4::ERROR::2016-11-01
>>>>> 11:07:20,964::supervdsmServer::96::SuperVdsm.ServerCallback::(wrapper)
>>>>> Error in getPathsStatus
>>>>>
>>>>> jsonrpc.Executor/4::DEBUG::2016-11-01
>>>>> 11:07:04,251::iscsi::434::Storage.ISCSI::(rescan) Performing SCSI
>>>>> scan, this will take up to 30 seconds
>>>>>
>>>>> jsonrpc.Executor/5::INFO::2016-11-01 11:07:19,413::iscsi::567::Stor
>>>>> age.ISCSI::(setRpFilterIfNeeded) iSCSI iface.net_ifacename not
>>>>> provided. Skipping.
>>>>>
>>>>> 11:09:15,753::iscsiadm::119::Storage.Misc.excCmd::(_runCmd)
>>>>> /usr/bin/taskset --cpu-list 0-7 /usr/bin/sudo -n /usr/sbin/iscsiadm -m
>>>>> session -R (cwd None)
>>>>>
>>>>> ======================================
>>>>>
>>>>> the other info please the attachment "bug-info.doc".
>>>>>
>>>>> this prolem on ovirt3.6 and 4.X ovirt and Glusterfs
>>>>> hyper-converged environment . how can I use iscsi storage type on ovirt
>>>>> and Glusterfs hyper-converged environment .Please help me !
>>>>>
>>>>> humaorong
>>>>>
>>>>> 2016-11-1
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>> _______________________________________________
>>>>> Users mailing list
>>>>> Users at ovirt.org
>>>>> http://lists.ovirt.org/mailman/listinfo/users
>>>>>
>>>>>
>>>>
>>>
>>> _______________________________________________
>>> Users mailing list
>>> Users at ovirt.org
>>> http://lists.ovirt.org/mailman/listinfo/users
>>>
>>>
>>
>> _______________________________________________
>> Users mailing list
>> Users at ovirt.org
>> http://lists.ovirt.org/mailman/listinfo/users
>>
>>
>>
>> _______________________________________________
>> Users mailing list
>> Users at ovirt.org
>> http://lists.ovirt.org/mailman/listinfo/users
>>
>>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.ovirt.org/pipermail/users/attachments/20161206/e5d38319/attachment-0001.html>
More information about the Users
mailing list