Okay, both the BZ and ML posts are interesting and helpful.  I'm kind of surprised there seems to be so much trouble and confusion for what I would have thought to be a very common setup.  Are most people using something else?

I think this gives me what I need for my next stab at doing this but I"m still puzzled on how to tear down what I have in oVirt so that I can redo it.  Specifically, I didn't see how to delete the existing iSCSI connections.  I've read that this can only be done through the REST API.  I have managed to redo the interfaces on my Hosts so that everything is now on just 2 NICs each, leaving 2 NICs free for a foothold on a new setup.  From all of my experimentation, it would appear that my only option is to create a new storage domain and export/import each disk volume one by one.  Maybe there's a migration option I have yet to see, but I don't see any way around creating a new storage domain here.

On 1/21/19 7:12 AM, Vinícius Ferrão wrote:
Hello people, in the past Maor Lipchuk (from RH) tried very hard to help me and Uwe but we was unable to converge on the solution.

This was discussed a year ago and on my understanding it still and oVirt bug. As today, if you simple “DuckDuckGo” for “ovirt iscsi multipath not working” the third link points to this bugzilla: https://bugzilla.redhat.com/show_bug.cgi?id=1474904

Which is the one I’ve mentioned and it’s extremely similar to John Florian case, which was my case too.

@John, take a look at the bugzilla link and see if the desired topology match with your case.

Regards,


On 21 Jan 2019, at 05:21, Eitan Raviv <eraviv@redhat.com> wrote:

Shani,
Can you help here with  iSCSI bonding?
Thanks

On Mon, Jan 21, 2019 at 7:51 AM Uwe Laverenz <uwe@laverenz.de> wrote:

Hi John,

Am 20.01.19 um 18:32 schrieb John Florian:

As for how to get there, whatever exactly that might look like, I'm also
having troubles figuring that out.  I figured I would transform the
setup described below into one where each host has:

 * 2 NICs bonded with LACP for my ovirtmgmt and "main" net
 * 1 NIC for my 1st storage net
 * 1 NIC for my 2nd storage net

This is exactly the setup I use. I have run this successfully with
CentOS/LIO and FreeNAS iSCSI targets with good performance.

In short:

- 2 separate, isolated networks for iSCSI with dedicated adapters
  on hosts and iSCSI target
- jumbo frames enabled
- no VLANs config needed on hosts, untagged VLANs on switch
- do _not_ use LACP, let multipathd handle failovers

Same experience as Vinicius: what did _not_ work for me is the
iSCSI-Bonding in OVirt. It seems to require that all storage IPs are
reachable from all other IPs, which is not the case in every setup.

To get multipathing to work I use multipath directly:

https://www.mail-archive.com/users@ovirt.org/msg42735.html

I will post a bonnie++ result later. If you need more details please let
me know.

cu,
Uwe
_______________________________________________
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-leave@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/
List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/E2QKV7CZR27NT6MRSNL352KLOQ5OAGDR/


_______________________________________________
Users mailing list -- users@ovirt.org
To unsubscribe send an email to users-leave@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/
List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/NSE5BCLJSIFDX2VDZRBRLODEH3ZCPYWN/