[ovirt-users] VDSM overwrites network config

Alan Cowles alan.cowles at gmail.com
Tue Apr 11 13:46:26 UTC 2017


Hey Dan,

Here is a layout of my environment setup, and the primary issue I have run
into, hopefully it helps to clear up the confusion.

Configure 2 hosts identically, eth0, eth1, eth0.502 (mgmt), eth1.504 (NFS)
and do the self-hosted engine install on one host.
As part of the install you have to identify the network that rhevm needs to
go on, and the network you need to mount your first NFS storage domain
from. In this case, eth0.502, and eth1.504 respectively.
That host, and the rhev-m hosted engine come up, however if you notice,
under Networks for the default datacenter, only rhevm exists as an actual
network.
You need to create a new network for NFS, in order to mount ISO/Data
storage domains, even though the engine setup has already mounted NFS via
eth1.504.
When you go to assign this new network from the engine, you cannot place it
on the eth1.504 only directly on eth0 or eth1 iteself.
Thus I have to be sure to tag that network in the rhev-m engine
When the tagged NFS network is placed on eth1, it looks like it breaks the
already existing NFS mount that supports the hosted engine, and causes
items to be non-responsive.
Rebooting the host at this stage, items still don't come up correctly and
the hosted engine remains down.
If I console to it, and manually setup the eth0 & eth1 interfaces, eth0.502
& eth1.504 VLANS,  and rhevm & NFS bridges, and reboot the host, the host
and the engine come up wonderfully with the defined networks vlan tagged,
and placed on the appropriate tag interfaces.

I then go to deploy a 2nd host as an additional hosted engine, and find
that I can select eth0.502 for rhevm and eth1.504 for nfs during the deploy
stages. But when it gets to the stage where it requires to you define the
networks that exist in the current cluster in order to activate the host
and proceed, I'm stuck in the same spot with applying networks, I can only
place them on the eth0/eth1 interfaces. I select ignore to exit the hosted
engine deployment wizard, and attempt to manually apply them, hoping to
repeat the steps from node 1 but was finding myself in a pickle because
starting VDSM would overwrite the network configs I had defined manually.
Why it does this on one host, and not on the other still perplexes me.

What I ended up doing once my primary host was rebuilt using the
appropriate bridges and vlan tagged interfaces, was reinstalling my 2nd
host completely, and configuring it as a self-hosted engine additional
host. This time it imports the network config from the first host
completely here, and I wind up with all tagged interfaces working correctly
and VDSM running as designed.

I guess the thing that bothered me mainly is the functionality in assigning
the networks in rhev manager, as it shows the vlans as sub-interfaces of
eth0/eth1, but doesn't let you assign networks to them, and the odd
behavior of VDSM overwriting configs on one host, but not the other?

I'll admit the setup I have is convoluted, but it's what I have to work
with for this project.

Thank you very much for the time and advice thus far.

On Sun, Apr 9, 2017 at 4:33 AM, Dan Kenigsberg <danken at redhat.com> wrote:

> On Fri, Apr 7, 2017 at 4:24 PM, Alan Cowles <alan.cowles at gmail.com> wrote:
> > Hey guys,
> >
> > I'm in a lab setup currently with 2 hosts, running RHEV-3.5, with a
> > self-hosted engine on RHEL 6.9 servers. I am doing this in order to plot
> out
> > a production upgrade I am planning going forward to 4.0, and I'm a bit
> stuck
> > and I'm hoping it's ok to ask questions here concerning this product and
> > version.
> >
> > In my lab, I have many vlans trunked on my switchports, so I have to
> create
> > individual vlan interfaces on my RHEL install. During the install, I am
> able
> > to pick my ifcfg-eth0.502 interface for rhevm, and ifcfg-eth1.504
> interface
> > for NFS, access the storage and create my self-hosted engine. The issue
> I am
> > running into is that I get into RHEV-M, and I am continuing to set the
> hosts
> > up or add other hosts, when I go to move my NFS network to host2 it only
> > allows me to select the base eth1 adapter, and not the VLAN tagged
> version.
> > I am able to tag the VLAN in the RHEV-M configured network itself, but
> this
> > has the unfortunate side effect of tagging a network on top of the
> already
> > tagged interface on host1, taking down NFS and the self hosted engine.
> >
> > I am able to access the console of host1, and I configure the ifcfg
> files,
> > vlan files, and bridge files to be on the correct interfaces, and I get
> my
> > host back up, and my RHEV-M back up. However when I try to make these
> manual
> > changes to host2 and get it up, the changes to these files are completely
> > overwritten the moment the host reboots connected to vdsmd start-up.
>
> If that was your only issue, I would have recommended you to read
> https://www.ovirt.org/blog/2016/05/modify-ifcfg-files/ and implement a
> hook that would leave the configuration as you wanted it.
>
>
> >
> > Right now, I have vdsmd disabled, and I have host2 configured the way I
> need
> > it to be with the rhevm bridge on eth0.502, the NFS bridge on eth1.504,
> and
> > my VMNet "guest" bridge on eth1.500, however that leaves me with a
> useless
> > host from RHEV standards.
> >
> > I've checked several different conf files to see where vdsmd is pulling
> it's
> > configuration from but I can't find it, or find a way to modify it to
> fit my
> > needs.
> >
> > Any advice or pointers here would be greatly appreciated. Thank you all
> in
> > advance.
>
> Pardon me for not clearly understanding the problem at hand.
>
> Could you specify your Engine-defined network names and vlan IDs? Can
> you specify the ifcfgs that you'd like to see on your hosts, and the
> ones re-generated on reboot?
>
> Dan.
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.ovirt.org/pipermail/users/attachments/20170411/a2bf1640/attachment.html>


More information about the Users mailing list