On 02/12/2014 11:21 PM, Gianluca Cecchi wrote:
On Wed, Feb 12, 2014 at 6:18 PM, ml ml wrote:
> I guess the brick details are stored in the postgres database and everything
> else after will fail?!
Yes, we fixed the issue with resolving brick's host while syncing with
gluster CLI in oVirt 3.4. However, when you use multiple addresses, you
will need to use the workaround below.
>
> Am i the only one with dedicated migration/storage interfaces? :)
>
> Thanks,
> Mario
>
One of the workarounds I found and that works for me as I'm not using
dns is this:
- for engine host node1 and node two have ip on mgmt
- for node1 and node2 their own ip addresses are on dedicated gluster network
so for example
10.4.4.x = mgmt
192.168.3.x = gluster dedicated
before:
on engine
/etc/hosts
10.4.4.58 node01
10.4.4.59 node02
10.4.4.60 engine
on node01
10.4.4.58 node01
10.4.4.59 node02
10.4.4.60 engine
after:
on engine (the same as before)
/etc/hosts
10.4.4.58 node01
10.4.4.59 node02
10.4.4.60 engine
on node01
#10.4.4.58 node01
#10.4.4.59 node02
192.168.3.1 node01
192.168.3.3 node02
10.4.4.60 engine
No operations on RDBMS.
Thanks, Gianluca!
I will update the wiki page so that this workaround is clear.
HIH,
Gianluca