[ovirt-users] ovirt with glusterfs - big test - unwanted results
paf1 at email.cz
paf1 at email.cz
Thu Mar 31 12:09:05 UTC 2016
Hello,
we tried the following test - with unwanted results
input:
5 node gluster
A = replica 3 with arbiter 1 ( node1+node2+arbiter on node 5 )
B = replica 3 with arbiter 1 ( node3+node4+arbiter on node 5 )
C = distributed replica 3 arbiter 1 ( node1+node2, node3+node4, each
arbiter on node 5)
node 5 has only arbiter replica ( 4x )
TEST:
1) directly reboot one node - OK ( is not important which ( data node
or arbiter node ))
2) directly reboot two nodes - OK ( if nodes are not from the same
replica )
3) directly reboot three nodes - yes, this is the main problem and a
questions ....
- rebooted all three nodes from replica "B" ( not so possible, but
who knows ... )
- all VMs with data on this replica was paused ( no data access ) - OK
- all VMs running on replica "B" nodes lost ( started manually,
later )( datas on other replicas ) - acceptable
BUT
- !!! all oVIrt domains went down !! - master domain is on replica
"A" which lost only one member from three !!!
so we are not expecting that all domain will go down, especially
master with 2 live members.
Results:
- the whole cluster unreachable until at all domains up - depent of
all nodes up !!!
- all paused VMs started back - OK
- rest of all VMs rebooted and runnig - OK
Questions:
1) why all domains down if master domain ( on replica "A" ) has two
runnig members ( 2 of 3 ) ??
2) how to fix that colaps without waiting to all nodes up ? ( in
worste case if node has HW error eg. ) ??
3) which oVirt cluster policy can prevent that situation ?? ( if
any )
regs.
Pavel
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.ovirt.org/pipermail/users/attachments/20160331/50a0eaf1/attachment-0001.html>
More information about the Users
mailing list