[ovirt-users] Client-quorum not met - Distributed-Replicate gluster volume

George Skorup george at mwcomm.com
Mon Feb 16 12:00:24 EST 2015


You would need six storage hosts in total to maintain quorum if even one 
of the hosts goes down. There's no way to decide who's right with 
replica 2. When you have 2 out of 3 online, majority rules.

I have a four node cluster doing replica 4, no distribute. I can take 
one host down. If two are down, quorum is not met and the volumes go 
read-only. Same issue applies, only 50% is online.

On 2/16/2015 5:20 AM, Wesley Schaft wrote:
> Hi,
>
> I've set up 4 oVirt nodes with Gluster storage to provide high available virtual machines.
> The Gluster volumes are Distributed-Replicate with a replica count of 2.
>
> The extra volume options are configured:
>
> cat /var/lib/glusterd/groups/virt
> quick-read=off
> read-ahead=off
> io-cache=off
> stat-prefetch=off
> eager-lock=enable
> remote-dio=enable
> quorum-type=auto
> server-quorum-type=server
>
>
> Volume for the self-hosted engine:
> gluster volume info engine
>
> Volume Name: engine
> Type: Distributed-Replicate
> Volume ID: 9e7a3265-1e91-46e1-a0ba-09c5cc1fc1c1
> Status: Started
> Number of Bricks: 2 x 2 = 4
> Transport-type: tcp
> Bricks:
> Brick1: gluster004:/gluster/engine/004
> Brick2: gluster005:/gluster/engine/005
> Brick3: gluster006:/gluster/engine/006
> Brick4: gluster007:/gluster/engine/007
> Options Reconfigured:
> cluster.quorum-type: auto
> storage.owner-gid: 36
> storage.owner-uid: 36
> cluster.server-quorum-type: server
> network.remote-dio: enable
> cluster.eager-lock: enable
> performance.stat-prefetch: off
> performance.io-cache: off
> performance.read-ahead: off
> performance.quick-read: off
> network.ping-timeout: 10
>
>
> Volume for the virtual machines:
> gluster volume info data
>
> Volume Name: data
> Type: Distributed-Replicate
> Volume ID: 896db323-7ac4-4023-82a6-a8815a4d06b4
> Status: Started
> Number of Bricks: 2 x 2 = 4
> Transport-type: tcp
> Bricks:
> Brick1: gluster004:/gluster/data/004
> Brick2: gluster005:/gluster/data/005
> Brick3: gluster006:/gluster/data/006
> Brick4: gluster007:/gluster/data/007
> Options Reconfigured:
> cluster.quorum-type: auto
> performance.quick-read: off
> performance.read-ahead: off
> performance.io-cache: off
> performance.stat-prefetch: off
> cluster.eager-lock: enable
> network.remote-dio: enable
> storage.owner-uid: 36
> storage.owner-gid: 36
> cluster.server-quorum-type: server
> network.ping-timeout: 10
>
>
> Everything seems to be working fine.
> However, when I stop the storage network on gluster004 or gluster006, client-quorum is lost.
> Client-quorum isn't lost when the storage network is stopped on gluster005 or gluster007.
>
> [2015-02-16 07:05:58.541531] W [MSGID: 108001] [afr-common.c:3635:afr_notify] 0-data-replicate-1: Client-quorum is not met
> [2015-02-16 07:05:58.541579] W [MSGID: 108001] [afr-common.c:3635:afr_notify] 0-engine-replicate-1: Client-quorum is not met
>
> And as a result, the volumes are read-only and the VM's are paused.
>
> I've added a "dummy" gluster node for quorum use (no bricks, only running glusterd), but that didn't help.
>
> gluster peer status
> Number of Peers: 4
>
> Hostname: gluster005
> Uuid: 6c5253b4-b1c6-4d0a-9e6b-1f3efc1e8086
> State: Peer in Cluster (Connected)
>
> Hostname: gluster006
> Uuid: 4b3d15c4-2de0-4d2e-aa4c-3981e47dadbd
> State: Peer in Cluster (Connected)
>
> Hostname: gluster007
> Uuid: 165e9ada-addb-496e-abf7-4a4efda4d5d3
> State: Peer in Cluster (Connected)
>
> Hostname: glusterdummy
> Uuid: 3ef8177b-2394-429b-a58e-ecf0f6ce79a0
> State: Peer in Cluster (Connected)
>
>
> The 4 nodes are running CentOS 7, with the following oVirt / Gluster packages:
>
> glusterfs-3.6.2-1.el7.x86_64
> glusterfs-api-3.6.2-1.el7.x86_64
> glusterfs-cli-3.6.2-1.el7.x86_64
> glusterfs-fuse-3.6.2-1.el7.x86_64
> glusterfs-libs-3.6.2-1.el7.x86_64
> glusterfs-rdma-3.6.2-1.el7.x86_64
> glusterfs-server-3.6.2-1.el7.x86_64
> ovirt-engine-sdk-python-3.5.1.0-1.el7.centos.noarch
> ovirt-host-deploy-1.3.1-1.el7.noarch
> ovirt-hosted-engine-ha-1.2.5-1.el7.centos.noarch
> ovirt-hosted-engine-setup-1.2.2-1.el7.centos.noarch
> vdsm-gluster-4.16.10-8.gitc937927.el7.noarch
>
>
> The self-hosted engine is running CentOS 6 with ovirt-engine-3.5.1-1.el6.noarch
>
> Regards,
> Wesley
>
>
>
>
>
> _______________________________________________
> Users mailing list
> Users at ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users



More information about the Users mailing list