On Mon, Nov 7, 2016 at 3:27 PM, Gary Pedretty <gary(a)ravnalaska.net> wrote:
[root@fai-kvm-1-gfs admin]# gluster volume status data2
Status of volume: data2
Gluster process TCP Port RDMA Port Online
Pid
------------------------------------------------------------
------------------
Brick fai-kvm-1-vmn.ravnalaska.net:/kvm2/gl
uster/data2/brick 49156 0 Y
3484
Brick fai-kvm-2-vmn.ravnalaska.net:/kvm2/gl
uster/data2/brick 49156 0 Y
34791
Brick fai-kvm-3-vmn.ravnalaska.net:/kvm2/gl
uster/data2/brick 49156 0 Y
177340
Brick fai-kvm-4-vmn.ravnalaska.net:/kvm2/gl
uster/data2/brick 49152 0 Y
146038
NFS Server on localhost 2049 0 Y
40844
Self-heal Daemon on localhost N/A N/A Y
40865
NFS Server on
fai-kvm-2-gfs.ravnalaska.net 2049 0 Y
99905
Self-heal Daemon on fai-kvm-2-gfs.ravnalask
a.net N/A N/A Y
99915
NFS Server on
fai-kvm-4-gfs.ravnalaska.net 2049 0 Y
176305
Self-heal Daemon on fai-kvm-4-gfs.ravnalask
a.net N/A N/A Y
176326
NFS Server on
fai-kvm-3-gfs.ravnalaska.net 2049 0 Y
226271
Self-heal Daemon on fai-kvm-3-gfs.ravnalask
a.net N/A N/A Y
226287
Task Status of Volume data2
------------------------------------------------------------
------------------
There are no active volume tasks
[root@fai-kvm-1-gfs admin]# gluster volume info data2
Volume Name: data2
Type: Striped-Replicate
Volume ID: 20f85c9a-541b-4df4-9dba-44c5179bbfb0
Status: Started
Number of Bricks: 1 x 2 x 2 = 4
Transport-type: tcp
Bricks:
Brick1: fai-kvm-1-vmn.ravnalaska.net:/kvm2/gluster/data2/brick
Brick2: fai-kvm-2-vmn.ravnalaska.net:/kvm2/gluster/data2/brick
Brick3: fai-kvm-3-vmn.ravnalaska.net:/kvm2/gluster/data2/brick
Brick4: fai-kvm-4-vmn.ravnalaska.net:/kvm2/gluster/data2/brick
Options Reconfigured:
performance.readdir-ahead: on
performance.quick-read: off
performance.read-ahead: off
performance.io-cache: off
performance.stat-prefetch: off
cluster.eager-lock: enable
network.remote-dio: enable
cluster.quorum-type: auto
cluster.server-quorum-type: server
storage.owner-uid: 36
storage.owner-gid: 36
See attached file for the mount log.
Striped-Replicate is no longer supported in GlusterFS upstream. Instead,
you should be using a Distribute-Replicate with sharding enabled. Also when
using a gluster volume as storage domain, it is recommended to use replica
3.
From the mount logs, there is no indication as to why the volume is
unmounted frequently. Could you try again with a replica 3 volume that has
sharding enabled?
Gary
------------------------------------------------------------------------
Gary Pedretty gary(a)ravnalaska.net
<gary(a)eraalaska.net>
Systems Manager
www.flyravn.com
Ravn Alaska /\ 907-450-7251
5245 Airport Industrial Road / \/\ 907-450-7238 fax
Fairbanks, Alaska 99709 /\ / \ \ Second greatest commandment
Serving All of Alaska / \/ /\ \ \/\ “Love your neighbor as
Really loving the record green up date! Summmer!! yourself” Matt 22:39
------------------------------------------------------------------------
On Nov 6, 2016, at 9:50 PM, Sahina Bose <sabose(a)redhat.com> wrote:
However your volume configuration seems suspect -"stripe 2 replica 2". Can
you provide gluster volume info of your second storage domain gluster
volume? The mount logs of the volume (under
/var/log/glusterfs/rhev-datacenter..<volname>.log)
from the host where the volume is being mounted will also help.