On Mon, Nov 7, 2016 at 3:27 PM, Gary Pedretty <gary@ravnalaska.net> wrote:
[root@fai-kvm-1-gfs admin]# gluster volume status data2
Status of volume: data2
Gluster process                             TCP Port  RDMA Port  Online  Pid
------------------------------------------------------------------------------
uster/data2/brick                           49156     0          Y       3484
uster/data2/brick                           49156     0          Y       34791
uster/data2/brick                           49156     0          Y       177340
uster/data2/brick                           49152     0          Y       146038
NFS Server on localhost                     2049      0          Y       40844
Self-heal Daemon on localhost               N/A       N/A        Y       40865
NFS Server on fai-kvm-2-gfs.ravnalaska.net  2049      0          Y       99905
Self-heal Daemon on fai-kvm-2-gfs.ravnalask
a.net                                       N/A       N/A        Y       99915
NFS Server on fai-kvm-4-gfs.ravnalaska.net  2049      0          Y       176305
Self-heal Daemon on fai-kvm-4-gfs.ravnalask
a.net                                       N/A       N/A        Y       176326
NFS Server on fai-kvm-3-gfs.ravnalaska.net  2049      0          Y       226271
Self-heal Daemon on fai-kvm-3-gfs.ravnalask
a.net                                       N/A       N/A        Y       226287

Task Status of Volume data2
------------------------------------------------------------------------------
There are no active volume tasks


[root@fai-kvm-1-gfs admin]# gluster volume info data2

Volume Name: data2
Type: Striped-Replicate
Volume ID: 20f85c9a-541b-4df4-9dba-44c5179bbfb0
Status: Started
Number of Bricks: 1 x 2 x 2 = 4
Transport-type: tcp
Bricks:
Brick1: fai-kvm-1-vmn.ravnalaska.net:/kvm2/gluster/data2/brick
Brick2: fai-kvm-2-vmn.ravnalaska.net:/kvm2/gluster/data2/brick
Brick3: fai-kvm-3-vmn.ravnalaska.net:/kvm2/gluster/data2/brick
Brick4: fai-kvm-4-vmn.ravnalaska.net:/kvm2/gluster/data2/brick
Options Reconfigured:
performance.readdir-ahead: on
performance.quick-read: off
performance.read-ahead: off
performance.io-cache: off
performance.stat-prefetch: off
cluster.eager-lock: enable
network.remote-dio: enable
cluster.quorum-type: auto
cluster.server-quorum-type: server
storage.owner-uid: 36
storage.owner-gid: 36


See attached file for the mount log.


Striped-Replicate is no longer supported in GlusterFS upstream. Instead, you should be using a Distribute-Replicate with sharding enabled. Also when using a gluster volume as storage domain, it is recommended to use replica 3.

From the mount logs, there is no indication as to why the volume is unmounted frequently. Could you try again with a replica 3 volume that has sharding enabled?
 

Gary


------------------------------------------------------------------------
Gary Pedretty                                        gary@ravnalaska.net
Systems Manager                                          www.flyravn.com
Ravn Alaska                           /\                    907-450-7251
5245 Airport Industrial Road         /  \/\             907-450-7238 fax
Fairbanks, Alaska  99709        /\  /    \ \ Second greatest commandment
Serving All of Alaska          /  \/  /\  \ \/\   “Love your neighbor as
Really loving the record green up date! Summmer!!   yourself” Matt 22:39
------------------------------------------------------------------------












On Nov 6, 2016, at 9:50 PM, Sahina Bose <sabose@redhat.com> wrote:

However your volume configuration seems suspect -"stripe 2 replica 2". Can you provide gluster volume info of your second storage domain gluster volume? The mount logs of the volume (under /var/log/glusterfs/rhev-datacenter..<volname>.log) from the host where the volume is being mounted will also help.