
On Mon, Nov 7, 2016 at 3:27 PM, Gary Pedretty <gary@ravnalaska.net> wrote:
[root@fai-kvm-1-gfs admin]# gluster volume status data2 Status of volume: data2 Gluster process TCP Port RDMA Port Online Pid ------------------------------------------------------------ ------------------ Brick fai-kvm-1-vmn.ravnalaska.net:/kvm2/gl uster/data2/brick 49156 0 Y 3484 Brick fai-kvm-2-vmn.ravnalaska.net:/kvm2/gl uster/data2/brick 49156 0 Y 34791 Brick fai-kvm-3-vmn.ravnalaska.net:/kvm2/gl uster/data2/brick 49156 0 Y 177340 Brick fai-kvm-4-vmn.ravnalaska.net:/kvm2/gl uster/data2/brick 49152 0 Y 146038 NFS Server on localhost 2049 0 Y 40844 Self-heal Daemon on localhost N/A N/A Y 40865 NFS Server on fai-kvm-2-gfs.ravnalaska.net 2049 0 Y 99905 Self-heal Daemon on fai-kvm-2-gfs.ravnalask a.net N/A N/A Y 99915 NFS Server on fai-kvm-4-gfs.ravnalaska.net 2049 0 Y 176305 Self-heal Daemon on fai-kvm-4-gfs.ravnalask a.net N/A N/A Y 176326 NFS Server on fai-kvm-3-gfs.ravnalaska.net 2049 0 Y 226271 Self-heal Daemon on fai-kvm-3-gfs.ravnalask a.net N/A N/A Y 226287
Task Status of Volume data2 ------------------------------------------------------------ ------------------ There are no active volume tasks
[root@fai-kvm-1-gfs admin]# gluster volume info data2
Volume Name: data2 Type: Striped-Replicate Volume ID: 20f85c9a-541b-4df4-9dba-44c5179bbfb0 Status: Started Number of Bricks: 1 x 2 x 2 = 4 Transport-type: tcp Bricks: Brick1: fai-kvm-1-vmn.ravnalaska.net:/kvm2/gluster/data2/brick Brick2: fai-kvm-2-vmn.ravnalaska.net:/kvm2/gluster/data2/brick Brick3: fai-kvm-3-vmn.ravnalaska.net:/kvm2/gluster/data2/brick Brick4: fai-kvm-4-vmn.ravnalaska.net:/kvm2/gluster/data2/brick Options Reconfigured: performance.readdir-ahead: on performance.quick-read: off performance.read-ahead: off performance.io-cache: off performance.stat-prefetch: off cluster.eager-lock: enable network.remote-dio: enable cluster.quorum-type: auto cluster.server-quorum-type: server storage.owner-uid: 36 storage.owner-gid: 36
See attached file for the mount log.
Striped-Replicate is no longer supported in GlusterFS upstream. Instead, you should be using a Distribute-Replicate with sharding enabled. Also when using a gluster volume as storage domain, it is recommended to use replica 3.
From the mount logs, there is no indication as to why the volume is unmounted frequently. Could you try again with a replica 3 volume that has sharding enabled?
Gary
------------------------------------------------------------------------ Gary Pedretty gary@ravnalaska.net <gary@eraalaska.net> Systems Manager www.flyravn.com Ravn Alaska /\ 907-450-7251 5245 Airport Industrial Road / \/\ 907-450-7238 fax Fairbanks, Alaska 99709 /\ / \ \ Second greatest commandment Serving All of Alaska / \/ /\ \ \/\ “Love your neighbor as Really loving the record green up date! Summmer!! yourself” Matt 22:39 ------------------------------------------------------------------------
On Nov 6, 2016, at 9:50 PM, Sahina Bose <sabose@redhat.com> wrote:
However your volume configuration seems suspect -"stripe 2 replica 2". Can you provide gluster volume info of your second storage domain gluster volume? The mount logs of the volume (under /var/log/glusterfs/rhev-datacenter..<volname>.log) from the host where the volume is being mounted will also help.