[ovirt-users] Multiple Data Storage Domains
Sahina Bose
sabose at redhat.com
Mon Nov 7 11:10:29 UTC 2016
On Mon, Nov 7, 2016 at 3:27 PM, Gary Pedretty <gary at ravnalaska.net> wrote:
> [root at fai-kvm-1-gfs admin]# gluster volume status data2
> Status of volume: data2
> Gluster process TCP Port RDMA Port Online
> Pid
> ------------------------------------------------------------
> ------------------
> Brick fai-kvm-1-vmn.ravnalaska.net:/kvm2/gl
> uster/data2/brick 49156 0 Y
> 3484
> Brick fai-kvm-2-vmn.ravnalaska.net:/kvm2/gl
> uster/data2/brick 49156 0 Y
> 34791
> Brick fai-kvm-3-vmn.ravnalaska.net:/kvm2/gl
> uster/data2/brick 49156 0 Y
> 177340
> Brick fai-kvm-4-vmn.ravnalaska.net:/kvm2/gl
> uster/data2/brick 49152 0 Y
> 146038
> NFS Server on localhost 2049 0 Y
> 40844
> Self-heal Daemon on localhost N/A N/A Y
> 40865
> NFS Server on fai-kvm-2-gfs.ravnalaska.net 2049 0 Y
> 99905
> Self-heal Daemon on fai-kvm-2-gfs.ravnalask
> a.net N/A N/A Y
> 99915
> NFS Server on fai-kvm-4-gfs.ravnalaska.net 2049 0 Y
> 176305
> Self-heal Daemon on fai-kvm-4-gfs.ravnalask
> a.net N/A N/A Y
> 176326
> NFS Server on fai-kvm-3-gfs.ravnalaska.net 2049 0 Y
> 226271
> Self-heal Daemon on fai-kvm-3-gfs.ravnalask
> a.net N/A N/A Y
> 226287
>
> Task Status of Volume data2
> ------------------------------------------------------------
> ------------------
> There are no active volume tasks
>
>
> [root at fai-kvm-1-gfs admin]# gluster volume info data2
>
> Volume Name: data2
> Type: Striped-Replicate
> Volume ID: 20f85c9a-541b-4df4-9dba-44c5179bbfb0
> Status: Started
> Number of Bricks: 1 x 2 x 2 = 4
> Transport-type: tcp
> Bricks:
> Brick1: fai-kvm-1-vmn.ravnalaska.net:/kvm2/gluster/data2/brick
> Brick2: fai-kvm-2-vmn.ravnalaska.net:/kvm2/gluster/data2/brick
> Brick3: fai-kvm-3-vmn.ravnalaska.net:/kvm2/gluster/data2/brick
> Brick4: fai-kvm-4-vmn.ravnalaska.net:/kvm2/gluster/data2/brick
> Options Reconfigured:
> performance.readdir-ahead: on
> performance.quick-read: off
> performance.read-ahead: off
> performance.io-cache: off
> performance.stat-prefetch: off
> cluster.eager-lock: enable
> network.remote-dio: enable
> cluster.quorum-type: auto
> cluster.server-quorum-type: server
> storage.owner-uid: 36
> storage.owner-gid: 36
>
>
> See attached file for the mount log.
>
Striped-Replicate is no longer supported in GlusterFS upstream. Instead,
you should be using a Distribute-Replicate with sharding enabled. Also when
using a gluster volume as storage domain, it is recommended to use replica
3.
>From the mount logs, there is no indication as to why the volume is
unmounted frequently. Could you try again with a replica 3 volume that has
sharding enabled?
>
> Gary
>
>
> ------------------------------------------------------------------------
> Gary Pedretty gary at ravnalaska.net
> <gary at eraalaska.net>
> Systems Manager www.flyravn.com
> Ravn Alaska /\ 907-450-7251
> 5245 Airport Industrial Road / \/\ 907-450-7238 fax
> Fairbanks, Alaska 99709 /\ / \ \ Second greatest commandment
> Serving All of Alaska / \/ /\ \ \/\ “Love your neighbor as
> Really loving the record green up date! Summmer!! yourself” Matt 22:39
> ------------------------------------------------------------------------
>
>
>
>
>
>
>
>
>
>
>
>
> On Nov 6, 2016, at 9:50 PM, Sahina Bose <sabose at redhat.com> wrote:
>
> However your volume configuration seems suspect -"stripe 2 replica 2". Can
> you provide gluster volume info of your second storage domain gluster
> volume? The mount logs of the volume (under /var/log/glusterfs/rhev-datacenter..<volname>.log)
> from the host where the volume is being mounted will also help.
>
>
>
>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.ovirt.org/pipermail/users/attachments/20161107/300adaf0/attachment-0001.html>
More information about the Users
mailing list