[ovirt-users] glusterfs tips/questions
Kanagaraj
kmayilsa at redhat.com
Wed May 21 13:26:29 UTC 2014
What are the steps which led this situation?
Did you re-install one of the nodes after forming the cluster or reboot
which could have changed the ip?
On 05/21/2014 03:43 PM, Gabi C wrote:
> On afected node:
>
> gluster peer status
>
> gluster peer status
> Number of Peers: 3
>
> Hostname: 10.125.1.194
> Uuid: 85c2a08c-a955-47cc-a924-cf66c6814654
> State: Peer in Cluster (Connected)
>
> Hostname: 10.125.1.196
> Uuid: c22e41b8-2818-4a96-a6df-a237517836d6
> State: Peer in Cluster (Connected)
>
> Hostname: 10.125.1.194
> Uuid: 85c2a08c-a955-47cc-a924-cf66c6814654
> State: Peer in Cluster (Connected)
>
>
>
>
>
> ls -la /var/lib/gluster
>
>
>
> ls -la /var/lib/glusterd/peers/
> total 20
> drwxr-xr-x. 2 root root 4096 May 21 11:10 .
> drwxr-xr-x. 9 root root 4096 May 21 11:09 ..
> -rw-------. 1 root root 73 May 21 11:10
> 85c2a08c-a955-47cc-a924-cf66c6814654
> -rw-------. 1 root root 73 May 21 10:52
> c22e41b8-2818-4a96-a6df-a237517836d6
> -rw-------. 1 root root 73 May 21 11:10
> d95558a0-a306-4812-aec2-a361a9ddde3e
>
>
> Shoul I delete d95558a0-a306-4812-aec2-a361a9ddde3e??
>
>
>
>
>
> On Wed, May 21, 2014 at 12:00 PM, Kanagaraj <kmayilsa at redhat.com
> <mailto:kmayilsa at redhat.com>> wrote:
>
>
> On 05/21/2014 02:04 PM, Gabi C wrote:
>> Hello!
>>
>> I have an ovirt setup, 3.4.1, up-to date, with gluster package
>> 3.5.0-3.fc19 on all 3 nodes. Glusterfs setup is replicated on 3
>> bricks. On 2 nodes 'gluster peeer status' raise 2 peer connected
>> with it's UUID. On third node 'gluster peer status' raise 3
>> peers, out of which, two reffer to same node/IP but different UUID.
>
> in every node you can find the peers in /var/lib/glusterd/peers/
>
> you can get the uuid of the current node using the command
> "gluster system:: uuid get"
>
> From this you can find which file is wrong in the above location.
>
> [Adding gluster-users at ovirt.org <mailto:gluster-users at ovirt.org>]
>
>>
>> What I have tried:
>> - stopped gluster volumes, put 3rd node in maintenace, reboor ->
>> no effect;
>> - stopped volumes, removed bricks belonging to 3rd node, readded
>> it, start volumes but still no effect.
>>
>>
>> Any ideas, hints?
>>
>> TIA
>>
>>
>> _______________________________________________
>> Users mailing list
>> Users at ovirt.org <mailto:Users at ovirt.org>
>> http://lists.ovirt.org/mailman/listinfo/users
>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.ovirt.org/pipermail/users/attachments/20140521/05e76bb6/attachment-0001.html>
More information about the Users
mailing list