[ovirt-users] glusterfs tips/questions

Gabi C gabicr at gmail.com
Wed May 21 13:04:43 UTC 2014


..or should I:

-stop volumes
-remove brick belonging to the affected node
-remove afected node/peer
-add thenode, brick then start volumes?



On Wed, May 21, 2014 at 1:13 PM, Gabi C <gabicr at gmail.com> wrote:

> On afected node:
>
> gluster peer status
>
> gluster peer status
> Number of Peers: 3
>
> Hostname: 10.125.1.194
> Uuid: 85c2a08c-a955-47cc-a924-cf66c6814654
> State: Peer in Cluster (Connected)
>
> Hostname: 10.125.1.196
> Uuid: c22e41b8-2818-4a96-a6df-a237517836d6
> State: Peer in Cluster (Connected)
>
> Hostname: 10.125.1.194
> Uuid: 85c2a08c-a955-47cc-a924-cf66c6814654
> State: Peer in Cluster (Connected)
>
>
>
>
>
> ls -la /var/lib/gluster
>
>
>
> ls -la /var/lib/glusterd/peers/
> total 20
> drwxr-xr-x. 2 root root 4096 May 21 11:10 .
> drwxr-xr-x. 9 root root 4096 May 21 11:09 ..
> -rw-------. 1 root root   73 May 21 11:10
> 85c2a08c-a955-47cc-a924-cf66c6814654
> -rw-------. 1 root root   73 May 21 10:52
> c22e41b8-2818-4a96-a6df-a237517836d6
> -rw-------. 1 root root   73 May 21 11:10
> d95558a0-a306-4812-aec2-a361a9ddde3e
>
>
> Shoul I delete d95558a0-a306-4812-aec2-a361a9ddde3e??
>
>
>
>
>
> On Wed, May 21, 2014 at 12:00 PM, Kanagaraj <kmayilsa at redhat.com> wrote:
>
>>
>> On 05/21/2014 02:04 PM, Gabi C wrote:
>>
>>   Hello!
>>
>>  I have an ovirt setup, 3.4.1, up-to date, with gluster package
>> 3.5.0-3.fc19 on all 3 nodes. Glusterfs setup is replicated on 3 bricks. On
>> 2 nodes 'gluster peeer status' raise 2 peer connected with it's UUID. On
>> third node 'gluster peer status' raise 3 peers, out of which, two reffer to
>> same node/IP but different UUID.
>>
>>
>> in every node you can find the peers in /var/lib/glusterd/peers/
>>
>> you can get the uuid of the current node using the command "gluster
>> system:: uuid get"
>>
>> From this you can find which file is wrong in the above location.
>>
>> [Adding gluster-users at ovirt.org]
>>
>>
>>  What I have tried:
>>  - stopped gluster volumes, put 3rd node in maintenace, reboor -> no
>> effect;
>>  - stopped  volumes, removed bricks belonging to 3rd node, readded it,
>> start volumes but still no effect.
>>
>>
>>  Any ideas, hints?
>>
>>  TIA
>>
>>
>> _______________________________________________
>> Users mailing listUsers at ovirt.orghttp://lists.ovirt.org/mailman/listinfo/users
>>
>>
>>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.ovirt.org/pipermail/users/attachments/20140521/1f5fdd87/attachment-0001.html>


More information about the Users mailing list