[ovirt-users] glusterfs tips/questions

Gabi C gabicr at gmail.com
Wed May 21 13:38:40 UTC 2014


Hello!


I haven't change the IP, nor reinstall nodes. All nodes are updated via
yum. All I can think of was that after having some issue with gluster,from
WebGUI I deleted VM, deactivate and detach storage domains ( I have 2) ,
than, *manually*, from one of the nodes , remove bricks, then detach peers,
probe them, add bricks again, bring the volume up, and readd storage
domains from the webGUI.


On Wed, May 21, 2014 at 4:26 PM, Kanagaraj <kmayilsa at redhat.com> wrote:

>  What are the steps which led this situation?
>
> Did you re-install one of the nodes after forming the cluster or reboot
> which could have changed the ip?
>
>
>
> On 05/21/2014 03:43 PM, Gabi C wrote:
>
>  On afected node:
>
> gluster peer status
>
> gluster peer status
> Number of Peers: 3
>
> Hostname: 10.125.1.194
> Uuid: 85c2a08c-a955-47cc-a924-cf66c6814654
> State: Peer in Cluster (Connected)
>
> Hostname: 10.125.1.196
> Uuid: c22e41b8-2818-4a96-a6df-a237517836d6
> State: Peer in Cluster (Connected)
>
> Hostname: 10.125.1.194
> Uuid: 85c2a08c-a955-47cc-a924-cf66c6814654
> State: Peer in Cluster (Connected)
>
>
>
>
>
>  ls -la /var/lib/gluster
>
>
>
> ls -la /var/lib/glusterd/peers/
> total 20
> drwxr-xr-x. 2 root root 4096 May 21 11:10 .
> drwxr-xr-x. 9 root root 4096 May 21 11:09 ..
> -rw-------. 1 root root   73 May 21 11:10
> 85c2a08c-a955-47cc-a924-cf66c6814654
> -rw-------. 1 root root   73 May 21 10:52
> c22e41b8-2818-4a96-a6df-a237517836d6
> -rw-------. 1 root root   73 May 21 11:10
> d95558a0-a306-4812-aec2-a361a9ddde3e
>
>
>  Shoul I delete d95558a0-a306-4812-aec2-a361a9ddde3e??
>
>
>
>
>
> On Wed, May 21, 2014 at 12:00 PM, Kanagaraj <kmayilsa at redhat.com> wrote:
>
>>
>> On 05/21/2014 02:04 PM, Gabi C wrote:
>>
>>   Hello!
>>
>>  I have an ovirt setup, 3.4.1, up-to date, with gluster package
>> 3.5.0-3.fc19 on all 3 nodes. Glusterfs setup is replicated on 3 bricks. On
>> 2 nodes 'gluster peeer status' raise 2 peer connected with it's UUID. On
>> third node 'gluster peer status' raise 3 peers, out of which, two reffer to
>> same node/IP but different UUID.
>>
>>
>>  in every node you can find the peers in /var/lib/glusterd/peers/
>>
>> you can get the uuid of the current node using the command "gluster
>> system:: uuid get"
>>
>> From this you can find which file is wrong in the above location.
>>
>> [Adding gluster-users at ovirt.org]
>>
>>
>>  What I have tried:
>>  - stopped gluster volumes, put 3rd node in maintenace, reboor -> no
>> effect;
>>  - stopped  volumes, removed bricks belonging to 3rd node, readded it,
>> start volumes but still no effect.
>>
>>
>>  Any ideas, hints?
>>
>>  TIA
>>
>>
>>  _______________________________________________
>> Users mailing listUsers at ovirt.orghttp://lists.ovirt.org/mailman/listinfo/users
>>
>>
>>
>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.ovirt.org/pipermail/users/attachments/20140521/dc5a5777/attachment-0001.html>


More information about the Users mailing list