Hello!


I haven't change the IP, nor reinstall nodes. All nodes are updated via yum. All I can think of was that after having some issue with gluster,from WebGUI I deleted VM, deactivate and detach storage domains ( I have 2) , than, manually, from one of the nodes , remove bricks, then detach peers, probe them, add bricks again, bring the volume up, and readd storage domains from the webGUI.


On Wed, May 21, 2014 at 4:26 PM, Kanagaraj <kmayilsa@redhat.com> wrote:
What are the steps which led this situation?

Did you re-install one of the nodes after forming the cluster or reboot which could have changed the ip?



On 05/21/2014 03:43 PM, Gabi C wrote:
On afected node:

gluster peer status

gluster peer status
Number of Peers: 3

Hostname: 10.125.1.194
Uuid: 85c2a08c-a955-47cc-a924-cf66c6814654
State: Peer in Cluster (Connected)

Hostname: 10.125.1.196
Uuid: c22e41b8-2818-4a96-a6df-a237517836d6
State: Peer in Cluster (Connected)

Hostname: 10.125.1.194
Uuid: 85c2a08c-a955-47cc-a924-cf66c6814654
State: Peer in Cluster (Connected)





ls -la /var/lib/gluster



ls -la /var/lib/glusterd/peers/
total 20
drwxr-xr-x. 2 root root 4096 May 21 11:10 .
drwxr-xr-x. 9 root root 4096 May 21 11:09 ..
-rw-------. 1 root root   73 May 21 11:10 85c2a08c-a955-47cc-a924-cf66c6814654
-rw-------. 1 root root   73 May 21 10:52 c22e41b8-2818-4a96-a6df-a237517836d6
-rw-------. 1 root root   73 May 21 11:10 d95558a0-a306-4812-aec2-a361a9ddde3e


Shoul I delete d95558a0-a306-4812-aec2-a361a9ddde3e??





On Wed, May 21, 2014 at 12:00 PM, Kanagaraj <kmayilsa@redhat.com> wrote:

On 05/21/2014 02:04 PM, Gabi C wrote:
Hello!

I have an ovirt setup, 3.4.1, up-to date, with gluster package 3.5.0-3.fc19 on all 3 nodes. Glusterfs setup is replicated on 3 bricks. On 2 nodes 'gluster peeer status' raise 2 peer connected with it's UUID. On third node 'gluster peer status' raise 3 peers, out of which, two reffer to same node/IP but different UUID.

in every node you can find the peers in /var/lib/glusterd/peers/

you can get the uuid of the current node using the command "gluster system:: uuid get"

From this you can find which file is wrong in the above location.

[Adding gluster-users@ovirt.org]


What I have tried:
- stopped gluster volumes, put 3rd node in maintenace, reboor -> no effect;
- stopped  volumes, removed bricks belonging to 3rd node, readded it, start volumes but still no effect.


Any ideas, hints?

TIA


_______________________________________________
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users