[ovirt-users] upgrade issues

Demeter Tibor tdemeter at itsmart.hu
Tue Aug 18 12:54:40 UTC 2015


Hi,
1, r710cluster is contain node0,node1,node2.
2, I'm sorry, supervdsm.log wasn't attached. what nodes on are need to know it?
3, it was on node0

[root at node1 ~]# gluster peer status
Number of Peers: 1

Hostname: 172.16.0.12
Uuid: 0ccdaa63-758a-43a4-93f2-25214a3cfa12
State: Accepted peer request (Connected)

[root at node1 ~]# gluster volume info
No volumes present

:(
After this I did it on node2

[root at node2 ~]# gluster peer status
Number of Peers: 2

Hostname: 172.16.0.10
Uuid: 664273d9-a77f-476d-ad09-3e126c4b19a7
State: Peer in Cluster (Connected)

Hostname: 172.16.0.11
Uuid: dc7a1611-f653-4d64-a426-b5ef589de289
State: Accepted peer request (Connected)

[root at node2 ~]# gluster volume info
 
Volume Name: g2sata
Type: Replicate
Volume ID: 49d76fc8-853e-4c7d-82a5-b12ec98dadd8
Status: Started
Number of Bricks: 1 x 2 = 2
Transport-type: tcp
Bricks:
Brick1: 172.16.0.10:/data/sata/brick2
Brick2: 172.16.0.12:/data/sata/brick2
Options Reconfigured:
nfs.disable: on
user.cifs: disable
auth.allow: 172.16.*
storage.owner-uid: 36
storage.owner-gid: 36
 
Volume Name: g4sata
Type: Replicate
Volume ID: f26ed231-c951-431f-8a2f-e8818b58cfb4
Status: Started
Number of Bricks: 1 x 2 = 2
Transport-type: tcp
Bricks:
Brick1: 172.16.0.10:/data/sata/iso
Brick2: 172.16.0.12:/data/sata/iso
Options Reconfigured:
nfs.disable: off
user.cifs: disable
auth.allow: 172.16.0.*
storage.owner-uid: 36
storage.owner-gid: 36

Many thanks,


Tibor



----- 2015. aug.. 17., 16:03, Sahina Bose sabose at redhat.com írta:

> From the engine.log - the gluster volumes are queried on host "node1"
> which returns no volumes.
> 
> 1. Your cluster "r710cluster1" - which nodes are added to it? node1
> alone or node0 and node2 as well?
> 
> 2. Was the attached supervdsm.log from node1?
> 
> 3. Which node was the below "gluster volume info"  output from? What is
> the output of "gluster peer status" and "gluster volume info" on node1?
> 
> 
> 
> On 08/17/2015 12:49 PM, Demeter Tibor wrote:
>> Dear Sahina,
>>
>> Thank you for your reply.
>>
>> Volume Name: g2sata
>> Type: Replicate
>> Volume ID: 49d76fc8-853e-4c7d-82a5-b12ec98dadd8
>> Status: Started
>> Number of Bricks: 1 x 2 = 2
>> Transport-type: tcp
>> Bricks:
>> Brick1: 172.16.0.10:/data/sata/brick2
>> Brick2: 172.16.0.12:/data/sata/brick2
>> Options Reconfigured:
>> nfs.disable: on
>> user.cifs: disable
>> auth.allow: 172.16.*
>> storage.owner-uid: 36
>> storage.owner-gid: 36
>>   
>> Volume Name: g4sata
>> Type: Replicate
>> Volume ID: f26ed231-c951-431f-8a2f-e8818b58cfb4
>> Status: Started
>> Number of Bricks: 1 x 2 = 2
>> Transport-type: tcp
>> Bricks:
>> Brick1: 172.16.0.10:/data/sata/iso
>> Brick2: 172.16.0.12:/data/sata/iso
>> Options Reconfigured:
>> nfs.disable: off
>> user.cifs: disable
>> auth.allow: 172.16.0.*
>> storage.owner-uid: 36
>> storage.owner-gid: 36
>>
>> Also, I have attached the logs .
>>
>> Thanks in advance,
>>
>>
>>
>> Tibor
>>
>> ----- 2015. aug.. 17., 8:40, Sahina Bose sabose at redhat.com írta:
>>
>>> Please provide output of "gluster volume info" command,  vdsm.log &
>>> engine.log
>>>
>>> There could be a mismatch between node information in engine database
>>> and gluster - one of the reasons is because the gluster server uuid
>>> changed on the node and we will need to see why.
>>>
>>>
>>> On 08/17/2015 12:35 AM, Demeter Tibor wrote:
>>>> Hi All,
>>>>
>>>> I have to upgrade ovirt 3.5.0 to 3.5.3. We have a 3 node system and we have a
>>>> gluster replica beetwen 2 node of these 3 servers.
>>>> I had gluster volume beetwen node0 and node2
>>>> But I wanted to do a new volume beetwen node1 and node2.
>>>> It didn't work, it complety kill my node1, because glusterd does not stated.I
>>>> always have to got always error: gluster peer rejected (related to node1).
>>>> I followed this article
>>>> http://www.gluster.org/community/documentation/index.php/Resolving_Peer_Rejected
>>>> and it was good, my gluster service already worked, but ovirt got these errors
>>>>
>>>> 	
>>>> Detected deletion of volume g2sata on cluster r710cluster1, and deleted it from
>>>> engine DB.
>>>> Detected deletion of volume g4sata on cluster r710cluster1, and deleted it from
>>>> engine DB.
>>>> 	
>>>> And ovirt does not see my gluster volumes anymore.
>>>>
>>>> I've checked with gluster volume status and gluster volume heal g2sata info, it
>>>> seems to be working, my VMs are ok.
>>>>
>>>>
>>>> How can I reimport my lost volumes to ovirt?
>>>>
>>>> Thanks in advance,
>>>>
>>>> Tibor
>>>> _______________________________________________
>>>> Users mailing list
>>>> Users at ovirt.org
> >>> http://lists.ovirt.org/mailman/listinfo/users



More information about the Users mailing list