[ovirt-users] GetGlusterVolumeAdvancedDetailsQuery & GetGlusterVolumeProfileInfoQuery when using seperate storage network
knarra
knarra at redhat.com
Wed May 6 18:15:31 UTC 2015
On 05/06/2015 11:22 PM, Jorick Astrego wrote:
>
>
> On 05/06/2015 06:24 PM, knarra wrote:
>> On 05/06/2015 06:59 PM, Jorick Astrego wrote:
>>>
>>>
>>> On 05/06/2015 02:49 PM, knarra wrote:
>>>> On 05/06/2015 05:33 PM, Jorick Astrego wrote:
>>>>> Hi,
>>>>>
>>>>> Looking forward to bz 1049994 Allow choosing network interface
>>>>> for gluster domain traffic.
>>>>>
>>>>> Currently I have the bricks on a different storage network and
>>>>> can't get the volume details or profile it.
>>>>>
>>>>> Will this be handled in 3.6 properly? I don't see any changes in
>>>>> gerrit regarding this, but I can be ovelooking it.
>>>>>
>>>>> The errors I get currently:
>>>>>
>>>>> Could not fetch brick profile stats
>>>>>
>>>>> 2015-05-06 10:34:22,430 ERROR
>>>>> [org.ovirt.engine.core.bll.gluster.GetGlusterVolumeProfileInfoQuery]
>>>>> (ajp--127.0.0.1-8702-27) Query
>>>>> GetGlusterVolumeProfileInfoQuery failed. Exception message is
>>>>> null : java.lang.NullPointerException:
>>>>> java.lang.NullPointerException
>>>>>
>>>>> and
>>>>>
>>>>> Error in fetching the brick details, please try again.
>>>>>
>>>>> 2015-05-06 10:36:14,205 ERROR
>>>>> [org.ovirt.engine.core.bll.gluster.GetGlusterVolumeAdvancedDetailsQuery]
>>>>> (ajp--127.0.0.1-8702-55) Query
>>>>> GetGlusterVolumeAdvancedDetailsQuery failed. Exception message
>>>>> is VdcBLLException: Volume status failed
>>>>> error: Staging failed on *.*.*.*. Error: No brick
>>>>> glustertest1.netbulae.test/gluster/brick1 in volume data
>>>>> Staging failed on *.*.*.*. Error: No brick
>>>>> glustertest1.netbulae.test:/gluster/brick1 in volume data
>>>>> return code: -1 (Failed with error GlusterVolumeStatusFailed
>>>>> and code 4157) :
>>>>> org.ovirt.engine.core.common.errors.VdcBLLException:
>>>>> VdcBLLException: Volume status failed
>>>>> error: Staging failed on *.*.*.*. Error: No brick
>>>>> glustertest1.netbulae.test:/gluster/brick1 in volume data
>>>>> Staging failed on *.*.*.*. Error: No brick
>>>>> glustertest1.netbulae.test:/gluster/brick1 in volume data
>>>>> return code: -1 (Failed with error GlusterVolumeStatusFailed
>>>>> and code 4157):
>>>>> org.ovirt.engine.core.common.errors.VdcBLLException:
>>>>> VdcBLLException: Volume status failed
>>>>> error: Staging failed on *.*.*.*. Error: No brick
>>>>> glustertest1.netbulae.test:/gluster/brick1 in volume data
>>>>> Staging failed on *.*.*.*. Error: No brick
>>>>> glustertest1.netbulae.test:/gluster/brick1 in volume data
>>>>> return code: -1 (Failed with error GlusterVolumeStatusFailed
>>>>> and code 4157)
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>> Met vriendelijke groet, With kind regards,
>>>>>
>>>>> Jorick Astrego*
>>>>>
>>>>> Netbulae Virtualization Experts *
>>>>> ------------------------------------------------------------------------
>>>>> Tel: 053 20 30 270 info at netbulae.eu Staalsteden 4-3A KvK 08198180
>>>>> Fax: 053 20 30 271 www.netbulae.eu 7547 TA Enschede BTW
>>>>> NL821234584B01
>>>>>
>>>>>
>>>>> ------------------------------------------------------------------------
>>>>>
>>>>>
>>>>>
>>>>> _______________________________________________
>>>>> Users mailing list
>>>>> Users at ovirt.org
>>>>> http://lists.ovirt.org/mailman/listinfo/users
>>>> Hi,
>>>>
>>>> Can you please check what does your gluster peer status on each
>>>> of your nodes return? I guess they are in disconnected state and
>>>> that is the reason you are not able to view these details.
>>>>
>>>> Thanks
>>>> kasturi
>>>>
>>> On the nodes it gives me the following:
>>>
>>> gluster peer status
>>> Connection failed. Please check if gluster daemon is operational.
>>>
>> This means that glusterd is not running on this node. you could check
>> the status of glusterd by running the command 'service glusterd status'.
>>
>> please start glusterd by running the command 'service glusterd start'
>> on both of your nodes.
>>
>> Ideally when glusterd goes down node in ovirt should move to
>> non-operational. Because of this BZ 1207150 as of now it is not
>> changing the state to non operational.
>
> There is no glusterd on the compute nodes in our setup, we have two
> clusters. One for virt hosts only and one for GlusterFS only.
>
>
> Like I said, everything is Up and running fine. It's just that I can't
> get the stats because the hostname != GlusterFS NIC ip
>
>
>>>
>>>
>>> But everyting is up and ovirt found the manually configured volume
>>> perfectly. But the hostname it lists as glustertest1.netbulae.test
>>> is not what my volume uses for communication as I created the volume
>>> using the ip's of the storage network.
>>>
>>>
>>> gluster peer status
>>> Number of Peers: 2
>>>
>>> Hostname: 10.1.1.3
>>> Uuid: 1cc0875e-1699-42ae-aed2-9152667ed5af
>>> State: Peer in Cluster (Connected)
>>>
>>> Hostname: 10.1.1.2
>>> Uuid: a0b3ac13-7388-441a-a238-1deb023cab6c
>>> State: Peer in Cluster (Connected)
>>>
>> Did you import already existing cluster ?
>
> No, I provisioned the nodes. Added them to our glusterfs cluster (with
> virt service disabled) and created the volume manually.
>
> oVirt auto-discovered the manual created volume after that.
>
> Error: No brick *glustertest1.netbulae.test*:/gluster/brick1 in
> volume data
>
> Hostname: *10.1.1.3*
>
> Things should work better in 3.6 (bz1049994), but I don't see any code
> changes to "GetGlusterVolumeProfileInfoQuery" linked to this in Bugzilla.
>
HI Jorik,
For more information on using separate storage network please refer
to the following feature page.
http://www.ovirt.org/Features/Select_Network_For_Gluster
Thanks
kasturi.
>
>
>
>
>
>
> Met vriendelijke groet, With kind regards,
>
> Jorick Astrego*
>
> Netbulae Virtualization Experts *
> ------------------------------------------------------------------------
> Tel: 053 20 30 270 info at netbulae.eu Staalsteden 4-3A KvK 08198180
> Fax: 053 20 30 271 www.netbulae.eu 7547 TA Enschede BTW NL821234584B01
>
>
> ------------------------------------------------------------------------
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.ovirt.org/pipermail/users/attachments/20150506/77149181/attachment-0001.html>
More information about the Users
mailing list