[ovirt-users] GetGlusterVolumeAdvancedDetailsQuery & GetGlusterVolumeProfileInfoQuery when using seperate storage network
Jorick Astrego
j.astrego at netbulae.eu
Wed May 6 17:52:00 UTC 2015
On 05/06/2015 06:24 PM, knarra wrote:
> On 05/06/2015 06:59 PM, Jorick Astrego wrote:
>>
>>
>> On 05/06/2015 02:49 PM, knarra wrote:
>>> On 05/06/2015 05:33 PM, Jorick Astrego wrote:
>>>> Hi,
>>>>
>>>> Looking forward to bz 1049994 Allow choosing network interface
>>>> for gluster domain traffic.
>>>>
>>>> Currently I have the bricks on a different storage network and
>>>> can't get the volume details or profile it.
>>>>
>>>> Will this be handled in 3.6 properly? I don't see any changes in
>>>> gerrit regarding this, but I can be ovelooking it.
>>>>
>>>> The errors I get currently:
>>>>
>>>> Could not fetch brick profile stats
>>>>
>>>> 2015-05-06 10:34:22,430 ERROR
>>>> [org.ovirt.engine.core.bll.gluster.GetGlusterVolumeProfileInfoQuery]
>>>> (ajp--127.0.0.1-8702-27) Query GetGlusterVolumeProfileInfoQuery
>>>> failed. Exception message is null :
>>>> java.lang.NullPointerException: java.lang.NullPointerException
>>>>
>>>> and
>>>>
>>>> Error in fetching the brick details, please try again.
>>>>
>>>> 2015-05-06 10:36:14,205 ERROR
>>>> [org.ovirt.engine.core.bll.gluster.GetGlusterVolumeAdvancedDetailsQuery]
>>>> (ajp--127.0.0.1-8702-55) Query
>>>> GetGlusterVolumeAdvancedDetailsQuery failed. Exception message
>>>> is VdcBLLException: Volume status failed
>>>> error: Staging failed on *.*.*.*. Error: No brick
>>>> glustertest1.netbulae.test/gluster/brick1 in volume data
>>>> Staging failed on *.*.*.*. Error: No brick
>>>> glustertest1.netbulae.test:/gluster/brick1 in volume data
>>>> return code: -1 (Failed with error GlusterVolumeStatusFailed
>>>> and code 4157) :
>>>> org.ovirt.engine.core.common.errors.VdcBLLException:
>>>> VdcBLLException: Volume status failed
>>>> error: Staging failed on *.*.*.*. Error: No brick
>>>> glustertest1.netbulae.test:/gluster/brick1 in volume data
>>>> Staging failed on *.*.*.*. Error: No brick
>>>> glustertest1.netbulae.test:/gluster/brick1 in volume data
>>>> return code: -1 (Failed with error GlusterVolumeStatusFailed
>>>> and code 4157):
>>>> org.ovirt.engine.core.common.errors.VdcBLLException:
>>>> VdcBLLException: Volume status failed
>>>> error: Staging failed on *.*.*.*. Error: No brick
>>>> glustertest1.netbulae.test:/gluster/brick1 in volume data
>>>> Staging failed on *.*.*.*. Error: No brick
>>>> glustertest1.netbulae.test:/gluster/brick1 in volume data
>>>> return code: -1 (Failed with error GlusterVolumeStatusFailed
>>>> and code 4157)
>>>>
>>>>
>>>>
>>>>
>>>>
>>>> Met vriendelijke groet, With kind regards,
>>>>
>>>> Jorick Astrego*
>>>>
>>>> Netbulae Virtualization Experts *
>>>> ------------------------------------------------------------------------
>>>> Tel: 053 20 30 270 info at netbulae.eu Staalsteden 4-3A KvK 08198180
>>>> Fax: 053 20 30 271 www.netbulae.eu 7547 TA Enschede BTW
>>>> NL821234584B01
>>>>
>>>>
>>>> ------------------------------------------------------------------------
>>>>
>>>>
>>>>
>>>> _______________________________________________
>>>> Users mailing list
>>>> Users at ovirt.org
>>>> http://lists.ovirt.org/mailman/listinfo/users
>>> Hi,
>>>
>>> Can you please check what does your gluster peer status on each
>>> of your nodes return? I guess they are in disconnected state and
>>> that is the reason you are not able to view these details.
>>>
>>> Thanks
>>> kasturi
>>>
>> On the nodes it gives me the following:
>>
>> gluster peer status
>> Connection failed. Please check if gluster daemon is operational.
>>
> This means that glusterd is not running on this node. you could check
> the status of glusterd by running the command 'service glusterd status'.
>
> please start glusterd by running the command 'service glusterd start'
> on both of your nodes.
>
> Ideally when glusterd goes down node in ovirt should move to
> non-operational. Because of this BZ 1207150 as of now it is not
> changing the state to non operational.
There is no glusterd on the compute nodes in our setup, we have two
clusters. One for virt hosts only and one for GlusterFS only.
Like I said, everything is Up and running fine. It's just that I can't
get the stats because the hostname != GlusterFS NIC ip
>>
>>
>> But everyting is up and ovirt found the manually configured volume
>> perfectly. But the hostname it lists as glustertest1.netbulae.test is
>> not what my volume uses for communication as I created the volume
>> using the ip's of the storage network.
>>
>>
>> gluster peer status
>> Number of Peers: 2
>>
>> Hostname: 10.1.1.3
>> Uuid: 1cc0875e-1699-42ae-aed2-9152667ed5af
>> State: Peer in Cluster (Connected)
>>
>> Hostname: 10.1.1.2
>> Uuid: a0b3ac13-7388-441a-a238-1deb023cab6c
>> State: Peer in Cluster (Connected)
>>
> Did you import already existing cluster ?
No, I provisioned the nodes. Added them to our glusterfs cluster (with
virt service disabled) and created the volume manually.
oVirt auto-discovered the manual created volume after that.
Error: No brick *glustertest1.netbulae.test*:/gluster/brick1 in
volume data
Hostname: *10.1.1.3*
Things should work better in 3.6 (bz1049994), but I don't see any code
changes to "GetGlusterVolumeProfileInfoQuery" linked to this in Bugzilla.
Met vriendelijke groet, With kind regards,
Jorick Astrego
Netbulae Virtualization Experts
----------------
Tel: 053 20 30 270 info at netbulae.eu Staalsteden 4-3A KvK 08198180
Fax: 053 20 30 271 www.netbulae.eu 7547 TA Enschede BTW NL821234584B01
----------------
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.ovirt.org/pipermail/users/attachments/20150506/22bad742/attachment-0001.html>
More information about the Users
mailing list