[Users] Problem with creating a glusterfs volume

Dominic Kaiser dominic at bostonvineyard.org
Fri Sep 21 15:09:33 UTC 2012


I can mount to another computer with this command:

mount -o mountproto=tcp,vers=3 -t nfs gfs1.bostonvineyard.org:/data
/home/administrator/test

So volumes work but I get a 500 error timeout when trying to add as a
storage domain in ovirt.  weird?

dk

On Fri, Sep 21, 2012 at 10:44 AM, Dominic Kaiser <dominic at bostonvineyard.org
> wrote:

> Hey All,
>
> So I finally found the problem.  Cheap NIC's.  Installed Intel NIC's no
> problems creating gluster volumes and distributed replicated ones.
>  Broadcom and Realtek yuk!  So now I am trying to mount the gluster volume
> as a nfs mount and am having a problem.  It is timing out like it is
> blocked by a firewall.
>
> I am trying to:  mount -t nfs gfs1.bostonvineyard.org:/export
> /home/administrator/test
>
> Here is gfs1 tail vdsm.log
>
> [root at gfs1 vdsm]# tail vdsm.log
> Thread-88731::DEBUG::2012-09-21
> 10:35:56,566::resourceManager::844::ResourceManager.Owner::(cancelAll)
> Owner.cancelAll requests {}
> Thread-88731::DEBUG::2012-09-21
> 10:35:56,567::task::978::TaskManager.Task::(_decref)
> Task=`01b69eed-de59-4e87-8b28-5268b5dcbb50`::ref 0 aborting False
> Thread-88737::DEBUG::2012-09-21
> 10:36:06,890::task::588::TaskManager.Task::(_updateState)
> Task=`f70222ad-f8b4-4733-9526-eff1d214ebd8`::moving from state init ->
> state preparing
> Thread-88737::INFO::2012-09-21
> 10:36:06,891::logUtils::37::dispatcher::(wrapper) Run and protect:
> repoStats(options=None)
> Thread-88737::INFO::2012-09-21
> 10:36:06,891::logUtils::39::dispatcher::(wrapper) Run and protect:
> repoStats, Return response: {}
> Thread-88737::DEBUG::2012-09-21
> 10:36:06,891::task::1172::TaskManager.Task::(prepare)
> Task=`f70222ad-f8b4-4733-9526-eff1d214ebd8`::finished: {}
> Thread-88737::DEBUG::2012-09-21
> 10:36:06,892::task::588::TaskManager.Task::(_updateState)
> Task=`f70222ad-f8b4-4733-9526-eff1d214ebd8`::moving from state preparing ->
> state finished
> Thread-88737::DEBUG::2012-09-21
> 10:36:06,892::resourceManager::809::ResourceManager.Owner::(releaseAll)
> Owner.releaseAll requests {} resources {}
> Thread-88737::DEBUG::2012-09-21
> 10:36:06,892::resourceManager::844::ResourceManager.Owner::(cancelAll)
> Owner.cancelAll requests {}
> Thread-88737::DEBUG::2012-09-21
> 10:36:06,893::task::978::TaskManager.Task::(_decref)
> Task=`f70222ad-f8b4-4733-9526-eff1d214ebd8`::ref 0 aborting False
>
> Do you know why I can not connect via NFS?  Using an older kernel not 3.5
> and iptables are off.
>
> Dominic
>
>
> On Mon, Sep 10, 2012 at 12:20 PM, Haim Ateya <hateya at redhat.com> wrote:
>
>> On 09/10/2012 06:27 PM, Dominic Kaiser wrote:
>>
>>> Here is the message and the logs again except zipped I failed the first
>>> delivery:
>>>
>>> Ok here are the logs 4 node and 1 engine log.  Tried making /data folder
>>> owned by root and then tried by 36:36 neither worked.  Name of volume is
>>> data to match folders on nodes also.
>>>
>>> Let me know what you think,
>>>
>>> Dominic
>>>
>>
>> this is the actual failure (taken from gfs2vdsm.log).
>>
>> Thread-332442::DEBUG::2012-09-**10 10:28:05,788::BindingXMLRPC::**859::vds::(wrapper)
>> client [10.3.0.241]::call volumeCreate with ('data', ['10.4.0.97:/data',
>> '10.4.0.98:/data', '10.4.0.99:/data', '10.4.0.100:/data'],
>>  2, 0, ['TCP']) {} flowID [406f2c8e]
>> MainProcess|Thread-332442::**DEBUG::2012-09-10
>> 10:28:05,792::__init__::1249::**Storage.Misc.excCmd::(_log)
>> '/usr/sbin/gluster --mode=script volume create data replica 2 transport TCP
>> 10.4.0.97:/data 10.4.0.98:/data 10
>> .4.0.99:/data 10.4.0.100:/data' (cwd None)
>> MainProcess|Thread-332442::**DEBUG::2012-09-10
>> 10:28:05,900::__init__::1249::**Storage.Misc.excCmd::(_log) FAILED:
>> <err> = 'Host 10.4.0.99 not a friend\n'; <rc> = 255
>> MainProcess|Thread-332442::**ERROR::2012-09-10
>> 10:28:05,900::supervdsmServer:**:76::SuperVdsm.ServerCallback:**:(wrapper)
>> Error in wrapper
>> Traceback (most recent call last):
>>   File "/usr/share/vdsm/**supervdsmServer.py", line 74, in wrapper
>>     return func(*args, **kwargs)
>>   File "/usr/share/vdsm/**supervdsmServer.py", line 286, in wrapper
>>     return func(*args, **kwargs)
>>   File "/usr/share/vdsm/gluster/cli.**py", line 46, in wrapper
>>     return func(*args, **kwargs)
>>   File "/usr/share/vdsm/gluster/cli.**py", line 176, in volumeCreate
>>     raise ge.**GlusterVolumeCreateFailedExcep**tion(rc, out, err)
>> GlusterVolumeCreateFailedExcep**tion: Volume create failed
>> error: Host 10.4.0.99 not a friend
>> return code: 255
>> Thread-332442::ERROR::2012-09-**10 10:28:05,901::BindingXMLRPC::**877::vds::(wrapper)
>> unexpected error
>> Traceback (most recent call last):
>>   File "/usr/share/vdsm/**BindingXMLRPC.py", line 864, in wrapper
>>     res = f(*args, **kwargs)
>>   File "/usr/share/vdsm/gluster/api.**py", line 32, in wrapper
>>     rv = func(*args, **kwargs)
>>   File "/usr/share/vdsm/gluster/api.**py", line 87, in volumeCreate
>>     transportList)
>>   File "/usr/share/vdsm/supervdsm.py"**, line 67, in __call__
>>     return callMethod()
>>   File "/usr/share/vdsm/supervdsm.py"**, line 65, in <lambda>
>>     **kwargs)
>>   File "<string>", line 2, in glusterVolumeCreate
>>   File "/usr/lib64/python2.7/**multiprocessing/managers.py", line 759,
>> in _callmethod
>>     kind, result = conn.recv()
>> TypeError: ('__init__() takes exactly 4 arguments (1 given)', <class
>> 'gluster.exception.**GlusterVolumeCreateFailedExcep**tion'>, ())
>>
>> can you please run  gluster peer status on all your nodes ? also, it
>> appears that '10.4.0.99' is problematic, can you try create the volume
>> without it ?
>>
>>
>>
>>> On Mon, Sep 10, 2012 at 11:24 AM, Dominic Kaiser <
>>> dominic at bostonvineyard.org <mailto:dominic@**bostonvineyard.org<dominic at bostonvineyard.org>>>
>>> wrote:
>>>
>>>     Here are the other two logs forgot them.
>>>
>>>     dk
>>>
>>>
>>>     On Mon, Sep 10, 2012 at 11:19 AM, Dominic Kaiser
>>>     <dominic at bostonvineyard.org <mailto:dominic@**bostonvineyard.org<dominic at bostonvineyard.org>
>>> >>
>>>
>>>     wrote:
>>>
>>>         Ok here are the logs 4 node and 1 engine log.  Tried making
>>>         /data folder owned by root and then tried by 36:36 neither
>>>         worked.  Name of volume is data to match folders on nodes also.
>>>
>>>         Let me know what you think,
>>>
>>>         Dominic
>>>
>>>
>>>         On Thu, Sep 6, 2012 at 8:33 AM, Maxim Burgerhout
>>>         <maxim at wzzrd.com <mailto:maxim at wzzrd.com>> wrote:
>>>
>>>             I just ran into this as well, and it seems that you have
>>>             to either reformat previously used gluster bricks or
>>>             manually tweak some extended attributes.
>>>
>>>             Maybe this helps you in setting up your gluster volume,
>>>             Dominic?
>>>
>>>             More info here:
>>>             https://bugzilla.redhat.com/**show_bug.cgi?id=812214<https://bugzilla.redhat.com/show_bug.cgi?id=812214>
>>>
>>>
>>>             Maxim Burgerhout
>>>             maxim at wzzrd.com <mailto:maxim at wzzrd.com>
>>>
>>>             ----------------
>>>             EB11 5E56 E648 9D99 E8EF 05FB C513 6FD4 1302 B48A
>>>
>>>
>>>
>>>
>>>
>>>             On Thu, Sep 6, 2012 at 7:50 AM, Shireesh Anjal
>>>             <sanjal at redhat.com <mailto:sanjal at redhat.com>> wrote:
>>>
>>>                 Hi Dominic,
>>>
>>>                 Looking at the engine log immediately after trying to
>>>                 create the volume should tell us on which node the
>>>                 gluster volume creation was attempted. Then looking at
>>>                 the vdsm log on that node should help us identifying
>>>                 the exact reason for failure.
>>>
>>>                 In case this doesn't help you, can you please simulate
>>>                 the issue again and send back all the 5 log files?
>>>                 (engine.log from engine server and vdsm.log from the 4
>>>                 nodes)
>>>
>>>                 Regards,
>>>                 Shireesh
>>>
>>>
>>>                 On Wednesday 05 September 2012 11:50 PM, Dominic
>>>                 Kaiser wrote:
>>>
>>>>                 So I have a problem creating glusterfs volumes.  Here
>>>>                 is the install:
>>>>
>>>>                  1. Ovirt 3.1
>>>>                  2. 4 Nodes are Fedora 17 with kernel 3.3.4 -
>>>>                     5.fc17.x86_64
>>>>                  3. 4 nodes peer joined and running
>>>>                  4. 4 nodes added as hosts to ovirt
>>>>                  5. created a directory on each node this path /data
>>>>                  6. chmod 36.36 -R /data all nodes
>>>>                  7. went back to ovirt and created a
>>>>
>>>>                     distributed/replicated volume and added the 4
>>>>                     nodes with brick path of /data
>>>>
>>>>                 I received this error:
>>>>
>>>>                 Creation of Gluster Volume maingfs1 failed.
>>>>
>>>>                 I went and looked at the vdsm logs on the nodes and
>>>>                 the ovirt server which did not say much.  Where else
>>>>                 should I look?  Also this error is vague what does it
>>>>                 mean?
>>>>
>>>>
>>>>                 --                 Dominic Kaiser
>>>>                 Greater Boston Vineyard
>>>>                 Director of Operations
>>>>
>>>>                 cell: 617-230-1412 <tel:617-230-1412>
>>>>                 fax: 617-252-0238 <tel:617-252-0238>
>>>>                 email: dominic at bostonvineyard.org
>>>>                 <mailto:dominic@**bostonvineyard.org<dominic at bostonvineyard.org>
>>>> >
>>>>
>>>>
>>>>
>>>>
>>>>                 ______________________________**_________________
>>>>                 Users mailing list
>>>>                 Users at ovirt.org  <mailto:Users at ovirt.org>
>>>>                 http://lists.ovirt.org/**mailman/listinfo/users<http://lists.ovirt.org/mailman/listinfo/users>
>>>>
>>>
>>>
>>>                 ______________________________**_________________
>>>                 Users mailing list
>>>                 Users at ovirt.org <mailto:Users at ovirt.org>
>>>
>>>                 http://lists.ovirt.org/**mailman/listinfo/users<http://lists.ovirt.org/mailman/listinfo/users>
>>>
>>>
>>>
>>>
>>>
>>>         --         Dominic Kaiser
>>>         Greater Boston Vineyard
>>>         Director of Operations
>>>
>>>         cell: 617-230-1412 <tel:617-230-1412>
>>>         fax: 617-252-0238 <tel:617-252-0238>
>>>         email: dominic at bostonvineyard.org
>>>         <mailto:dominic@**bostonvineyard.org<dominic at bostonvineyard.org>
>>> >
>>>
>>>
>>>
>>>
>>>
>>>
>>>     --     Dominic Kaiser
>>>     Greater Boston Vineyard
>>>     Director of Operations
>>>
>>>     cell: 617-230-1412 <tel:617-230-1412>
>>>     fax: 617-252-0238 <tel:617-252-0238>
>>>     email: dominic at bostonvineyard.org <mailto:dominic@**
>>> bostonvineyard.org <dominic at bostonvineyard.org>>
>>>
>>>
>>>
>>>
>>>
>>>
>>> --
>>> Dominic Kaiser
>>> Greater Boston Vineyard
>>> Director of Operations
>>>
>>> cell: 617-230-1412
>>> fax: 617-252-0238
>>> email: dominic at bostonvineyard.org <mailto:dominic@**bostonvineyard.org<dominic at bostonvineyard.org>
>>> >
>>>
>>>
>>>
>>>
>>>
>>> ______________________________**_________________
>>> Users mailing list
>>> Users at ovirt.org
>>> http://lists.ovirt.org/**mailman/listinfo/users<http://lists.ovirt.org/mailman/listinfo/users>
>>>
>>
>>
>
>
> --
> Dominic Kaiser
> Greater Boston Vineyard
> Director of Operations
>
> cell: 617-230-1412
> fax: 617-252-0238
> email: dominic at bostonvineyard.org
>
>
>


-- 
Dominic Kaiser
Greater Boston Vineyard
Director of Operations

cell: 617-230-1412
fax: 617-252-0238
email: dominic at bostonvineyard.org
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.ovirt.org/pipermail/users/attachments/20120921/1858218e/attachment-0001.html>


More information about the Users mailing list