[Users] Problem with creating a glusterfs volume
Dominic Kaiser
dominic at bostonvineyard.org
Fri Sep 21 14:44:35 UTC 2012
Hey All,
So I finally found the problem. Cheap NIC's. Installed Intel NIC's no
problems creating gluster volumes and distributed replicated ones.
Broadcom and Realtek yuk! So now I am trying to mount the gluster volume
as a nfs mount and am having a problem. It is timing out like it is
blocked by a firewall.
I am trying to: mount -t nfs gfs1.bostonvineyard.org:/export
/home/administrator/test
Here is gfs1 tail vdsm.log
[root at gfs1 vdsm]# tail vdsm.log
Thread-88731::DEBUG::2012-09-21
10:35:56,566::resourceManager::844::ResourceManager.Owner::(cancelAll)
Owner.cancelAll requests {}
Thread-88731::DEBUG::2012-09-21
10:35:56,567::task::978::TaskManager.Task::(_decref)
Task=`01b69eed-de59-4e87-8b28-5268b5dcbb50`::ref 0 aborting False
Thread-88737::DEBUG::2012-09-21
10:36:06,890::task::588::TaskManager.Task::(_updateState)
Task=`f70222ad-f8b4-4733-9526-eff1d214ebd8`::moving from state init ->
state preparing
Thread-88737::INFO::2012-09-21
10:36:06,891::logUtils::37::dispatcher::(wrapper) Run and protect:
repoStats(options=None)
Thread-88737::INFO::2012-09-21
10:36:06,891::logUtils::39::dispatcher::(wrapper) Run and protect:
repoStats, Return response: {}
Thread-88737::DEBUG::2012-09-21
10:36:06,891::task::1172::TaskManager.Task::(prepare)
Task=`f70222ad-f8b4-4733-9526-eff1d214ebd8`::finished: {}
Thread-88737::DEBUG::2012-09-21
10:36:06,892::task::588::TaskManager.Task::(_updateState)
Task=`f70222ad-f8b4-4733-9526-eff1d214ebd8`::moving from state preparing ->
state finished
Thread-88737::DEBUG::2012-09-21
10:36:06,892::resourceManager::809::ResourceManager.Owner::(releaseAll)
Owner.releaseAll requests {} resources {}
Thread-88737::DEBUG::2012-09-21
10:36:06,892::resourceManager::844::ResourceManager.Owner::(cancelAll)
Owner.cancelAll requests {}
Thread-88737::DEBUG::2012-09-21
10:36:06,893::task::978::TaskManager.Task::(_decref)
Task=`f70222ad-f8b4-4733-9526-eff1d214ebd8`::ref 0 aborting False
Do you know why I can not connect via NFS? Using an older kernel not 3.5
and iptables are off.
Dominic
On Mon, Sep 10, 2012 at 12:20 PM, Haim Ateya <hateya at redhat.com> wrote:
> On 09/10/2012 06:27 PM, Dominic Kaiser wrote:
>
>> Here is the message and the logs again except zipped I failed the first
>> delivery:
>>
>> Ok here are the logs 4 node and 1 engine log. Tried making /data folder
>> owned by root and then tried by 36:36 neither worked. Name of volume is
>> data to match folders on nodes also.
>>
>> Let me know what you think,
>>
>> Dominic
>>
>
> this is the actual failure (taken from gfs2vdsm.log).
>
> Thread-332442::DEBUG::2012-09-**10 10:28:05,788::BindingXMLRPC::**859::vds::(wrapper)
> client [10.3.0.241]::call volumeCreate with ('data', ['10.4.0.97:/data',
> '10.4.0.98:/data', '10.4.0.99:/data', '10.4.0.100:/data'],
> 2, 0, ['TCP']) {} flowID [406f2c8e]
> MainProcess|Thread-332442::**DEBUG::2012-09-10
> 10:28:05,792::__init__::1249::**Storage.Misc.excCmd::(_log)
> '/usr/sbin/gluster --mode=script volume create data replica 2 transport TCP
> 10.4.0.97:/data 10.4.0.98:/data 10
> .4.0.99:/data 10.4.0.100:/data' (cwd None)
> MainProcess|Thread-332442::**DEBUG::2012-09-10
> 10:28:05,900::__init__::1249::**Storage.Misc.excCmd::(_log) FAILED: <err>
> = 'Host 10.4.0.99 not a friend\n'; <rc> = 255
> MainProcess|Thread-332442::**ERROR::2012-09-10
> 10:28:05,900::supervdsmServer:**:76::SuperVdsm.ServerCallback:**:(wrapper)
> Error in wrapper
> Traceback (most recent call last):
> File "/usr/share/vdsm/**supervdsmServer.py", line 74, in wrapper
> return func(*args, **kwargs)
> File "/usr/share/vdsm/**supervdsmServer.py", line 286, in wrapper
> return func(*args, **kwargs)
> File "/usr/share/vdsm/gluster/cli.**py", line 46, in wrapper
> return func(*args, **kwargs)
> File "/usr/share/vdsm/gluster/cli.**py", line 176, in volumeCreate
> raise ge.**GlusterVolumeCreateFailedExcep**tion(rc, out, err)
> GlusterVolumeCreateFailedExcep**tion: Volume create failed
> error: Host 10.4.0.99 not a friend
> return code: 255
> Thread-332442::ERROR::2012-09-**10 10:28:05,901::BindingXMLRPC::**877::vds::(wrapper)
> unexpected error
> Traceback (most recent call last):
> File "/usr/share/vdsm/**BindingXMLRPC.py", line 864, in wrapper
> res = f(*args, **kwargs)
> File "/usr/share/vdsm/gluster/api.**py", line 32, in wrapper
> rv = func(*args, **kwargs)
> File "/usr/share/vdsm/gluster/api.**py", line 87, in volumeCreate
> transportList)
> File "/usr/share/vdsm/supervdsm.py"**, line 67, in __call__
> return callMethod()
> File "/usr/share/vdsm/supervdsm.py"**, line 65, in <lambda>
> **kwargs)
> File "<string>", line 2, in glusterVolumeCreate
> File "/usr/lib64/python2.7/**multiprocessing/managers.py", line 759, in
> _callmethod
> kind, result = conn.recv()
> TypeError: ('__init__() takes exactly 4 arguments (1 given)', <class
> 'gluster.exception.**GlusterVolumeCreateFailedExcep**tion'>, ())
>
> can you please run gluster peer status on all your nodes ? also, it
> appears that '10.4.0.99' is problematic, can you try create the volume
> without it ?
>
>
>
>> On Mon, Sep 10, 2012 at 11:24 AM, Dominic Kaiser <
>> dominic at bostonvineyard.org <mailto:dominic@**bostonvineyard.org<dominic at bostonvineyard.org>>>
>> wrote:
>>
>> Here are the other two logs forgot them.
>>
>> dk
>>
>>
>> On Mon, Sep 10, 2012 at 11:19 AM, Dominic Kaiser
>> <dominic at bostonvineyard.org <mailto:dominic@**bostonvineyard.org<dominic at bostonvineyard.org>
>> >>
>>
>> wrote:
>>
>> Ok here are the logs 4 node and 1 engine log. Tried making
>> /data folder owned by root and then tried by 36:36 neither
>> worked. Name of volume is data to match folders on nodes also.
>>
>> Let me know what you think,
>>
>> Dominic
>>
>>
>> On Thu, Sep 6, 2012 at 8:33 AM, Maxim Burgerhout
>> <maxim at wzzrd.com <mailto:maxim at wzzrd.com>> wrote:
>>
>> I just ran into this as well, and it seems that you have
>> to either reformat previously used gluster bricks or
>> manually tweak some extended attributes.
>>
>> Maybe this helps you in setting up your gluster volume,
>> Dominic?
>>
>> More info here:
>> https://bugzilla.redhat.com/**show_bug.cgi?id=812214<https://bugzilla.redhat.com/show_bug.cgi?id=812214>
>>
>>
>> Maxim Burgerhout
>> maxim at wzzrd.com <mailto:maxim at wzzrd.com>
>>
>> ----------------
>> EB11 5E56 E648 9D99 E8EF 05FB C513 6FD4 1302 B48A
>>
>>
>>
>>
>>
>> On Thu, Sep 6, 2012 at 7:50 AM, Shireesh Anjal
>> <sanjal at redhat.com <mailto:sanjal at redhat.com>> wrote:
>>
>> Hi Dominic,
>>
>> Looking at the engine log immediately after trying to
>> create the volume should tell us on which node the
>> gluster volume creation was attempted. Then looking at
>> the vdsm log on that node should help us identifying
>> the exact reason for failure.
>>
>> In case this doesn't help you, can you please simulate
>> the issue again and send back all the 5 log files?
>> (engine.log from engine server and vdsm.log from the 4
>> nodes)
>>
>> Regards,
>> Shireesh
>>
>>
>> On Wednesday 05 September 2012 11:50 PM, Dominic
>> Kaiser wrote:
>>
>>> So I have a problem creating glusterfs volumes. Here
>>> is the install:
>>>
>>> 1. Ovirt 3.1
>>> 2. 4 Nodes are Fedora 17 with kernel 3.3.4 -
>>> 5.fc17.x86_64
>>> 3. 4 nodes peer joined and running
>>> 4. 4 nodes added as hosts to ovirt
>>> 5. created a directory on each node this path /data
>>> 6. chmod 36.36 -R /data all nodes
>>> 7. went back to ovirt and created a
>>>
>>> distributed/replicated volume and added the 4
>>> nodes with brick path of /data
>>>
>>> I received this error:
>>>
>>> Creation of Gluster Volume maingfs1 failed.
>>>
>>> I went and looked at the vdsm logs on the nodes and
>>> the ovirt server which did not say much. Where else
>>> should I look? Also this error is vague what does it
>>> mean?
>>>
>>>
>>> -- Dominic Kaiser
>>> Greater Boston Vineyard
>>> Director of Operations
>>>
>>> cell: 617-230-1412 <tel:617-230-1412>
>>> fax: 617-252-0238 <tel:617-252-0238>
>>> email: dominic at bostonvineyard.org
>>> <mailto:dominic@**bostonvineyard.org<dominic at bostonvineyard.org>
>>> >
>>>
>>>
>>>
>>>
>>> ______________________________**_________________
>>> Users mailing list
>>> Users at ovirt.org <mailto:Users at ovirt.org>
>>> http://lists.ovirt.org/**mailman/listinfo/users<http://lists.ovirt.org/mailman/listinfo/users>
>>>
>>
>>
>> ______________________________**_________________
>> Users mailing list
>> Users at ovirt.org <mailto:Users at ovirt.org>
>>
>> http://lists.ovirt.org/**mailman/listinfo/users<http://lists.ovirt.org/mailman/listinfo/users>
>>
>>
>>
>>
>>
>> -- Dominic Kaiser
>> Greater Boston Vineyard
>> Director of Operations
>>
>> cell: 617-230-1412 <tel:617-230-1412>
>> fax: 617-252-0238 <tel:617-252-0238>
>> email: dominic at bostonvineyard.org
>> <mailto:dominic@**bostonvineyard.org <dominic at bostonvineyard.org>
>> >
>>
>>
>>
>>
>>
>>
>> -- Dominic Kaiser
>> Greater Boston Vineyard
>> Director of Operations
>>
>> cell: 617-230-1412 <tel:617-230-1412>
>> fax: 617-252-0238 <tel:617-252-0238>
>> email: dominic at bostonvineyard.org <mailto:dominic@**
>> bostonvineyard.org <dominic at bostonvineyard.org>>
>>
>>
>>
>>
>>
>>
>> --
>> Dominic Kaiser
>> Greater Boston Vineyard
>> Director of Operations
>>
>> cell: 617-230-1412
>> fax: 617-252-0238
>> email: dominic at bostonvineyard.org <mailto:dominic@**bostonvineyard.org<dominic at bostonvineyard.org>
>> >
>>
>>
>>
>>
>>
>> ______________________________**_________________
>> Users mailing list
>> Users at ovirt.org
>> http://lists.ovirt.org/**mailman/listinfo/users<http://lists.ovirt.org/mailman/listinfo/users>
>>
>
>
--
Dominic Kaiser
Greater Boston Vineyard
Director of Operations
cell: 617-230-1412
fax: 617-252-0238
email: dominic at bostonvineyard.org
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.ovirt.org/pipermail/users/attachments/20120921/8e298a58/attachment-0001.html>
More information about the Users
mailing list