[Users] Problem with creating a glusterfs volume
Dominic Kaiser
dominic at bostonvineyard.org
Fri Sep 21 15:51:03 UTC 2012
I noticed something. If I am trying to mount the gluster share from
another computer and do not include mounproto=tcp it times out. vers=3 or
4 does not matter. Could this be why I can not add it from the engine gui?
dk
On Fri, Sep 21, 2012 at 11:12 AM, Dominic Kaiser <dominic at bostonvineyard.org
> wrote:
> Here is the engine.log info:
>
> [root at ovirt ovirt-engine]# tail engine.log
> 2012-09-21 11:10:00,007 INFO
> [org.ovirt.engine.core.bll.AutoRecoveryManager]
> (QuartzScheduler_Worker-49) Autorecovering 0 hosts
> 2012-09-21 11:10:00,007 INFO
> [org.ovirt.engine.core.bll.AutoRecoveryManager]
> (QuartzScheduler_Worker-49) Checking autorecoverable hosts done
> 2012-09-21 11:10:00,008 INFO
> [org.ovirt.engine.core.bll.AutoRecoveryManager]
> (QuartzScheduler_Worker-49) Checking autorecoverable storage domains
> 2012-09-21 11:10:00,009 INFO
> [org.ovirt.engine.core.bll.AutoRecoveryManager]
> (QuartzScheduler_Worker-49) Autorecovering 0 storage domains
> 2012-09-21 11:10:00,010 INFO
> [org.ovirt.engine.core.bll.AutoRecoveryManager]
> (QuartzScheduler_Worker-49) Checking autorecoverable storage domains done
> 2012-09-21 11:10:22,710 ERROR
> [org.ovirt.engine.core.engineencryptutils.EncryptionUtils]
> (QuartzScheduler_Worker-84) Failed to decryptData must not be longer than
> 256 bytes
> 2012-09-21 11:10:22,726 ERROR
> [org.ovirt.engine.core.engineencryptutils.EncryptionUtils]
> (QuartzScheduler_Worker-12) Failed to decryptData must start with zero
> 2012-09-21 11:10:54,519 INFO
> [org.ovirt.engine.core.bll.storage.RemoveStorageServerConnectionCommand]
> (ajp--0.0.0.0-8009-11) [3769be9c] Running command:
> RemoveStorageServerConnectionCommand internal: false. Entities affected :
> ID: aaa00000-0000-0000-0000-123456789aaa Type: System
> 2012-09-21 11:10:54,537 INFO
> [org.ovirt.engine.core.vdsbroker.vdsbroker.DisconnectStorageServerVDSCommand]
> (ajp--0.0.0.0-8009-11) [3769be9c] START,
> DisconnectStorageServerVDSCommand(vdsId =
> 3822e6c0-0295-11e2-86e6-d74ad5358c03, storagePoolId =
> 00000000-0000-0000-0000-000000000000, storageType = NFS, connectionList =
> [{ id: null, connection: gfs1.bostonvineyard.org:/data };]), log id:
> 16dd4a1b
> 2012-09-21 11:10:56,417 INFO
> [org.ovirt.engine.core.vdsbroker.vdsbroker.DisconnectStorageServerVDSCommand]
> (ajp--0.0.0.0-8009-11) [3769be9c] FINISH,
> DisconnectStorageServerVDSCommand, return:
> {00000000-0000-0000-0000-000000000000=477}, log id: 16dd4a1b
>
> Thanks,
>
> dk
>
> On Fri, Sep 21, 2012 at 11:09 AM, Dominic Kaiser <
> dominic at bostonvineyard.org> wrote:
>
>> I can mount to another computer with this command:
>>
>> mount -o mountproto=tcp,vers=3 -t nfs gfs1.bostonvineyard.org:/data
>> /home/administrator/test
>>
>> So volumes work but I get a 500 error timeout when trying to add as a
>> storage domain in ovirt. weird?
>>
>> dk
>>
>> On Fri, Sep 21, 2012 at 10:44 AM, Dominic Kaiser <
>> dominic at bostonvineyard.org> wrote:
>>
>>> Hey All,
>>>
>>> So I finally found the problem. Cheap NIC's. Installed Intel NIC's no
>>> problems creating gluster volumes and distributed replicated ones.
>>> Broadcom and Realtek yuk! So now I am trying to mount the gluster volume
>>> as a nfs mount and am having a problem. It is timing out like it is
>>> blocked by a firewall.
>>>
>>> I am trying to: mount -t nfs gfs1.bostonvineyard.org:/export
>>> /home/administrator/test
>>>
>>> Here is gfs1 tail vdsm.log
>>>
>>> [root at gfs1 vdsm]# tail vdsm.log
>>> Thread-88731::DEBUG::2012-09-21
>>> 10:35:56,566::resourceManager::844::ResourceManager.Owner::(cancelAll)
>>> Owner.cancelAll requests {}
>>> Thread-88731::DEBUG::2012-09-21
>>> 10:35:56,567::task::978::TaskManager.Task::(_decref)
>>> Task=`01b69eed-de59-4e87-8b28-5268b5dcbb50`::ref 0 aborting False
>>> Thread-88737::DEBUG::2012-09-21
>>> 10:36:06,890::task::588::TaskManager.Task::(_updateState)
>>> Task=`f70222ad-f8b4-4733-9526-eff1d214ebd8`::moving from state init ->
>>> state preparing
>>> Thread-88737::INFO::2012-09-21
>>> 10:36:06,891::logUtils::37::dispatcher::(wrapper) Run and protect:
>>> repoStats(options=None)
>>> Thread-88737::INFO::2012-09-21
>>> 10:36:06,891::logUtils::39::dispatcher::(wrapper) Run and protect:
>>> repoStats, Return response: {}
>>> Thread-88737::DEBUG::2012-09-21
>>> 10:36:06,891::task::1172::TaskManager.Task::(prepare)
>>> Task=`f70222ad-f8b4-4733-9526-eff1d214ebd8`::finished: {}
>>> Thread-88737::DEBUG::2012-09-21
>>> 10:36:06,892::task::588::TaskManager.Task::(_updateState)
>>> Task=`f70222ad-f8b4-4733-9526-eff1d214ebd8`::moving from state preparing ->
>>> state finished
>>> Thread-88737::DEBUG::2012-09-21
>>> 10:36:06,892::resourceManager::809::ResourceManager.Owner::(releaseAll)
>>> Owner.releaseAll requests {} resources {}
>>> Thread-88737::DEBUG::2012-09-21
>>> 10:36:06,892::resourceManager::844::ResourceManager.Owner::(cancelAll)
>>> Owner.cancelAll requests {}
>>> Thread-88737::DEBUG::2012-09-21
>>> 10:36:06,893::task::978::TaskManager.Task::(_decref)
>>> Task=`f70222ad-f8b4-4733-9526-eff1d214ebd8`::ref 0 aborting False
>>>
>>> Do you know why I can not connect via NFS? Using an older kernel not
>>> 3.5 and iptables are off.
>>>
>>> Dominic
>>>
>>>
>>> On Mon, Sep 10, 2012 at 12:20 PM, Haim Ateya <hateya at redhat.com> wrote:
>>>
>>>> On 09/10/2012 06:27 PM, Dominic Kaiser wrote:
>>>>
>>>>> Here is the message and the logs again except zipped I failed the
>>>>> first delivery:
>>>>>
>>>>> Ok here are the logs 4 node and 1 engine log. Tried making /data
>>>>> folder owned by root and then tried by 36:36 neither worked. Name of
>>>>> volume is data to match folders on nodes also.
>>>>>
>>>>> Let me know what you think,
>>>>>
>>>>> Dominic
>>>>>
>>>>
>>>> this is the actual failure (taken from gfs2vdsm.log).
>>>>
>>>> Thread-332442::DEBUG::2012-09-**10 10:28:05,788::BindingXMLRPC::**859::vds::(wrapper)
>>>> client [10.3.0.241]::call volumeCreate with ('data', ['10.4.0.97:/data',
>>>> '10.4.0.98:/data', '10.4.0.99:/data', '10.4.0.100:/data'],
>>>> 2, 0, ['TCP']) {} flowID [406f2c8e]
>>>> MainProcess|Thread-332442::**DEBUG::2012-09-10
>>>> 10:28:05,792::__init__::1249::**Storage.Misc.excCmd::(_log)
>>>> '/usr/sbin/gluster --mode=script volume create data replica 2 transport TCP
>>>> 10.4.0.97:/data 10.4.0.98:/data 10
>>>> .4.0.99:/data 10.4.0.100:/data' (cwd None)
>>>> MainProcess|Thread-332442::**DEBUG::2012-09-10
>>>> 10:28:05,900::__init__::1249::**Storage.Misc.excCmd::(_log) FAILED:
>>>> <err> = 'Host 10.4.0.99 not a friend\n'; <rc> = 255
>>>> MainProcess|Thread-332442::**ERROR::2012-09-10
>>>> 10:28:05,900::supervdsmServer:**:76::SuperVdsm.ServerCallback:**:(wrapper)
>>>> Error in wrapper
>>>> Traceback (most recent call last):
>>>> File "/usr/share/vdsm/**supervdsmServer.py", line 74, in wrapper
>>>> return func(*args, **kwargs)
>>>> File "/usr/share/vdsm/**supervdsmServer.py", line 286, in wrapper
>>>> return func(*args, **kwargs)
>>>> File "/usr/share/vdsm/gluster/cli.**py", line 46, in wrapper
>>>> return func(*args, **kwargs)
>>>> File "/usr/share/vdsm/gluster/cli.**py", line 176, in volumeCreate
>>>> raise ge.**GlusterVolumeCreateFailedExcep**tion(rc, out, err)
>>>> GlusterVolumeCreateFailedExcep**tion: Volume create failed
>>>> error: Host 10.4.0.99 not a friend
>>>> return code: 255
>>>> Thread-332442::ERROR::2012-09-**10 10:28:05,901::BindingXMLRPC::**877::vds::(wrapper)
>>>> unexpected error
>>>> Traceback (most recent call last):
>>>> File "/usr/share/vdsm/**BindingXMLRPC.py", line 864, in wrapper
>>>> res = f(*args, **kwargs)
>>>> File "/usr/share/vdsm/gluster/api.**py", line 32, in wrapper
>>>> rv = func(*args, **kwargs)
>>>> File "/usr/share/vdsm/gluster/api.**py", line 87, in volumeCreate
>>>> transportList)
>>>> File "/usr/share/vdsm/supervdsm.py"**, line 67, in __call__
>>>> return callMethod()
>>>> File "/usr/share/vdsm/supervdsm.py"**, line 65, in <lambda>
>>>> **kwargs)
>>>> File "<string>", line 2, in glusterVolumeCreate
>>>> File "/usr/lib64/python2.7/**multiprocessing/managers.py", line 759,
>>>> in _callmethod
>>>> kind, result = conn.recv()
>>>> TypeError: ('__init__() takes exactly 4 arguments (1 given)', <class
>>>> 'gluster.exception.**GlusterVolumeCreateFailedExcep**tion'>, ())
>>>>
>>>> can you please run gluster peer status on all your nodes ? also, it
>>>> appears that '10.4.0.99' is problematic, can you try create the volume
>>>> without it ?
>>>>
>>>>
>>>>
>>>>> On Mon, Sep 10, 2012 at 11:24 AM, Dominic Kaiser <
>>>>> dominic at bostonvineyard.org <mailto:dominic@**bostonvineyard.org<dominic at bostonvineyard.org>>>
>>>>> wrote:
>>>>>
>>>>> Here are the other two logs forgot them.
>>>>>
>>>>> dk
>>>>>
>>>>>
>>>>> On Mon, Sep 10, 2012 at 11:19 AM, Dominic Kaiser
>>>>> <dominic at bostonvineyard.org <mailto:dominic@**bostonvineyard.org<dominic at bostonvineyard.org>
>>>>> >>
>>>>>
>>>>> wrote:
>>>>>
>>>>> Ok here are the logs 4 node and 1 engine log. Tried making
>>>>> /data folder owned by root and then tried by 36:36 neither
>>>>> worked. Name of volume is data to match folders on nodes also.
>>>>>
>>>>> Let me know what you think,
>>>>>
>>>>> Dominic
>>>>>
>>>>>
>>>>> On Thu, Sep 6, 2012 at 8:33 AM, Maxim Burgerhout
>>>>> <maxim at wzzrd.com <mailto:maxim at wzzrd.com>> wrote:
>>>>>
>>>>> I just ran into this as well, and it seems that you have
>>>>> to either reformat previously used gluster bricks or
>>>>> manually tweak some extended attributes.
>>>>>
>>>>> Maybe this helps you in setting up your gluster volume,
>>>>> Dominic?
>>>>>
>>>>> More info here:
>>>>> https://bugzilla.redhat.com/**show_bug.cgi?id=812214<https://bugzilla.redhat.com/show_bug.cgi?id=812214>
>>>>>
>>>>>
>>>>> Maxim Burgerhout
>>>>> maxim at wzzrd.com <mailto:maxim at wzzrd.com>
>>>>>
>>>>> ----------------
>>>>> EB11 5E56 E648 9D99 E8EF 05FB C513 6FD4 1302 B48A
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>> On Thu, Sep 6, 2012 at 7:50 AM, Shireesh Anjal
>>>>> <sanjal at redhat.com <mailto:sanjal at redhat.com>> wrote:
>>>>>
>>>>> Hi Dominic,
>>>>>
>>>>> Looking at the engine log immediately after trying to
>>>>> create the volume should tell us on which node the
>>>>> gluster volume creation was attempted. Then looking at
>>>>> the vdsm log on that node should help us identifying
>>>>> the exact reason for failure.
>>>>>
>>>>> In case this doesn't help you, can you please simulate
>>>>> the issue again and send back all the 5 log files?
>>>>> (engine.log from engine server and vdsm.log from the 4
>>>>> nodes)
>>>>>
>>>>> Regards,
>>>>> Shireesh
>>>>>
>>>>>
>>>>> On Wednesday 05 September 2012 11:50 PM, Dominic
>>>>> Kaiser wrote:
>>>>>
>>>>>> So I have a problem creating glusterfs volumes. Here
>>>>>> is the install:
>>>>>>
>>>>>> 1. Ovirt 3.1
>>>>>> 2. 4 Nodes are Fedora 17 with kernel 3.3.4 -
>>>>>> 5.fc17.x86_64
>>>>>> 3. 4 nodes peer joined and running
>>>>>> 4. 4 nodes added as hosts to ovirt
>>>>>> 5. created a directory on each node this path /data
>>>>>> 6. chmod 36.36 -R /data all nodes
>>>>>> 7. went back to ovirt and created a
>>>>>>
>>>>>> distributed/replicated volume and added the 4
>>>>>> nodes with brick path of /data
>>>>>>
>>>>>> I received this error:
>>>>>>
>>>>>> Creation of Gluster Volume maingfs1 failed.
>>>>>>
>>>>>> I went and looked at the vdsm logs on the nodes and
>>>>>> the ovirt server which did not say much. Where else
>>>>>> should I look? Also this error is vague what does it
>>>>>> mean?
>>>>>>
>>>>>>
>>>>>> -- Dominic Kaiser
>>>>>> Greater Boston Vineyard
>>>>>> Director of Operations
>>>>>>
>>>>>> cell: 617-230-1412 <tel:617-230-1412>
>>>>>> fax: 617-252-0238 <tel:617-252-0238>
>>>>>> email: dominic at bostonvineyard.org
>>>>>> <mailto:dominic@**bostonvineyard.org<dominic at bostonvineyard.org>
>>>>>> >
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>> ______________________________**_________________
>>>>>> Users mailing list
>>>>>> Users at ovirt.org <mailto:Users at ovirt.org>
>>>>>> http://lists.ovirt.org/**mailman/listinfo/users<http://lists.ovirt.org/mailman/listinfo/users>
>>>>>>
>>>>>
>>>>>
>>>>> ______________________________**_________________
>>>>> Users mailing list
>>>>> Users at ovirt.org <mailto:Users at ovirt.org>
>>>>>
>>>>> http://lists.ovirt.org/**mailman/listinfo/users<http://lists.ovirt.org/mailman/listinfo/users>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>> -- Dominic Kaiser
>>>>> Greater Boston Vineyard
>>>>> Director of Operations
>>>>>
>>>>> cell: 617-230-1412 <tel:617-230-1412>
>>>>> fax: 617-252-0238 <tel:617-252-0238>
>>>>> email: dominic at bostonvineyard.org
>>>>> <mailto:dominic@**bostonvineyard.org<dominic at bostonvineyard.org>
>>>>> >
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>> -- Dominic Kaiser
>>>>> Greater Boston Vineyard
>>>>> Director of Operations
>>>>>
>>>>> cell: 617-230-1412 <tel:617-230-1412>
>>>>> fax: 617-252-0238 <tel:617-252-0238>
>>>>> email: dominic at bostonvineyard.org <mailto:dominic@**
>>>>> bostonvineyard.org <dominic at bostonvineyard.org>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>> --
>>>>> Dominic Kaiser
>>>>> Greater Boston Vineyard
>>>>> Director of Operations
>>>>>
>>>>> cell: 617-230-1412
>>>>> fax: 617-252-0238
>>>>> email: dominic at bostonvineyard.org <mailto:dominic@**bostonvineyard.org<dominic at bostonvineyard.org>
>>>>> >
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>> ______________________________**_________________
>>>>> Users mailing list
>>>>> Users at ovirt.org
>>>>> http://lists.ovirt.org/**mailman/listinfo/users<http://lists.ovirt.org/mailman/listinfo/users>
>>>>>
>>>>
>>>>
>>>
>>>
>>> --
>>> Dominic Kaiser
>>> Greater Boston Vineyard
>>> Director of Operations
>>>
>>> cell: 617-230-1412
>>> fax: 617-252-0238
>>> email: dominic at bostonvineyard.org
>>>
>>>
>>>
>>
>>
>> --
>> Dominic Kaiser
>> Greater Boston Vineyard
>> Director of Operations
>>
>> cell: 617-230-1412
>> fax: 617-252-0238
>> email: dominic at bostonvineyard.org
>>
>>
>>
>
>
> --
> Dominic Kaiser
> Greater Boston Vineyard
> Director of Operations
>
> cell: 617-230-1412
> fax: 617-252-0238
> email: dominic at bostonvineyard.org
>
>
>
--
Dominic Kaiser
Greater Boston Vineyard
Director of Operations
cell: 617-230-1412
fax: 617-252-0238
email: dominic at bostonvineyard.org
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.ovirt.org/pipermail/users/attachments/20120921/b5a69fa7/attachment-0001.html>
More information about the Users
mailing list