[Users] Problem with creating a glusterfs volume
Jason Brooks
jbrooks at redhat.com
Fri Sep 21 22:36:14 UTC 2012
On 09/21/2012 08:09 AM, Dominic Kaiser wrote:
> I can mount to another computer with this command:
>
> mount -o mountproto=tcp,vers=3 -t nfs gfs1.bostonvineyard.org:/data
> /home/administrator/test
I notice that in your previous message, citing the mount that didn't
work, you were mounting :/export, and above you're mounting :/data. Can
you also mount the export volume from another computer?
>
> So volumes work but I get a 500 error timeout when trying to add as a
> storage domain in ovirt. weird?
>
> dk
>
> On Fri, Sep 21, 2012 at 10:44 AM, Dominic Kaiser
> <dominic at bostonvineyard.org <mailto:dominic at bostonvineyard.org>> wrote:
>
> Hey All,
>
> So I finally found the problem. Cheap NIC's. Installed Intel NIC's
> no problems creating gluster volumes and distributed replicated
> ones. Broadcom and Realtek yuk! So now I am trying to mount the
> gluster volume as a nfs mount and am having a problem. It is timing
> out like it is blocked by a firewall.
>
> I am trying to: mount -t nfs gfs1.bostonvineyard.org:/export
> /home/administrator/test
>
> Here is gfs1 tail vdsm.log
>
> [root at gfs1 vdsm]# tail vdsm.log
> Thread-88731::DEBUG::2012-09-21
> 10:35:56,566::resourceManager::844::ResourceManager.Owner::(cancelAll)
> Owner.cancelAll requests {}
> Thread-88731::DEBUG::2012-09-21
> 10:35:56,567::task::978::TaskManager.Task::(_decref)
> Task=`01b69eed-de59-4e87-8b28-5268b5dcbb50`::ref 0 aborting False
> Thread-88737::DEBUG::2012-09-21
> 10:36:06,890::task::588::TaskManager.Task::(_updateState)
> Task=`f70222ad-f8b4-4733-9526-eff1d214ebd8`::moving from state init
> -> state preparing
> Thread-88737::INFO::2012-09-21
> 10:36:06,891::logUtils::37::dispatcher::(wrapper) Run and protect:
> repoStats(options=None)
> Thread-88737::INFO::2012-09-21
> 10:36:06,891::logUtils::39::dispatcher::(wrapper) Run and protect:
> repoStats, Return response: {}
> Thread-88737::DEBUG::2012-09-21
> 10:36:06,891::task::1172::TaskManager.Task::(prepare)
> Task=`f70222ad-f8b4-4733-9526-eff1d214ebd8`::finished: {}
> Thread-88737::DEBUG::2012-09-21
> 10:36:06,892::task::588::TaskManager.Task::(_updateState)
> Task=`f70222ad-f8b4-4733-9526-eff1d214ebd8`::moving from state
> preparing -> state finished
> Thread-88737::DEBUG::2012-09-21
> 10:36:06,892::resourceManager::809::ResourceManager.Owner::(releaseAll)
> Owner.releaseAll requests {} resources {}
> Thread-88737::DEBUG::2012-09-21
> 10:36:06,892::resourceManager::844::ResourceManager.Owner::(cancelAll)
> Owner.cancelAll requests {}
> Thread-88737::DEBUG::2012-09-21
> 10:36:06,893::task::978::TaskManager.Task::(_decref)
> Task=`f70222ad-f8b4-4733-9526-eff1d214ebd8`::ref 0 aborting False
>
> Do you know why I can not connect via NFS? Using an older kernel
> not 3.5 and iptables are off.
>
> Dominic
>
>
> On Mon, Sep 10, 2012 at 12:20 PM, Haim Ateya <hateya at redhat.com
> <mailto:hateya at redhat.com>> wrote:
>
> On 09/10/2012 06:27 PM, Dominic Kaiser wrote:
>
> Here is the message and the logs again except zipped I
> failed the first delivery:
>
> Ok here are the logs 4 node and 1 engine log. Tried making
> /data folder owned by root and then tried by 36:36 neither
> worked. Name of volume is data to match folders on nodes also.
>
> Let me know what you think,
>
> Dominic
>
>
> this is the actual failure (taken from gfs2vdsm.log).
>
> Thread-332442::DEBUG::2012-09-__10
> 10:28:05,788::BindingXMLRPC::__859::vds::(wrapper) client
> [10.3.0.241]::call volumeCreate with ('data',
> ['10.4.0.97:/data', '10.4.0.98:/data', '10.4.0.99:/data',
> '10.4.0.100:/data'],
> 2, 0, ['TCP']) {} flowID [406f2c8e]
> MainProcess|Thread-332442::__DEBUG::2012-09-10
> 10:28:05,792::__init__::1249::__Storage.Misc.excCmd::(_log)
> '/usr/sbin/gluster --mode=script volume create data replica 2
> transport TCP 10.4.0.97:/data 10.4.0.98:/data 10
> .4.0.99:/data 10.4.0.100:/data' (cwd None)
> MainProcess|Thread-332442::__DEBUG::2012-09-10
> 10:28:05,900::__init__::1249::__Storage.Misc.excCmd::(_log)
> FAILED: <err> = 'Host 10.4.0.99 not a friend\n'; <rc> = 255
> MainProcess|Thread-332442::__ERROR::2012-09-10
> 10:28:05,900::supervdsmServer:__:76::SuperVdsm.ServerCallback:__:(wrapper)
> Error in wrapper
> Traceback (most recent call last):
> File "/usr/share/vdsm/__supervdsmServer.py", line 74, in wrapper
> return func(*args, **kwargs)
> File "/usr/share/vdsm/__supervdsmServer.py", line 286, in wrapper
> return func(*args, **kwargs)
> File "/usr/share/vdsm/gluster/cli.__py", line 46, in wrapper
> return func(*args, **kwargs)
> File "/usr/share/vdsm/gluster/cli.__py", line 176, in
> volumeCreate
> raise ge.__GlusterVolumeCreateFailedExcep__tion(rc, out, err)
> GlusterVolumeCreateFailedExcep__tion: Volume create failed
> error: Host 10.4.0.99 not a friend
> return code: 255
> Thread-332442::ERROR::2012-09-__10
> 10:28:05,901::BindingXMLRPC::__877::vds::(wrapper) unexpected error
> Traceback (most recent call last):
> File "/usr/share/vdsm/__BindingXMLRPC.py", line 864, in wrapper
> res = f(*args, **kwargs)
> File "/usr/share/vdsm/gluster/api.__py", line 32, in wrapper
> rv = func(*args, **kwargs)
> File "/usr/share/vdsm/gluster/api.__py", line 87, in volumeCreate
> transportList)
> File "/usr/share/vdsm/supervdsm.py"__, line 67, in __call__
> return callMethod()
> File "/usr/share/vdsm/supervdsm.py"__, line 65, in <lambda>
> **kwargs)
> File "<string>", line 2, in glusterVolumeCreate
> File "/usr/lib64/python2.7/__multiprocessing/managers.py",
> line 759, in _callmethod
> kind, result = conn.recv()
> TypeError: ('__init__() takes exactly 4 arguments (1 given)',
> <class
> 'gluster.exception.__GlusterVolumeCreateFailedExcep__tion'>, ())
>
> can you please run gluster peer status on all your nodes ?
> also, it appears that '10.4.0.99' is problematic, can you try
> create the volume without it ?
>
>
>
> On Mon, Sep 10, 2012 at 11:24 AM, Dominic Kaiser
> <dominic at bostonvineyard.org
> <mailto:dominic at bostonvineyard.org>
> <mailto:dominic at __bostonvineyard.org
> <mailto:dominic at bostonvineyard.org>>> wrote:
>
> Here are the other two logs forgot them.
>
> dk
>
>
> On Mon, Sep 10, 2012 at 11:19 AM, Dominic Kaiser
> <dominic at bostonvineyard.org
> <mailto:dominic at bostonvineyard.org>
> <mailto:dominic at __bostonvineyard.org
> <mailto:dominic at bostonvineyard.org>>>
>
> wrote:
>
> Ok here are the logs 4 node and 1 engine log.
> Tried making
> /data folder owned by root and then tried by 36:36
> neither
> worked. Name of volume is data to match folders on
> nodes also.
>
> Let me know what you think,
>
> Dominic
>
>
> On Thu, Sep 6, 2012 at 8:33 AM, Maxim Burgerhout
> <maxim at wzzrd.com <mailto:maxim at wzzrd.com>
> <mailto:maxim at wzzrd.com <mailto:maxim at wzzrd.com>>> wrote:
>
> I just ran into this as well, and it seems that
> you have
> to either reformat previously used gluster
> bricks or
> manually tweak some extended attributes.
>
> Maybe this helps you in setting up your gluster
> volume,
> Dominic?
>
> More info here:
> https://bugzilla.redhat.com/__show_bug.cgi?id=812214
> <https://bugzilla.redhat.com/show_bug.cgi?id=812214>
>
>
> Maxim Burgerhout
> maxim at wzzrd.com <mailto:maxim at wzzrd.com>
> <mailto:maxim at wzzrd.com <mailto:maxim at wzzrd.com>>
>
> ----------------
> EB11 5E56 E648 9D99 E8EF 05FB C513 6FD4 1302 B48A
>
>
>
>
>
> On Thu, Sep 6, 2012 at 7:50 AM, Shireesh Anjal
> <sanjal at redhat.com <mailto:sanjal at redhat.com>
> <mailto:sanjal at redhat.com <mailto:sanjal at redhat.com>>> wrote:
>
> Hi Dominic,
>
> Looking at the engine log immediately after
> trying to
> create the volume should tell us on which
> node the
> gluster volume creation was attempted. Then
> looking at
> the vdsm log on that node should help us
> identifying
> the exact reason for failure.
>
> In case this doesn't help you, can you
> please simulate
> the issue again and send back all the 5 log
> files?
> (engine.log from engine server and vdsm.log
> from the 4
> nodes)
>
> Regards,
> Shireesh
>
>
> On Wednesday 05 September 2012 11:50 PM,
> Dominic
> Kaiser wrote:
>
> So I have a problem creating glusterfs
> volumes. Here
> is the install:
>
> 1. Ovirt 3.1
> 2. 4 Nodes are Fedora 17 with kernel
> 3.3.4 -
> 5.fc17.x86_64
> 3. 4 nodes peer joined and running
> 4. 4 nodes added as hosts to ovirt
> 5. created a directory on each node
> this path /data
> 6. chmod 36.36 -R /data all nodes
> 7. went back to ovirt and created a
>
> distributed/replicated volume and
> added the 4
> nodes with brick path of /data
>
> I received this error:
>
> Creation of Gluster Volume maingfs1 failed.
>
> I went and looked at the vdsm logs on
> the nodes and
> the ovirt server which did not say
> much. Where else
> should I look? Also this error is
> vague what does it
> mean?
>
>
> -- Dominic Kaiser
> Greater Boston Vineyard
> Director of Operations
>
> cell: 617-230-1412 <tel:617-230-1412>
> <tel:617-230-1412 <tel:617-230-1412>>
> fax: 617-252-0238 <tel:617-252-0238>
> <tel:617-252-0238 <tel:617-252-0238>>
> email: dominic at bostonvineyard.org
> <mailto:dominic at bostonvineyard.org>
> <mailto:dominic at __bostonvineyard.org
> <mailto:dominic at bostonvineyard.org>>
>
>
>
>
>
> _________________________________________________
> Users mailing list
> Users at ovirt.org <mailto:Users at ovirt.org>
> <mailto:Users at ovirt.org <mailto:Users at ovirt.org>>
> http://lists.ovirt.org/__mailman/listinfo/users
> <http://lists.ovirt.org/mailman/listinfo/users>
>
>
>
>
> _________________________________________________
> Users mailing list
> Users at ovirt.org <mailto:Users at ovirt.org>
> <mailto:Users at ovirt.org <mailto:Users at ovirt.org>>
>
> http://lists.ovirt.org/__mailman/listinfo/users
> <http://lists.ovirt.org/mailman/listinfo/users>
>
>
>
>
>
> -- Dominic Kaiser
> Greater Boston Vineyard
> Director of Operations
>
> cell: 617-230-1412 <tel:617-230-1412>
> <tel:617-230-1412 <tel:617-230-1412>>
> fax: 617-252-0238 <tel:617-252-0238>
> <tel:617-252-0238 <tel:617-252-0238>>
> email: dominic at bostonvineyard.org
> <mailto:dominic at bostonvineyard.org>
> <mailto:dominic at __bostonvineyard.org
> <mailto:dominic at bostonvineyard.org>>
>
>
>
>
>
>
> -- Dominic Kaiser
> Greater Boston Vineyard
> Director of Operations
>
> cell: 617-230-1412 <tel:617-230-1412> <tel:617-230-1412
> <tel:617-230-1412>>
> fax: 617-252-0238 <tel:617-252-0238> <tel:617-252-0238
> <tel:617-252-0238>>
> email: dominic at bostonvineyard.org
> <mailto:dominic at bostonvineyard.org>
> <mailto:dominic at __bostonvineyard.org
> <mailto:dominic at bostonvineyard.org>>
>
>
>
>
>
>
> --
> Dominic Kaiser
> Greater Boston Vineyard
> Director of Operations
>
> cell: 617-230-1412 <tel:617-230-1412>
> fax: 617-252-0238 <tel:617-252-0238>
> email: dominic at bostonvineyard.org
> <mailto:dominic at bostonvineyard.org>
> <mailto:dominic at __bostonvineyard.org
> <mailto:dominic at bostonvineyard.org>>
>
>
>
>
>
> _________________________________________________
> Users mailing list
> Users at ovirt.org <mailto:Users at ovirt.org>
> http://lists.ovirt.org/__mailman/listinfo/users
> <http://lists.ovirt.org/mailman/listinfo/users>
>
>
>
>
>
> --
> Dominic Kaiser
> Greater Boston Vineyard
> Director of Operations
>
> cell: 617-230-1412 <tel:617-230-1412>
> fax: 617-252-0238 <tel:617-252-0238>
> email: dominic at bostonvineyard.org <mailto:dominic at bostonvineyard.org>
>
>
>
>
>
> --
> Dominic Kaiser
> Greater Boston Vineyard
> Director of Operations
>
> cell: 617-230-1412
> fax: 617-252-0238
> email: dominic at bostonvineyard.org <mailto:dominic at bostonvineyard.org>
>
>
>
>
> _______________________________________________
> Users mailing list
> Users at ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
--
@jasonbrooks
More information about the Users
mailing list