Yes I can mount both to another computer.  Just not to ovirt.  I noticed on the other computer which is Ubuntu 12.04 if you leave mountproto=tcp out of the command it does not mount.  Does engine default to tcp?

Dk

On Sep 21, 2012 6:36 PM, "Jason Brooks" <jbrooks@redhat.com> wrote:
On 09/21/2012 08:09 AM, Dominic Kaiser wrote:
I can mount to another computer with this command:

mount -o mountproto=tcp,vers=3 -t nfs gfs1.bostonvineyard.org:/data
/home/administrator/test

I notice that in your previous message, citing the mount that didn't work, you were mounting :/export, and above you're mounting :/data. Can you also mount the export volume from another computer?



So volumes work but I get a 500 error timeout when trying to add as a
storage domain in ovirt.  weird?

dk

On Fri, Sep 21, 2012 at 10:44 AM, Dominic Kaiser
<dominic@bostonvineyard.org <mailto:dominic@bostonvineyard.org>> wrote:

    Hey All,

    So I finally found the problem.  Cheap NIC's.  Installed Intel NIC's
    no problems creating gluster volumes and distributed replicated
    ones.  Broadcom and Realtek yuk!  So now I am trying to mount the
    gluster volume as a nfs mount and am having a problem.  It is timing
    out like it is blocked by a firewall.

    I am trying to:  mount -t nfs gfs1.bostonvineyard.org:/export
    /home/administrator/test

    Here is gfs1 tail vdsm.log

    [root@gfs1 vdsm]# tail vdsm.log
    Thread-88731::DEBUG::2012-09-21
    10:35:56,566::resourceManager::844::ResourceManager.Owner::(cancelAll)
    Owner.cancelAll requests {}
    Thread-88731::DEBUG::2012-09-21
    10:35:56,567::task::978::TaskManager.Task::(_decref)
    Task=`01b69eed-de59-4e87-8b28-5268b5dcbb50`::ref 0 aborting False
    Thread-88737::DEBUG::2012-09-21
    10:36:06,890::task::588::TaskManager.Task::(_updateState)
    Task=`f70222ad-f8b4-4733-9526-eff1d214ebd8`::moving from state init
    -> state preparing
    Thread-88737::INFO::2012-09-21
    10:36:06,891::logUtils::37::dispatcher::(wrapper) Run and protect:
    repoStats(options=None)
    Thread-88737::INFO::2012-09-21
    10:36:06,891::logUtils::39::dispatcher::(wrapper) Run and protect:
    repoStats, Return response: {}
    Thread-88737::DEBUG::2012-09-21
    10:36:06,891::task::1172::TaskManager.Task::(prepare)
    Task=`f70222ad-f8b4-4733-9526-eff1d214ebd8`::finished: {}
    Thread-88737::DEBUG::2012-09-21
    10:36:06,892::task::588::TaskManager.Task::(_updateState)
    Task=`f70222ad-f8b4-4733-9526-eff1d214ebd8`::moving from state
    preparing -> state finished
    Thread-88737::DEBUG::2012-09-21
    10:36:06,892::resourceManager::809::ResourceManager.Owner::(releaseAll)
    Owner.releaseAll requests {} resources {}
    Thread-88737::DEBUG::2012-09-21
    10:36:06,892::resourceManager::844::ResourceManager.Owner::(cancelAll)
    Owner.cancelAll requests {}
    Thread-88737::DEBUG::2012-09-21
    10:36:06,893::task::978::TaskManager.Task::(_decref)
    Task=`f70222ad-f8b4-4733-9526-eff1d214ebd8`::ref 0 aborting False

    Do you know why I can not connect via NFS?  Using an older kernel
    not 3.5 and iptables are off.

    Dominic


    On Mon, Sep 10, 2012 at 12:20 PM, Haim Ateya <hateya@redhat.com
    <mailto:hateya@redhat.com>> wrote:

        On 09/10/2012 06:27 PM, Dominic Kaiser wrote:

            Here is the message and the logs again except zipped I
            failed the first delivery:

            Ok here are the logs 4 node and 1 engine log.  Tried making
            /data folder owned by root and then tried by 36:36 neither
            worked.  Name of volume is data to match folders on nodes also.

            Let me know what you think,

            Dominic


        this is the actual failure (taken from gfs2vdsm.log).

        Thread-332442::DEBUG::2012-09-__10
        10:28:05,788::BindingXMLRPC::__859::vds::(wrapper) client
        [10.3.0.241]::call volumeCreate with ('data',
        ['10.4.0.97:/data', '10.4.0.98:/data', '10.4.0.99:/data',
        '10.4.0.100:/data'],
          2, 0, ['TCP']) {} flowID [406f2c8e]
        MainProcess|Thread-332442::__DEBUG::2012-09-10
        10:28:05,792::__init__::1249::__Storage.Misc.excCmd::(_log)
        '/usr/sbin/gluster --mode=script volume create data replica 2
        transport TCP 10.4.0.97:/data 10.4.0.98:/data 10
        .4.0.99:/data 10.4.0.100:/data' (cwd None)
        MainProcess|Thread-332442::__DEBUG::2012-09-10
        10:28:05,900::__init__::1249::__Storage.Misc.excCmd::(_log)
        FAILED: <err> = 'Host 10.4.0.99 not a friend\n'; <rc> = 255
        MainProcess|Thread-332442::__ERROR::2012-09-10
        10:28:05,900::supervdsmServer:__:76::SuperVdsm.ServerCallback:__:(wrapper)
        Error in wrapper
        Traceback (most recent call last):
           File "/usr/share/vdsm/__supervdsmServer.py", line 74, in wrapper
             return func(*args, **kwargs)
           File "/usr/share/vdsm/__supervdsmServer.py", line 286, in wrapper
             return func(*args, **kwargs)
           File "/usr/share/vdsm/gluster/cli.__py", line 46, in wrapper
             return func(*args, **kwargs)
           File "/usr/share/vdsm/gluster/cli.__py", line 176, in
        volumeCreate
             raise ge.__GlusterVolumeCreateFailedExcep__tion(rc, out, err)
        GlusterVolumeCreateFailedExcep__tion: Volume create failed
        error: Host 10.4.0.99 not a friend
        return code: 255
        Thread-332442::ERROR::2012-09-__10
        10:28:05,901::BindingXMLRPC::__877::vds::(wrapper) unexpected error
        Traceback (most recent call last):
           File "/usr/share/vdsm/__BindingXMLRPC.py", line 864, in wrapper
             res = f(*args, **kwargs)
           File "/usr/share/vdsm/gluster/api.__py", line 32, in wrapper
             rv = func(*args, **kwargs)
           File "/usr/share/vdsm/gluster/api.__py", line 87, in volumeCreate
             transportList)
           File "/usr/share/vdsm/supervdsm.py"__, line 67, in __call__
             return callMethod()
           File "/usr/share/vdsm/supervdsm.py"__, line 65, in <lambda>
             **kwargs)
           File "<string>", line 2, in glusterVolumeCreate
           File "/usr/lib64/python2.7/__multiprocessing/managers.py",
        line 759, in _callmethod
             kind, result = conn.recv()
        TypeError: ('__init__() takes exactly 4 arguments (1 given)',
        <class
        'gluster.exception.__GlusterVolumeCreateFailedExcep__tion'>, ())

        can you please run  gluster peer status on all your nodes ?
        also, it appears that '10.4.0.99' is problematic, can you try
        create the volume without it ?



            On Mon, Sep 10, 2012 at 11:24 AM, Dominic Kaiser
            <dominic@bostonvineyard.org
            <mailto:dominic@bostonvineyard.org>
            <mailto:dominic@__bostonvineyard.org
            <mailto:dominic@bostonvineyard.org>>> wrote:

                 Here are the other two logs forgot them.

                 dk


                 On Mon, Sep 10, 2012 at 11:19 AM, Dominic Kaiser
                 <dominic@bostonvineyard.org
            <mailto:dominic@bostonvineyard.org>
            <mailto:dominic@__bostonvineyard.org
            <mailto:dominic@bostonvineyard.org>>>

                 wrote:

                     Ok here are the logs 4 node and 1 engine log.
              Tried making
                     /data folder owned by root and then tried by 36:36
            neither
                     worked.  Name of volume is data to match folders on
            nodes also.

                     Let me know what you think,

                     Dominic


                     On Thu, Sep 6, 2012 at 8:33 AM, Maxim Burgerhout
                     <maxim@wzzrd.com <mailto:maxim@wzzrd.com>
            <mailto:maxim@wzzrd.com <mailto:maxim@wzzrd.com>>> wrote:

                         I just ran into this as well, and it seems that
            you have
                         to either reformat previously used gluster
            bricks or
                         manually tweak some extended attributes.

                         Maybe this helps you in setting up your gluster
            volume,
                         Dominic?

                         More info here:
            https://bugzilla.redhat.com/__show_bug.cgi?id=812214
            <https://bugzilla.redhat.com/show_bug.cgi?id=812214>


                         Maxim Burgerhout
            maxim@wzzrd.com <mailto:maxim@wzzrd.com>
            <mailto:maxim@wzzrd.com <mailto:maxim@wzzrd.com>>

                         ----------------
                         EB11 5E56 E648 9D99 E8EF 05FB C513 6FD4 1302 B48A





                         On Thu, Sep 6, 2012 at 7:50 AM, Shireesh Anjal
                         <sanjal@redhat.com <mailto:sanjal@redhat.com>
            <mailto:sanjal@redhat.com <mailto:sanjal@redhat.com>>> wrote:

                             Hi Dominic,

                             Looking at the engine log immediately after
            trying to
                             create the volume should tell us on which
            node the
                             gluster volume creation was attempted. Then
            looking at
                             the vdsm log on that node should help us
            identifying
                             the exact reason for failure.

                             In case this doesn't help you, can you
            please simulate
                             the issue again and send back all the 5 log
            files?
                             (engine.log from engine server and vdsm.log
            from the 4
                             nodes)

                             Regards,
                             Shireesh


                             On Wednesday 05 September 2012 11:50 PM,
            Dominic
                             Kaiser wrote:

                                 So I have a problem creating glusterfs
                volumes.  Here
                                 is the install:

                                  1. Ovirt 3.1
                                  2. 4 Nodes are Fedora 17 with kernel
                3.3.4 -
                                     5.fc17.x86_64
                                  3. 4 nodes peer joined and running
                                  4. 4 nodes added as hosts to ovirt
                                  5. created a directory on each node
                this path /data
                                  6. chmod 36.36 -R /data all nodes
                                  7. went back to ovirt and created a

                                     distributed/replicated volume and
                added the 4
                                     nodes with brick path of /data

                                 I received this error:

                                 Creation of Gluster Volume maingfs1 failed.

                                 I went and looked at the vdsm logs on
                the nodes and
                                 the ovirt server which did not say
                much.  Where else
                                 should I look?  Also this error is
                vague what does it
                                 mean?


                                 --                 Dominic Kaiser
                                 Greater Boston Vineyard
                                 Director of Operations

                                 cell: 617-230-1412 <tel:617-230-1412>
                <tel:617-230-1412 <tel:617-230-1412>>
                                 fax: 617-252-0238 <tel:617-252-0238>
                <tel:617-252-0238 <tel:617-252-0238>>
                                 email: dominic@bostonvineyard.org
                <mailto:dominic@bostonvineyard.org>
                                 <mailto:dominic@__bostonvineyard.org
                <mailto:dominic@bostonvineyard.org>>





                _________________________________________________
                                 Users mailing list
                Users@ovirt.org <mailto:Users@ovirt.org>
                  <mailto:Users@ovirt.org <mailto:Users@ovirt.org>>
                http://lists.ovirt.org/__mailman/listinfo/users
                <http://lists.ovirt.org/mailman/listinfo/users>




            _________________________________________________
                             Users mailing list
            Users@ovirt.org <mailto:Users@ovirt.org>
            <mailto:Users@ovirt.org <mailto:Users@ovirt.org>>

            http://lists.ovirt.org/__mailman/listinfo/users
            <http://lists.ovirt.org/mailman/listinfo/users>





                     --         Dominic Kaiser
                     Greater Boston Vineyard
                     Director of Operations

                     cell: 617-230-1412 <tel:617-230-1412>
            <tel:617-230-1412 <tel:617-230-1412>>
                     fax: 617-252-0238 <tel:617-252-0238>
            <tel:617-252-0238 <tel:617-252-0238>>
                     email: dominic@bostonvineyard.org
            <mailto:dominic@bostonvineyard.org>
                     <mailto:dominic@__bostonvineyard.org
            <mailto:dominic@bostonvineyard.org>>






                 --     Dominic Kaiser
                 Greater Boston Vineyard
                 Director of Operations

                 cell: 617-230-1412 <tel:617-230-1412> <tel:617-230-1412
            <tel:617-230-1412>>
                 fax: 617-252-0238 <tel:617-252-0238> <tel:617-252-0238
            <tel:617-252-0238>>
                 email: dominic@bostonvineyard.org
            <mailto:dominic@bostonvineyard.org>
            <mailto:dominic@__bostonvineyard.org
            <mailto:dominic@bostonvineyard.org>>






            --
            Dominic Kaiser
            Greater Boston Vineyard
            Director of Operations

            cell: 617-230-1412 <tel:617-230-1412>
            fax: 617-252-0238 <tel:617-252-0238>
            email: dominic@bostonvineyard.org
            <mailto:dominic@bostonvineyard.org>
            <mailto:dominic@__bostonvineyard.org
            <mailto:dominic@bostonvineyard.org>>





            _________________________________________________
            Users mailing list
            Users@ovirt.org <mailto:Users@ovirt.org>
            http://lists.ovirt.org/__mailman/listinfo/users
            <http://lists.ovirt.org/mailman/listinfo/users>





    --
    Dominic Kaiser
    Greater Boston Vineyard
    Director of Operations

    cell: 617-230-1412 <tel:617-230-1412>
    fax: 617-252-0238 <tel:617-252-0238>
    email: dominic@bostonvineyard.org <mailto:dominic@bostonvineyard.org>





--
Dominic Kaiser
Greater Boston Vineyard
Director of Operations

cell: 617-230-1412
fax: 617-252-0238
email: dominic@bostonvineyard.org <mailto:dominic@bostonvineyard.org>




_______________________________________________
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users



--

@jasonbrooks