[ovirt-users] Gluster command [<UNKNOWN>] failed on server
Ramesh Nachimuthu
rnachimu at redhat.com
Thu Sep 3 11:55:31 UTC 2015
Can u post the output of 'gluster peer status' on the gluster node?
Regards,
Ramesh
On 09/03/2015 05:10 PM, suporte at logicworks.pt wrote:
> Hi,
>
> I just installed Version 3.5.3.1-1.el7.centos, on centos 7.1, no HE.
>
> for storage, I have only one server with glusterfs:
> glusterfs-fuse-3.7.3-1.el7.x86_64
> glusterfs-server-3.7.3-1.el7.x86_64
> glusterfs-libs-3.7.3-1.el7.x86_64
> glusterfs-client-xlators-3.7.3-1.el7.x86_64
> glusterfs-api-3.7.3-1.el7.x86_64
> glusterfs-3.7.3-1.el7.x86_64
> glusterfs-cli-3.7.3-1.el7.x86_64
>
> # service glusterd status
> Redirecting to /bin/systemctl status glusterd.service
> glusterd.service - GlusterFS, a clustered file-system server
> Loaded: loaded (/usr/lib/systemd/system/glusterd.service; enabled)
> Active: active (running) since Thu 2015-09-03 11
> <callto:2015-09-03%2011>:23:32 WEST; 10min ago
> Process: 1153 ExecStart=/usr/sbin/glusterd -p /var/run/glusterd.pid
> (code=exited, status=0/SUCCESS)
> Main PID: 1387 (glusterd)
> CGroup: /system.slice/glusterd.service
> ââ1387 /usr/sbin/glusterd -p /var/run/glusterd.pid
> ââ2314 /usr/sbin/glusterfsd -s gfs3.acloud.pt --volfile-id
> gv0.gfs...
>
> Sep 03 11:23:31 gfs3.domain.pt systemd[1]: Starting GlusterFS, a
> clustered f....
> Sep 03 11:23:32 gfs3.domain.pt systemd[1]: Started GlusterFS, a
> clustered fi....
> Hint: Some lines were ellipsized, use -l to show in full.
>
>
> Everything was running until I need to restart the node (host), after
> that I was not ables to make the host active. This is the error message:
> Gluster command [<UNKNOWN>] failed on server
>
>
> I also disable JSON protocol, but no success
>
> vdsm.log:
> Thread-14::DEBUG::2015-09-03 11
> <callto:2015-09-03%2011>:37:23,131::BindingXMLRPC::1133::vds::(wrapper) client
> [192.168.6.200 <callto:192.168.6.200>]::call getHardwareInfo with () {}
> Thread-14::DEBUG::2015-09-03 11
> <callto:2015-09-03%2011>:37:23,132::BindingXMLRPC::1140::vds::(wrapper) return
> getHardwareInfo with {'status': {'message': 'Done', 'code': 0},
> 'info': {'systemProductName': 'PRIMERGY RX2520 M1',
> 'systemSerialNumber': 'YLSK005705', 'systemFamily': 'SERVER',
> 'systemVersion': 'GS01', 'systemUUID':
> '4600EA20-2BFF-B34F-B607-DBF9F6B278CE', 'systemManufacturer': 'FUJITSU'}}
> Thread-14::DEBUG::2015-09-03 11
> <callto:2015-09-03%2011>:37:23,266::BindingXMLRPC::1133::vds::(wrapper) client
> [192.168.6.200 <callto:192.168.6.200>]::call hostsList with () {}
> flowID [4acc5233]
> Thread-14::ERROR::2015-09-03 11
> <callto:2015-09-03%2011>:37:23,279::BindingXMLRPC::1149::vds::(wrapper) vdsm
> exception occured
> Traceback (most recent call last):
> File "/usr/share/vdsm/rpc/BindingXMLRPC.py", line 1136, in wrapper
> res = f(*args, **kwargs)
> File "/usr/share/vdsm/gluster/api.py", line 54, in wrapper
> rv = func(*args, **kwargs)
> File "/usr/share/vdsm/gluster/api.py", line 251, in hostsList
> return {'hosts': self.svdsmProxy.glusterPeerStatus()}
> File "/usr/share/vdsm/supervdsm.py", line 50, in __call__
> return callMethod()
> File "/usr/share/vdsm/supervdsm.py", line 48, in <lambda>
> **kwargs)
> File "<string>", line 2, in glusterPeerStatus
> File "/usr/lib64/python2.7/multiprocessing/managers.py", line 773,
> in _callmethod
> raise convert_to_error(kind, result)
> GlusterCmdExecFailedException: Command execution failed
> error: Connection failed. Please check if gluster daemon is operational.
> return code: 1
>
>
> supervdsm.log:
> MainProcess|Thread-14::DEBUG::2015-09-03 11
> <callto:2015-09-03%2011>:37:23,131::supervdsmServer::102::SuperVdsm.ServerCallback::(wrapper)
> call getHardwareInfo with () {}
> MainProcess|Thread-14::DEBUG::2015-09-03 11
> <callto:2015-09-03%2011>:37:23,132::supervdsmServer::109::SuperVdsm.ServerCallback::(wrapper)
> return getHardwareInfo with {'systemProductName': 'PRIMERGY RX2520
> M1', 'systemSerialNumber': 'YLSK005705', 'systemFamily': 'SERVER',
> 'systemVersion': 'GS01', 'systemUUID':
> '4600EA20-2BFF-B34F-B607-DBF9F6B278CE', 'systemManufacturer': 'FUJITSU'}
> MainProcess|Thread-14::DEBUG::2015-09-03 11
> <callto:2015-09-03%2011>:37:23,266::supervdsmServer::102::SuperVdsm.ServerCallback::(wrapper)
> call wrapper with () {}
> MainProcess|Thread-14::DEBUG::2015-09-03 11
> <callto:2015-09-03%2011>:37:23,267::utils::739::root::(execCmd)
> /usr/sbin/gluster --mode=script peer status --xml (cwd None)
> MainProcess|Thread-14::DEBUG::2015-09-03 11
> <callto:2015-09-03%2011>:37:23,278::utils::759::root::(execCmd)
> FAILED: <err> = ''; <rc> = 1
> MainProcess|Thread-14::ERROR::2015-09-03 11
> <callto:2015-09-03%2011>:37:23,279::supervdsmServer::106::SuperVdsm.ServerCallback::(wrapper)
> Error in wrapper
> Traceback (most recent call last):
> File "/usr/share/vdsm/supervdsmServer", line 104, in wrapper
> res = func(*args, **kwargs)
> File "/usr/share/vdsm/supervdsmServer", line 414, in wrapper
> return func(*args, **kwargs)
> File "/usr/share/vdsm/gluster/__init__.py", line 31, in wrapper
> return func(*args, **kwargs)
> File "/usr/share/vdsm/gluster/cli.py", line 909, in peerStatus
> xmltree = _execGlusterXml(command)
> File "/usr/share/vdsm/gluster/cli.py", line 90, in _execGlusterXml
> raise ge.GlusterCmdExecFailedException(rc, out, err)
> GlusterCmdExecFailedException: Command execution failed
> error: Connection failed. Please check if gluster daemon is operational.
> return code: 1
>
>
>
> Any idea?
>
> Thanks
>
> José
>
>
> --
> ------------------------------------------------------------------------
> Jose Ferradeira
> http://www.logicworks.pt
>
>
>
> _______________________________________________
> Users mailing list
> Users at ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.ovirt.org/pipermail/users/attachments/20150903/ae1f7168/attachment-0001.html>
More information about the Users
mailing list