I post a bug in centos about a problem with
NetworkManager-wait-online.service, this service is executed before
network-online.target
I had a problem with mount drbd in a centos7 at boot time. I think it could
be similar to the problem to mount gluster in boot time in centos7.
I modified the service adding one line that wait to NetworkManager to set
up the ip address in the interface that is used to connect de drbd.
the ip address in eth0 is: 10.88.0.1
# cat /usr/lib/systemd/system/NetworkManager-wait-online.service
[Unit]
Description=Network Manager Wait Online
Requisite=NetworkManager.service
After=NetworkManager.service
Wants=network.target
Before=network.target network-online.target
[Service]
Type=oneshot
ExecStart=/usr/bin/nm-online -q --timeout=30
ExecStart=/bin/sh -c 'until ip a s eth0 | grep -q 10.88; do usleep 100000;
done'
Alberto Larraz
2014-12-09 10:54 GMT+01:00 Punit Dambiwal <hypunit(a)gmail.com>:
Dear Kaushal,
I tried various method...but still the same error...it seems it's gluster
bug..is there any body can suggest work-around here ??
Thanks,
Punit
On Sun, Dec 7, 2014 at 8:40 PM, Punit Dambiwal <hypunit(a)gmail.com> wrote:
> Hi Kaushal,
>
> Still the same error...even try your suggested workaround :-
>
> -------------------
> Can you replace 'Before=network-online.target' with
> 'Wants=network-online.target' and try the boot again? This should
> force the network to be online before starting GlusterD.
> -------------------
>
> Thanks,
> Punit
>
> On Sat, Dec 6, 2014 at 11:44 AM, Punit Dambiwal <hypunit(a)gmail.com>
> wrote:
>
>> Hi Kaushal,
>>
>> I already have all the hosts entry in the /etc/hosts for the easy
>> resolution....i will try your method in the glusterd.services and check and
>> let you know....weather problem solve or not.....
>>
>> On Fri, Dec 5, 2014 at 9:50 PM, Kaushal M <kshlmster(a)gmail.com> wrote:
>>
>>> Can you replace 'Before=network-online.target' with
>>> 'Wants=network-online.target' and try the boot again? This should
>>> force the network to be online before starting GlusterD.
>>>
>>> If even that fails, you could try adding an entry into /etc/hosts with
>>> the hostname of the system. This should prevent any more failures.
>>>
>>> I still don't believe it's a problem with Gluster. Gluster uses apis
>>> provided by the system to perform name resolution. These definitely
>>> work correctly because you can start GlusterD later. Since the
>>> resolution failure only happens during boot, it points to system or
>>> network setup issues during boot. To me it seems like the network
>>> isn't completely setup at that point of time.
>>>
>>> ~kaushal
>>>
>>> On Fri, Dec 5, 2014 at 12:47 PM, Punit Dambiwal <hypunit(a)gmail.com>
>>> wrote:
>>> > Hi Kaushal,
>>> >
>>> > It seems it's bug in glusterfs 3.6....even i manage my systemd to
>>> start the
>>> > network service before glusterd...but it's still fail...
>>> >
>>> > ---------------
>>> > [Unit]
>>> > Description=GlusterFS, a clustered file-system server
>>> > After=network.target rpcbind.service
>>> > Before=network-online.target
>>> >
>>> > [Service]
>>> > Type=forking
>>> > PIDFile=/var/run/glusterd.pid
>>> > LimitNOFILE=65536
>>> > ExecStartPre=/etc/rc.d/init.d/network start
>>> > ExecStart=/usr/sbin/glusterd -p /var/run/glusterd.pid
>>> > KillMode=process
>>> >
>>> > [Install]
>>> > WantedBy=multi-user.target
>>> > ----------------
>>> >
>>> > Thanks,
>>> > Punit
>>> >
>>> > On Wed, Dec 3, 2014 at 8:56 PM, Kaushal M <kshlmster(a)gmail.com>
>>> wrote:
>>> >>
>>> >> I just remembered this.
>>> >>
>>> >> There was another user having a similar issue of GlusterD failing
to
>>> >> start on the mailing list a while back. The cause of his problem
was
>>> >> the way his network was brought up.
>>> >> IIRC, he was using a static network configuration. The problem
>>> >> vanished when he began using dhcp. Or it might have been he was
using
>>> >> dhcp.service and it got solved after switching to NetworkManager.
>>> >>
>>> >> This could be one more thing you could look at.
>>> >>
>>> >> I'll try to find the mail thread to see if it was the same
problem
>>> as you.
>>> >>
>>> >> ~kaushal
>>> >>
>>> >> On Wed, Dec 3, 2014 at 6:22 PM, Kaushal M
<kshlmster(a)gmail.com>
>>> wrote:
>>> >> > I don't know much about how the network target is brought
up in
>>> >> > CentOS7, but I'll try as much as I can.
>>> >> >
>>> >> > It seems to me that, after the network has been brought up and
by
>>> the
>>> >> > time GlusterD is started,
>>> >> > a. The machine hasn't yet recieved it's hostname, or
>>> >> > b. It hasn't yet registered with the name server.
>>> >> >
>>> >> > This is causing name resolution failures.
>>> >> >
>>> >> > I don't know if the network target could come up without
the
>>> machine
>>> >> > getting its hostname, so I'm pretty sure it's not a.
>>> >> >
>>> >> > So it seems to be b. But these kind of signing in happens only
in
>>> DDNS
>>> >> > systems, which doesn't seem to be the case for you.
>>> >> >
>>> >> > Both of these reasons might be wrong (most likely wrong).
You'd do
>>> >> > good if you could ask for help from someone with more
experience in
>>> >> > systemd + networking.
>>> >> >
>>> >> > ~kaushal
>>> >> >
>>> >> > On Wed, Dec 3, 2014 at 10:54 AM, Punit Dambiwal
<hypunit(a)gmail.com
>>> >
>>> >> > wrote:
>>> >> >> Hi Kaushal,
>>> >> >>
>>> >> >> This is the host...which i rebooted...would you mind to let
me
>>> know how
>>> >> >> i
>>> >> >> can make the glusterd sevice come up after network...i am
using
>>> >> >> centos7...if
>>> >> >> network is the issue...
>>> >> >>
>>> >> >> On Wed, Dec 3, 2014 at 11:54 AM, Kaushal M
<kshlmster(a)gmail.com>
>>> wrote:
>>> >> >>>
>>> >> >>> This peer cannot be identified.
>>> >> >>>
>>> >> >>> " [2014-12-03 02:29:25.998153] D
>>> >> >>>
[glusterd-peer-utils.c:121:glusterd_peerinfo_find_by_hostname]
>>> >> >>> 0-management:
>>> >> >>> Unable to find friend:
cpu05.zne01.hkg1.ovt.36stack.com"
>>> >> >>>
>>> >> >>> I don't know why this address is not being resolved
during boot
>>> time.
>>> >> >>> If
>>> >> >>> this is a valid peer, the the only reason I can think
of this
>>> that the
>>> >> >>> network is not up.
>>> >> >>>
>>> >> >>> If you had previously detached the peer forcefully, the
that
>>> could
>>> >> >>> have
>>> >> >>> left stale entries in some volumes. In this case as
well,
>>> GlusterD
>>> >> >>> will fail
>>> >> >>> to identify the peer.
>>> >> >>>
>>> >> >>> Do either of these reasons seem a possibility to you?
>>> >> >>>
>>> >> >>> On Dec 3, 2014 8:07 AM, "Punit Dambiwal"
<hypunit(a)gmail.com>
>>> wrote:
>>> >> >>>>
>>> >> >>>> Hi Kaushal,
>>> >> >>>>
>>> >> >>>> Please find the logs here :-
http://ur1.ca/iyoe5
and
>>> >> >>>>
http://ur1.ca/iyoed
>>> >> >>>>
>>> >> >>>> On Tue, Dec 2, 2014 at 10:43 PM, Kaushal M
<kshlmster(a)gmail.com
>>> >
>>> >> >>>> wrote:
>>> >> >>>>>
>>> >> >>>>> Hey Punit,
>>> >> >>>>> In the logs you've provided, GlusterD
appears to be running
>>> >> >>>>> correctly.
>>> >> >>>>> Could you provide the logs for the time period
when GlusterD
>>> >> >>>>> attempts to
>>> >> >>>>> start but fails.
>>> >> >>>>>
>>> >> >>>>> ~kaushal
>>> >> >>>>>
>>> >> >>>>> On Dec 2, 2014 8:03 PM, "Punit
Dambiwal" <hypunit(a)gmail.com>
>>> wrote:
>>> >> >>>>>>
>>> >> >>>>>> Hi Kaushal,
>>> >> >>>>>>
>>> >> >>>>>> Please find the logs here :-
http://ur1.ca/iyhs5 and
>>> >> >>>>>>
http://ur1.ca/iyhue
>>> >> >>>>>>
>>> >> >>>>>> Thanks,
>>> >> >>>>>> punit
>>> >> >>>>>>
>>> >> >>>>>>
>>> >> >>>>>> On Tue, Dec 2, 2014 at 12:00 PM, Kaushal M
<
>>> kshlmster(a)gmail.com>
>>> >> >>>>>> wrote:
>>> >> >>>>>>>
>>> >> >>>>>>> Hey Punit,
>>> >> >>>>>>> Could you start Glusterd in debug mode
and provide the logs
>>> here?
>>> >> >>>>>>> To start it in debug mode, append
'-LDEBUG' to the ExecStart
>>> line
>>> >> >>>>>>> in
>>> >> >>>>>>> the service file.
>>> >> >>>>>>>
>>> >> >>>>>>> ~kaushal
>>> >> >>>>>>>
>>> >> >>>>>>> On Mon, Dec 1, 2014 at 9:05 AM, Punit
Dambiwal <
>>> hypunit(a)gmail.com>
>>> >> >>>>>>> wrote:
>>> >> >>>>>>> > Hi,
>>> >> >>>>>>> >
>>> >> >>>>>>> > Can Any body help me on this ??
>>> >> >>>>>>> >
>>> >> >>>>>>> > On Thu, Nov 27, 2014 at 9:29 AM,
Punit Dambiwal
>>> >> >>>>>>> > <hypunit(a)gmail.com>
>>> >> >>>>>>> > wrote:
>>> >> >>>>>>> >>
>>> >> >>>>>>> >> Hi Kaushal,
>>> >> >>>>>>> >>
>>> >> >>>>>>> >> Thanks for the detailed
reply....let me explain my setup
>>> first
>>> >> >>>>>>> >> :-
>>> >> >>>>>>> >>
>>> >> >>>>>>> >> 1. Ovirt Engine
>>> >> >>>>>>> >> 2. 4* host as well as storage
machine (Host and gluster
>>> >> >>>>>>> >> combined)
>>> >> >>>>>>> >> 3. Every host has 24
bricks...
>>> >> >>>>>>> >>
>>> >> >>>>>>> >> Now whenever the host machine
reboot...it can come up but
>>> can
>>> >> >>>>>>> >> not
>>> >> >>>>>>> >> join the
>>> >> >>>>>>> >> cluster again and through the
following error "Gluster
>>> command
>>> >> >>>>>>> >> [<UNKNOWN>]
>>> >> >>>>>>> >> failed on server.."
>>> >> >>>>>>> >>
>>> >> >>>>>>> >> Please check my comment in
line :-
>>> >> >>>>>>> >>
>>> >> >>>>>>> >> 1. Use the same string for
doing the peer probe and for
>>> the
>>> >> >>>>>>> >> brick
>>> >> >>>>>>> >> address
>>> >> >>>>>>> >> during volume
create/add-brick. Ideally, we suggest you
>>> use
>>> >> >>>>>>> >> properly
>>> >> >>>>>>> >> resolvable FQDNs everywhere.
If that is not possible,
>>> then use
>>> >> >>>>>>> >> only
>>> >> >>>>>>> >> IP
>>> >> >>>>>>> >> addresses. Try to avoid short
names.
>>> >> >>>>>>> >> ---------------
>>> >> >>>>>>> >> [root@cpu05 ~]# gluster peer
status
>>> >> >>>>>>> >> Number of Peers: 3
>>> >> >>>>>>> >>
>>> >> >>>>>>> >> Hostname:
cpu03.stack.com
>>> >> >>>>>>> >> Uuid:
5729b8c4-e80d-4353-b456-6f467bddbdfb
>>> >> >>>>>>> >> State: Peer in Cluster
(Connected)
>>> >> >>>>>>> >>
>>> >> >>>>>>> >> Hostname:
cpu04.stack.com
>>> >> >>>>>>> >> Uuid:
d272b790-c4b2-4bed-ba68-793656e6d7b0
>>> >> >>>>>>> >> State: Peer in Cluster
(Connected)
>>> >> >>>>>>> >> Other names:
>>> >> >>>>>>> >> 10.10.0.8
>>> >> >>>>>>> >>
>>> >> >>>>>>> >> Hostname:
cpu02.stack.com
>>> >> >>>>>>> >> Uuid:
8d8a7041-950e-40d0-85f9-58d14340ca25
>>> >> >>>>>>> >> State: Peer in Cluster
(Connected)
>>> >> >>>>>>> >> [root@cpu05 ~]#
>>> >> >>>>>>> >> ----------------
>>> >> >>>>>>> >> 2. During boot up, make sure
to launch glusterd only
>>> after the
>>> >> >>>>>>> >> network is
>>> >> >>>>>>> >> up. This will allow the new
peer identification mechanism
>>> to do
>>> >> >>>>>>> >> its
>>> >> >>>>>>> >> job correctly.
>>> >> >>>>>>> >> >> I think the service
itself doing the same job....
>>> >> >>>>>>> >>
>>> >> >>>>>>> >> [root@cpu05 ~]# cat
>>> /usr/lib/systemd/system/glusterd.service
>>> >> >>>>>>> >> [Unit]
>>> >> >>>>>>> >> Description=GlusterFS, a
clustered file-system server
>>> >> >>>>>>> >> After=network.target
rpcbind.service
>>> >> >>>>>>> >> Before=network-online.target
>>> >> >>>>>>> >>
>>> >> >>>>>>> >> [Service]
>>> >> >>>>>>> >> Type=forking
>>> >> >>>>>>> >> PIDFile=/var/run/glusterd.pid
>>> >> >>>>>>> >> LimitNOFILE=65536
>>> >> >>>>>>> >> ExecStart=/usr/sbin/glusterd
-p /var/run/glusterd.pid
>>> >> >>>>>>> >> KillMode=process
>>> >> >>>>>>> >>
>>> >> >>>>>>> >> [Install]
>>> >> >>>>>>> >> WantedBy=multi-user.target
>>> >> >>>>>>> >> [root@cpu05 ~]#
>>> >> >>>>>>> >> --------------------
>>> >> >>>>>>> >>
>>> >> >>>>>>> >> gluster logs :-
>>> >> >>>>>>> >>
>>> >> >>>>>>> >> [2014-11-24 09:22:22.147471] I
[MSGID: 100030]
>>> >> >>>>>>> >> [glusterfsd.c:2018:main]
>>> >> >>>>>>> >> 0-/usr/sbin/glusterd: Started
running /usr/sbin/glusterd
>>> >> >>>>>>> >> version
>>> >> >>>>>>> >> 3.6.1
>>> >> >>>>>>> >> (args: /usr/sbin/glusterd -p
/var/run/glusterd.pid)
>>> >> >>>>>>> >> [2014-11-24 09:22:22.151565] I
[glusterd.c:1214:init]
>>> >> >>>>>>> >> 0-management:
>>> >> >>>>>>> >> Maximum allowed open file
descriptors set to 65536
>>> >> >>>>>>> >> [2014-11-24 09:22:22.151599] I
[glusterd.c:1259:init]
>>> >> >>>>>>> >> 0-management:
>>> >> >>>>>>> >> Using
>>> >> >>>>>>> >> /var/lib/glusterd as working
directory
>>> >> >>>>>>> >> [2014-11-24 09:22:22.155216]
W
>>> >> >>>>>>> >>
[rdma.c:4195:__gf_rdma_ctx_create]
>>> >> >>>>>>> >> 0-rpc-transport/rdma: rdma_cm
event channel creation
>>> failed (No
>>> >> >>>>>>> >> such device)
>>> >> >>>>>>> >> [2014-11-24 09:22:22.155264] E
[rdma.c:4483:init]
>>> >> >>>>>>> >> 0-rdma.management:
>>> >> >>>>>>> >> Failed to initialize IB
Device
>>> >> >>>>>>> >> [2014-11-24 09:22:22.155285]
E
>>> >> >>>>>>> >>
[rpc-transport.c:333:rpc_transport_load]
>>> >> >>>>>>> >> 0-rpc-transport:
'rdma' initialization failed
>>> >> >>>>>>> >> [2014-11-24 09:22:22.155354]
W
>>> >> >>>>>>> >>
[rpcsvc.c:1524:rpcsvc_transport_create]
>>> >> >>>>>>> >> 0-rpc-service: cannot create
listener, initing the
>>> transport
>>> >> >>>>>>> >> failed
>>> >> >>>>>>> >> [2014-11-24 09:22:22.156290]
I
>>> >> >>>>>>> >>
[glusterd.c:413:glusterd_check_gsync_present] 0-glusterd:
>>> >> >>>>>>> >> geo-replication
>>> >> >>>>>>> >> module not installed in the
system
>>> >> >>>>>>> >> [2014-11-24 09:22:22.161318]
I
>>> >> >>>>>>> >>
[glusterd-store.c:2043:glusterd_restore_op_version]
>>> 0-glusterd:
>>> >> >>>>>>> >> retrieved
>>> >> >>>>>>> >> op-version: 30600
>>> >> >>>>>>> >> [2014-11-24 09:22:22.821800]
I
>>> >> >>>>>>> >>
>>> [glusterd-handler.c:3146:glusterd_friend_add_from_peerinfo]
>>> >> >>>>>>> >> 0-management:
>>> >> >>>>>>> >> connect returned 0
>>> >> >>>>>>> >> [2014-11-24 09:22:22.825810]
I
>>> >> >>>>>>> >>
>>> [glusterd-handler.c:3146:glusterd_friend_add_from_peerinfo]
>>> >> >>>>>>> >> 0-management:
>>> >> >>>>>>> >> connect returned 0
>>> >> >>>>>>> >> [2014-11-24 09:22:22.828705]
I
>>> >> >>>>>>> >>
>>> [glusterd-handler.c:3146:glusterd_friend_add_from_peerinfo]
>>> >> >>>>>>> >> 0-management:
>>> >> >>>>>>> >> connect returned 0
>>> >> >>>>>>> >> [2014-11-24 09:22:22.828771]
I
>>> >> >>>>>>> >>
[rpc-clnt.c:969:rpc_clnt_connection_init]
>>> >> >>>>>>> >> 0-management: setting
frame-timeout to 600
>>> >> >>>>>>> >> [2014-11-24 09:22:22.832670]
I
>>> >> >>>>>>> >>
[rpc-clnt.c:969:rpc_clnt_connection_init]
>>> >> >>>>>>> >> 0-management: setting
frame-timeout to 600
>>> >> >>>>>>> >> [2014-11-24 09:22:22.835919]
I
>>> >> >>>>>>> >>
[rpc-clnt.c:969:rpc_clnt_connection_init]
>>> >> >>>>>>> >> 0-management: setting
frame-timeout to 600
>>> >> >>>>>>> >> [2014-11-24 09:22:22.840209]
E
>>> >> >>>>>>> >>
[glusterd-store.c:4248:glusterd_resolve_all_bricks]
>>> 0-glusterd:
>>> >> >>>>>>> >> resolve
>>> >> >>>>>>> >> brick failed in restore
>>> >> >>>>>>> >> [2014-11-24 09:22:22.840233] E
[xlator.c:425:xlator_init]
>>> >> >>>>>>> >> 0-management:
>>> >> >>>>>>> >> Initialization of volume
'management' failed, review your
>>> >> >>>>>>> >> volfile
>>> >> >>>>>>> >> again
>>> >> >>>>>>> >> [2014-11-24 09:22:22.840245]
E
>>> >> >>>>>>> >>
[graph.c:322:glusterfs_graph_init]
>>> >> >>>>>>> >> 0-management: initializing
translator failed
>>> >> >>>>>>> >> [2014-11-24 09:22:22.840264]
E
>>> >> >>>>>>> >>
[graph.c:525:glusterfs_graph_activate]
>>> >> >>>>>>> >> 0-graph: init failed
>>> >> >>>>>>> >> [2014-11-24 09:22:22.840754]
W
>>> >> >>>>>>> >>
[glusterfsd.c:1194:cleanup_and_exit]
>>> >> >>>>>>> >> (-->
>>> >> >>>>>>> >> 0-: received signum (0),
shutting down
>>> >> >>>>>>> >>
>>> >> >>>>>>> >> Thanks,
>>> >> >>>>>>> >> Punit
>>> >> >>>>>>> >>
>>> >> >>>>>>> >>
>>> >> >>>>>>> >>
>>> >> >>>>>>> >>
>>> >> >>>>>>> >> On Wed, Nov 26, 2014 at 7:14
PM, Kaushal M
>>> >> >>>>>>> >> <kshlmster(a)gmail.com>
>>> >> >>>>>>> >> wrote:
>>> >> >>>>>>> >>>
>>> >> >>>>>>> >>> Based on the logs I can
guess that glusterd is being
>>> started
>>> >> >>>>>>> >>> before
>>> >> >>>>>>> >>> the network has come up
and that the addresses given to
>>> bricks
>>> >> >>>>>>> >>> do
>>> >> >>>>>>> >>> not
>>> >> >>>>>>> >>> directly match the
addresses used in during peer probe.
>>> >> >>>>>>> >>>
>>> >> >>>>>>> >>> The gluster_after_reboot
log has the line "[2014-11-25
>>> >> >>>>>>> >>> 06:46:09.972113] E
>>> >> >>>>>>> >>>
[glusterd-store.c:2632:glusterd_resolve_all_bricks]
>>> >> >>>>>>> >>> 0-glusterd: resolve brick
failed in restore".
>>> >> >>>>>>> >>>
>>> >> >>>>>>> >>> Brick resolution fails
when glusterd cannot match the
>>> address
>>> >> >>>>>>> >>> for
>>> >> >>>>>>> >>> the
>>> >> >>>>>>> >>> brick, with one of the
peers. Brick resolution happens
>>> in two
>>> >> >>>>>>> >>> phases,
>>> >> >>>>>>> >>> 1. We first try to
identify the peer by performing string
>>> >> >>>>>>> >>> comparisions
>>> >> >>>>>>> >>> with the brick address and
the peer addresses (The peer
>>> names
>>> >> >>>>>>> >>> will
>>> >> >>>>>>> >>> be
>>> >> >>>>>>> >>> the names/addresses that
were given when the peer was
>>> probed).
>>> >> >>>>>>> >>> 2. If we don't find a
match from step 1, we will then
>>> resolve
>>> >> >>>>>>> >>> all
>>> >> >>>>>>> >>> the
>>> >> >>>>>>> >>> brick address and the peer
addresses into addrinfo
>>> structs,
>>> >> >>>>>>> >>> and
>>> >> >>>>>>> >>> then
>>> >> >>>>>>> >>> compare these structs to
find a match. This process
>>> should
>>> >> >>>>>>> >>> generally
>>> >> >>>>>>> >>> find a match if available.
This will fail only if the
>>> network
>>> >> >>>>>>> >>> is
>>> >> >>>>>>> >>> not
>>> >> >>>>>>> >>> up yet as we cannot
resolve addresses.
>>> >> >>>>>>> >>>
>>> >> >>>>>>> >>> The above steps are
applicable only to glusterfs versions
>>> >> >>>>>>> >>> >=3.6.
>>> >> >>>>>>> >>> They
>>> >> >>>>>>> >>> were introduced to reduce
problems with peer
>>> identification,
>>> >> >>>>>>> >>> like
>>> >> >>>>>>> >>> the
>>> >> >>>>>>> >>> one you encountered
>>> >> >>>>>>> >>>
>>> >> >>>>>>> >>> Since both of the steps
failed to find a match in one
>>> run, but
>>> >> >>>>>>> >>> succeeded later, we can
come to the conclusion that,
>>> >> >>>>>>> >>> a) the bricks don't
have the exact same string used in
>>> peer
>>> >> >>>>>>> >>> probe
>>> >> >>>>>>> >>> for
>>> >> >>>>>>> >>> their addresses as step 1
failed, and
>>> >> >>>>>>> >>> b) the network was not up
in the initial run, as step 2
>>> failed
>>> >> >>>>>>> >>> during
>>> >> >>>>>>> >>> the initial run, but
passed in the second run.
>>> >> >>>>>>> >>>
>>> >> >>>>>>> >>> Please let me know if my
conclusion is correct.
>>> >> >>>>>>> >>>
>>> >> >>>>>>> >>> If it is, you can solve
your problem in two ways.
>>> >> >>>>>>> >>> 1. Use the same string for
doing the peer probe and for
>>> the
>>> >> >>>>>>> >>> brick
>>> >> >>>>>>> >>> address during volume
create/add-brick. Ideally, we
>>> suggest
>>> >> >>>>>>> >>> you
>>> >> >>>>>>> >>> use
>>> >> >>>>>>> >>> properly resolvable FQDNs
everywhere. If that is not
>>> possible,
>>> >> >>>>>>> >>> then
>>> >> >>>>>>> >>> use only IP addresses. Try
to avoid short names.
>>> >> >>>>>>> >>> 2. During boot up, make
sure to launch glusterd only
>>> after the
>>> >> >>>>>>> >>> network
>>> >> >>>>>>> >>> is up. This will allow the
new peer identification
>>> mechanism
>>> >> >>>>>>> >>> to do
>>> >> >>>>>>> >>> its
>>> >> >>>>>>> >>> job correctly.
>>> >> >>>>>>> >>>
>>> >> >>>>>>> >>>
>>> >> >>>>>>> >>> If you have already
followed these steps and yet still
>>> hit the
>>> >> >>>>>>> >>> problem, then please
provide more information (setup,
>>> logs,
>>> >> >>>>>>> >>> etc.).
>>> >> >>>>>>> >>> It
>>> >> >>>>>>> >>> could be much different
problem that you are facing.
>>> >> >>>>>>> >>>
>>> >> >>>>>>> >>> ~kaushal
>>> >> >>>>>>> >>>
>>> >> >>>>>>> >>> On Wed, Nov 26, 2014 at
4:01 PM, Punit Dambiwal
>>> >> >>>>>>> >>> <hypunit(a)gmail.com>
>>> >> >>>>>>> >>> wrote:
>>> >> >>>>>>> >>> > Is there any one can
help on this ??
>>> >> >>>>>>> >>> >
>>> >> >>>>>>> >>> > Thanks,
>>> >> >>>>>>> >>> > punit
>>> >> >>>>>>> >>> >
>>> >> >>>>>>> >>> > On Wed, Nov 26, 2014
at 9:42 AM, Punit Dambiwal
>>> >> >>>>>>> >>> >
<hypunit(a)gmail.com>
>>> >> >>>>>>> >>> > wrote:
>>> >> >>>>>>> >>> >>
>>> >> >>>>>>> >>> >> Hi,
>>> >> >>>>>>> >>> >>
>>> >> >>>>>>> >>> >> My Glusterfs
version is :- glusterfs-3.6.1-1.el7
>>> >> >>>>>>> >>> >>
>>> >> >>>>>>> >>> >> On Wed, Nov 26,
2014 at 1:59 AM, Kanagaraj Mayilsamy
>>> >> >>>>>>> >>> >>
<kmayilsa(a)redhat.com>
>>> >> >>>>>>> >>> >> wrote:
>>> >> >>>>>>> >>> >>>
>>> >> >>>>>>> >>> >>>
[+Gluster-users(a)gluster.org]
>>> >> >>>>>>> >>> >>>
>>> >> >>>>>>> >>> >>>
"Initialization of volume 'management' failed,
>>> review your
>>> >> >>>>>>> >>> >>> volfile
>>> >> >>>>>>> >>> >>> again",
glusterd throws this error when the service
>>> is
>>> >> >>>>>>> >>> >>> started
>>> >> >>>>>>> >>> >>>
automatically
>>> >> >>>>>>> >>> >>> after the
reboot. But the service is successfully
>>> started
>>> >> >>>>>>> >>> >>> later
>>> >> >>>>>>> >>> >>> manually by
>>> >> >>>>>>> >>> >>> the user.
>>> >> >>>>>>> >>> >>>
>>> >> >>>>>>> >>> >>> can somebody
from gluster-users please help on this?
>>> >> >>>>>>> >>> >>>
>>> >> >>>>>>> >>> >>> glusterfs
version: 3.5.1
>>> >> >>>>>>> >>> >>>
>>> >> >>>>>>> >>> >>> Thanks,
>>> >> >>>>>>> >>> >>> Kanagaraj
>>> >> >>>>>>> >>> >>>
>>> >> >>>>>>> >>> >>> -----
Original Message -----
>>> >> >>>>>>> >>> >>> > From:
"Punit Dambiwal" <hypunit(a)gmail.com>
>>> >> >>>>>>> >>> >>> > To:
"Kanagaraj" <kmayilsa(a)redhat.com>
>>> >> >>>>>>> >>> >>> > Cc:
users(a)ovirt.org
>>> >> >>>>>>> >>> >>> > Sent:
Tuesday, November 25, 2014 7:24:45 PM
>>> >> >>>>>>> >>> >>> > Subject:
Re: [ovirt-users] Gluster command
>>> [<UNKNOWN>]
>>> >> >>>>>>> >>> >>> > failed
on
>>> >> >>>>>>> >>> >>> >
server...
>>> >> >>>>>>> >>> >>> >
>>> >> >>>>>>> >>> >>> > Hi
Kanagraj,
>>> >> >>>>>>> >>> >>> >
>>> >> >>>>>>> >>> >>> > Please
check the attached log files....i didn't
>>> find any
>>> >> >>>>>>> >>> >>> > thing
>>> >> >>>>>>> >>> >>> >
special....
>>> >> >>>>>>> >>> >>> >
>>> >> >>>>>>> >>> >>> > On Tue,
Nov 25, 2014 at 12:12 PM, Kanagaraj
>>> >> >>>>>>> >>> >>> >
<kmayilsa(a)redhat.com>
>>> >> >>>>>>> >>> >>> > wrote:
>>> >> >>>>>>> >>> >>> >
>>> >> >>>>>>> >>> >>> > > Do
you see any errors in
>>> >> >>>>>>> >>> >>> > >
>>> /var/log/glusterfs/etc-glusterfs-glusterd.vol.log or
>>> >> >>>>>>> >>> >>> > >
vdsm.log
>>> >> >>>>>>> >>> >>> > >
when
>>> >> >>>>>>> >>> >>> > >
the
>>> >> >>>>>>> >>> >>> > >
service is trying to start automatically after
>>> the
>>> >> >>>>>>> >>> >>> > >
reboot?
>>> >> >>>>>>> >>> >>> > >
>>> >> >>>>>>> >>> >>> > >
Thanks,
>>> >> >>>>>>> >>> >>> > >
Kanagaraj
>>> >> >>>>>>> >>> >>> > >
>>> >> >>>>>>> >>> >>> > >
>>> >> >>>>>>> >>> >>> > > On
11/24/2014 08:13 PM, Punit Dambiwal wrote:
>>> >> >>>>>>> >>> >>> > >
>>> >> >>>>>>> >>> >>> > > Hi
Kanagaraj,
>>> >> >>>>>>> >>> >>> > >
>>> >> >>>>>>> >>> >>> > >
Yes...once i will start the gluster service and
>>> then
>>> >> >>>>>>> >>> >>> > >
vdsmd
>>> >> >>>>>>> >>> >>> > >
...the
>>> >> >>>>>>> >>> >>> > >
host
>>> >> >>>>>>> >>> >>> > > can
connect to cluster...but the question is why
>>> it's
>>> >> >>>>>>> >>> >>> > >
not
>>> >> >>>>>>> >>> >>> > >
started
>>> >> >>>>>>> >>> >>> > >
even it
>>> >> >>>>>>> >>> >>> > > has
chkconfig enabled...
>>> >> >>>>>>> >>> >>> > >
>>> >> >>>>>>> >>> >>> > > I
have tested it in two host cluster
>>> >> >>>>>>> >>> >>> > >
environment...(Centos 6.6
>>> >> >>>>>>> >>> >>> > >
and
>>> >> >>>>>>> >>> >>> > >
centos 7.0) on both hypervisior cluster..it's
>>> failed
>>> >> >>>>>>> >>> >>> > > to
>>> >> >>>>>>> >>> >>> > >
reconnect
>>> >> >>>>>>> >>> >>> > > in
>>> >> >>>>>>> >>> >>> > > to
>>> >> >>>>>>> >>> >>> > >
cluster after reboot....
>>> >> >>>>>>> >>> >>> > >
>>> >> >>>>>>> >>> >>> > > In
both the environment glusterd enabled for
>>> next
>>> >> >>>>>>> >>> >>> > >
boot....but
>>> >> >>>>>>> >>> >>> > >
it's
>>> >> >>>>>>> >>> >>> > >
failed with the same error....seems it's bug in
>>> either
>>> >> >>>>>>> >>> >>> > >
gluster or
>>> >> >>>>>>> >>> >>> > >
Ovirt ??
>>> >> >>>>>>> >>> >>> > >
>>> >> >>>>>>> >>> >>> > >
Please help me to find the workaround here if
>>> can not
>>> >> >>>>>>> >>> >>> > >
resolve
>>> >> >>>>>>> >>> >>> > >
it...as
>>> >> >>>>>>> >>> >>> > >
without this the Host machine can not connect
>>> after
>>> >> >>>>>>> >>> >>> > >
reboot....that
>>> >> >>>>>>> >>> >>> > >
means
>>> >> >>>>>>> >>> >>> > >
engine will consider it as down and every time
>>> need to
>>> >> >>>>>>> >>> >>> > >
manually
>>> >> >>>>>>> >>> >>> > >
start
>>> >> >>>>>>> >>> >>> > >
the
>>> >> >>>>>>> >>> >>> > >
gluster service and vdsmd... ??
>>> >> >>>>>>> >>> >>> > >
>>> >> >>>>>>> >>> >>> > >
Thanks,
>>> >> >>>>>>> >>> >>> > >
Punit
>>> >> >>>>>>> >>> >>> > >
>>> >> >>>>>>> >>> >>> > > On
Mon, Nov 24, 2014 at 10:20 PM, Kanagaraj
>>> >> >>>>>>> >>> >>> > >
<kmayilsa(a)redhat.com>
>>> >> >>>>>>> >>> >>> > >
wrote:
>>> >> >>>>>>> >>> >>> > >
>>> >> >>>>>>> >>> >>> > >>
From vdsm.log "error: Connection failed. Please
>>> >> >>>>>>> >>> >>> > >>
check if
>>> >> >>>>>>> >>> >>> > >>
gluster
>>> >> >>>>>>> >>> >>> > >>
daemon
>>> >> >>>>>>> >>> >>> > >>
is operational."
>>> >> >>>>>>> >>> >>> >
>>
>>> >> >>>>>>> >>> >>> > >>
Starting glusterd service should fix this issue.
>>> >> >>>>>>> >>> >>> > >>
'service
>>> >> >>>>>>> >>> >>> > >>
glusterd
>>> >> >>>>>>> >>> >>> > >>
start'
>>> >> >>>>>>> >>> >>> > >>
But i am wondering why the glusterd was not
>>> started
>>> >> >>>>>>> >>> >>> > >>
automatically
>>> >> >>>>>>> >>> >>> > >>
after
>>> >> >>>>>>> >>> >>> > >>
the reboot.
>>> >> >>>>>>> >>> >>> >
>>
>>> >> >>>>>>> >>> >>> > >>
Thanks,
>>> >> >>>>>>> >>> >>> > >>
Kanagaraj
>>> >> >>>>>>> >>> >>> >
>>
>>> >> >>>>>>> >>> >>> >
>>
>>> >> >>>>>>> >>> >>> >
>>
>>> >> >>>>>>> >>> >>> > >>
On 11/24/2014 07:18 PM, Punit Dambiwal wrote:
>>> >> >>>>>>> >>> >>> >
>>
>>> >> >>>>>>> >>> >>> > >>
Hi Kanagaraj,
>>> >> >>>>>>> >>> >>> >
>>
>>> >> >>>>>>> >>> >>> > >>
Please find the attached VDSM logs :-
>>> >> >>>>>>> >>> >>> >
>>
>>> >> >>>>>>> >>> >>> > >>
----------------
>>> >> >>>>>>> >>> >>> > >>
Thread-13::DEBUG::2014-11-24
>>> >> >>>>>>> >>> >>> >
>>
>>> >> >>>>>>> >>> >>> >
>>
>>> >> >>>>>>> >>> >>> >
>>
>>> >> >>>>>>> >>> >>> >
>>
>>> >> >>>>>>> >>> >>> >
>>
>>>
21:41:17,182::resourceManager::977::Storage.ResourceManager.Owner::(cancelAll)
>>> >> >>>>>>> >>> >>> > >>
Owner.cancelAll requests {}
>>> >> >>>>>>> >>> >>> > >>
Thread-13::DEBUG::2014-11-24
>>> >> >>>>>>> >>> >>> >
>>
>>> >> >>>>>>> >>> >>> >
>>
>>> >> >>>>>>> >>> >>> >
>>
>>> 21:41:17,182::task::993::Storage.TaskManager.Task::(_decref)
>>> >> >>>>>>> >>> >>> >
>>
>>> Task=`1691d409-9b27-4585-8281-5ec26154367a`::ref 0
>>> >> >>>>>>> >>> >>> > >>
aborting
>>> >> >>>>>>> >>> >>> > >>
False
>>> >> >>>>>>> >>> >>> > >>
Thread-13::DEBUG::2014-11-24
>>> >> >>>>>>> >>> >>> >
>>
>>> >> >>>>>>> >>> >>> >
>>
>>> >> >>>>>>> >>> >>> >
>>
>>> >> >>>>>>> >>> >>> >
>>
>>> 21:41:32,393::task::595::Storage.TaskManager.Task::(_updateState)
>>> >> >>>>>>> >>> >>> >
>>
>>> Task=`994c7bc3-a236-4d03-a732-e068c7ed9ed4`::moving
>>> >> >>>>>>> >>> >>> > >>
from
>>> >> >>>>>>> >>> >>> > >>
state
>>> >> >>>>>>> >>> >>> > >>
init
>>> >> >>>>>>> >>> >>> > >>
->
>>> >> >>>>>>> >>> >>> > >>
state preparing
>>> >> >>>>>>> >>> >>> > >>
Thread-13::INFO::2014-11-24
>>> >> >>>>>>> >>> >>> >
>>
>>> 21:41:32,393::logUtils::44::dispatcher::(wrapper) Run
>>> >> >>>>>>> >>> >>> > >>
and
>>> >> >>>>>>> >>> >>> > >>
protect:
>>> >> >>>>>>> >>> >>> > >>
repoStats(options=None)
>>> >> >>>>>>> >>> >>> > >>
Thread-13::INFO::2014-11-24
>>> >> >>>>>>> >>> >>> >
>>
>>> 21:41:32,393::logUtils::47::dispatcher::(wrapper) Run
>>> >> >>>>>>> >>> >>> > >>
and
>>> >> >>>>>>> >>> >>> > >>
protect:
>>> >> >>>>>>> >>> >>> > >>
repoStats, Return response: {}
>>> >> >>>>>>> >>> >>> > >>
Thread-13::DEBUG::2014-11-24
>>> >> >>>>>>> >>> >>> >
>>
>>> >> >>>>>>> >>> >>> >
>>
>>> >> >>>>>>> >>> >>> >
>>
>>> 21:41:32,393::task::1191::Storage.TaskManager.Task::(prepare)
>>> >> >>>>>>> >>> >>> >
>>
>>> >> >>>>>>> >>> >>> >
>>
>>> Task=`994c7bc3-a236-4d03-a732-e068c7ed9ed4`::finished: {}
>>> >> >>>>>>> >>> >>> > >>
Thread-13::DEBUG::2014-11-24
>>> >> >>>>>>> >>> >>> >
>>
>>> >> >>>>>>> >>> >>> >
>>
>>> >> >>>>>>> >>> >>> >
>>
>>> >> >>>>>>> >>> >>> >
>>
>>> 21:41:32,394::task::595::Storage.TaskManager.Task::(_updateState)
>>> >> >>>>>>> >>> >>> >
>>
>>> Task=`994c7bc3-a236-4d03-a732-e068c7ed9ed4`::moving
>>> >> >>>>>>> >>> >>> > >>
from
>>> >> >>>>>>> >>> >>> > >>
state
>>> >> >>>>>>> >>> >>> > >>
preparing
>>> >> >>>>>>> >>> >>> > >>
->
>>> >> >>>>>>> >>> >>> > >>
state finished
>>> >> >>>>>>> >>> >>> > >>
Thread-13::DEBUG::2014-11-24
>>> >> >>>>>>> >>> >>> >
>>
>>> >> >>>>>>> >>> >>> >
>>
>>> >> >>>>>>> >>> >>> >
>>
>>> >> >>>>>>> >>> >>> >
>>
>>> >> >>>>>>> >>> >>> >
>>
>>>
21:41:32,394::resourceManager::940::Storage.ResourceManager.Owner::(releaseAll)
>>> >> >>>>>>> >>> >>> > >>
Owner.releaseAll requests {} resources {}
>>> >> >>>>>>> >>> >>> > >>
Thread-13::DEBUG::2014-11-24
>>> >> >>>>>>> >>> >>> >
>>
>>> >> >>>>>>> >>> >>> >
>>
>>> >> >>>>>>> >>> >>> >
>>
>>> >> >>>>>>> >>> >>> >
>>
>>> >> >>>>>>> >>> >>> >
>>
>>>
21:41:32,394::resourceManager::977::Storage.ResourceManager.Owner::(cancelAll)
>>> >> >>>>>>> >>> >>> > >>
Owner.cancelAll requests {}
>>> >> >>>>>>> >>> >>> > >>
Thread-13::DEBUG::2014-11-24
>>> >> >>>>>>> >>> >>> >
>>
>>> >> >>>>>>> >>> >>> >
>>
>>> >> >>>>>>> >>> >>> >
>>
>>> 21:41:32,394::task::993::Storage.TaskManager.Task::(_decref)
>>> >> >>>>>>> >>> >>> >
>>
>>> Task=`994c7bc3-a236-4d03-a732-e068c7ed9ed4`::ref 0
>>> >> >>>>>>> >>> >>> > >>
aborting
>>> >> >>>>>>> >>> >>> > >>
False
>>> >> >>>>>>> >>> >>> > >>
Thread-13::DEBUG::2014-11-24
>>> >> >>>>>>> >>> >>> >
>>
>>> 21:41:41,550::BindingXMLRPC::1132::vds::(wrapper)
>>> >> >>>>>>> >>> >>> > >>
client
>>> >> >>>>>>> >>> >>> > >>
[10.10.10.2]::call
>>> >> >>>>>>> >>> >>> > >>
getCapabilities with () {}
>>> >> >>>>>>> >>> >>> > >>
Thread-13::DEBUG::2014-11-24
>>> >> >>>>>>> >>> >>> > >>
21:41:41,553::utils::738::root::(execCmd)
>>> >> >>>>>>> >>> >>> > >>
/sbin/ip route show to 0.0.0.0/0 table all
>>> (cwd None)
>>> >> >>>>>>> >>> >>> > >>
Thread-13::DEBUG::2014-11-24
>>> >> >>>>>>> >>> >>> > >>
21:41:41,560::utils::758::root::(execCmd)
>>> >> >>>>>>> >>> >>> > >>
SUCCESS: <err> = ''; <rc> = 0
>>> >> >>>>>>> >>> >>> > >>
Thread-13::DEBUG::2014-11-24
>>> >> >>>>>>> >>> >>> >
>>
>>> 21:41:41,588::caps::728::root::(_getKeyPackages) rpm
>>> >> >>>>>>> >>> >>> > >>
package
>>> >> >>>>>>> >>> >>> > >>
('gluster-swift',) not found
>>> >> >>>>>>> >>> >>> > >>
Thread-13::DEBUG::2014-11-24
>>> >> >>>>>>> >>> >>> >
>>
>>> 21:41:41,592::caps::728::root::(_getKeyPackages) rpm
>>> >> >>>>>>> >>> >>> > >>
package
>>> >> >>>>>>> >>> >>> > >>
('gluster-swift-object',) not found
>>> >> >>>>>>> >>> >>> > >>
Thread-13::DEBUG::2014-11-24
>>> >> >>>>>>> >>> >>> >
>>
>>> 21:41:41,593::caps::728::root::(_getKeyPackages) rpm
>>> >> >>>>>>> >>> >>> > >>
package
>>> >> >>>>>>> >>> >>> > >>
('gluster-swift-plugin',) not found
>>> >> >>>>>>> >>> >>> > >>
Thread-13::DEBUG::2014-11-24
>>> >> >>>>>>> >>> >>> >
>>
>>> 21:41:41,598::caps::728::root::(_getKeyPackages) rpm
>>> >> >>>>>>> >>> >>> > >>
package
>>> >> >>>>>>> >>> >>> > >>
('gluster-swift-account',) not found
>>> >> >>>>>>> >>> >>> > >>
Thread-13::DEBUG::2014-11-24
>>> >> >>>>>>> >>> >>> >
>>
>>> 21:41:41,598::caps::728::root::(_getKeyPackages) rpm
>>> >> >>>>>>> >>> >>> > >>
package
>>> >> >>>>>>> >>> >>> > >>
('gluster-swift-proxy',) not found
>>> >> >>>>>>> >>> >>> > >>
Thread-13::DEBUG::2014-11-24
>>> >> >>>>>>> >>> >>> >
>>
>>> 21:41:41,598::caps::728::root::(_getKeyPackages) rpm
>>> >> >>>>>>> >>> >>> > >>
package
>>> >> >>>>>>> >>> >>> > >>
('gluster-swift-doc',) not found
>>> >> >>>>>>> >>> >>> > >>
Thread-13::DEBUG::2014-11-24
>>> >> >>>>>>> >>> >>> >
>>
>>> 21:41:41,599::caps::728::root::(_getKeyPackages) rpm
>>> >> >>>>>>> >>> >>> > >>
package
>>> >> >>>>>>> >>> >>> > >>
('gluster-swift-container',) not found
>>> >> >>>>>>> >>> >>> > >>
Thread-13::DEBUG::2014-11-24
>>> >> >>>>>>> >>> >>> >
>>
>>> 21:41:41,599::caps::728::root::(_getKeyPackages) rpm
>>> >> >>>>>>> >>> >>> > >>
package
>>> >> >>>>>>> >>> >>> > >>
('glusterfs-geo-replication',) not found
>>> >> >>>>>>> >>> >>> > >>
Thread-13::DEBUG::2014-11-24
>>> >> >>>>>>> >>> >>> > >>
21:41:41,600::caps::646::root::(get)
>>> >> >>>>>>> >>> >>> > >>
VirtioRNG DISABLED: libvirt version
>>> 0.10.2-29.el6_5.9
>>> >> >>>>>>> >>> >>> > >>
required
>>> >> >>>>>>> >>> >>> > >>
>=
>>> >> >>>>>>> >>> >>> > >>
0.10.2-31
>>> >> >>>>>>> >>> >>> > >>
Thread-13::DEBUG::2014-11-24
>>> >> >>>>>>> >>> >>> >
>>
>>> 21:41:41,603::BindingXMLRPC::1139::vds::(wrapper)
>>> >> >>>>>>> >>> >>> > >>
return
>>> >> >>>>>>> >>> >>> > >>
getCapabilities
>>> >> >>>>>>> >>> >>> > >>
with {'status': {'message': 'Done', 'code': 0},
>>> >> >>>>>>> >>> >>> > >>
'info':
>>> >> >>>>>>> >>> >>> > >>
{'HBAInventory':
>>> >> >>>>>>> >>> >>> > >>
{'iSCSI': [{'InitiatorName':
>>> >> >>>>>>> >>> >>> > >>
'iqn.1994-05.com.redhat:32151ce183c8'}],
>>> >> >>>>>>> >>> >>> > >>
'FC':
>>> >> >>>>>>> >>> >>> > >>
[]}, 'packages2': {'kernel': {'release':
>>> >> >>>>>>> >>> >>> > >>
'431.el6.x86_64',
>>> >> >>>>>>> >>> >>> > >>
'buildtime':
>>> >> >>>>>>> >>> >>> > >>
1385061309.0, 'version': '2.6.32'},
>>> 'glusterfs-rdma':
>>> >> >>>>>>> >>> >>> > >>
{'release':
>>> >> >>>>>>> >>> >>> > >>
'1.el6',
>>> >> >>>>>>> >>> >>> > >>
'buildtime': 1403622628L, 'version': '3.5.1'},
>>> >> >>>>>>> >>> >>> > >>
'glusterfs-fuse':
>>> >> >>>>>>> >>> >>> > >>
{'release': '1.el6', 'buildtime': 1403622628L,
>>> >> >>>>>>> >>> >>> > >>
'version':
>>> >> >>>>>>> >>> >>> > >>
'3.5.1'},
>>> >> >>>>>>> >>> >>> > >>
'spice-server': {'release': '6.el6_5.2',
>>> 'buildtime':
>>> >> >>>>>>> >>> >>> > >>
1402324637L,
>>> >> >>>>>>> >>> >>> > >>
'version': '0.12.4'}, 'vdsm': {'release':
>>> >> >>>>>>> >>> >>> > >>
'1.gitdb83943.el6',
>>> >> >>>>>>> >>> >>> > >>
'buildtime':
>>> >> >>>>>>> >>> >>> > >>
1412784567L, 'version': '4.16.7'}, 'qemu-kvm':
>>> >> >>>>>>> >>> >>> > >>
{'release':
>>> >> >>>>>>> >>> >>> > >>
'2.415.el6_5.10', 'buildtime': 1402435700L,
>>> >> >>>>>>> >>> >>> > >>
'version':
>>> >> >>>>>>> >>> >>> > >>
'0.12.1.2'},
>>> >> >>>>>>> >>> >>> > >>
'qemu-img': {'release': '2.415.el6_5.10',
>>> >> >>>>>>> >>> >>> > >>
'buildtime':
>>> >> >>>>>>> >>> >>> > >>
1402435700L,
>>> >> >>>>>>> >>> >>> > >>
'version': '0.12.1.2'}, 'libvirt': {'release':
>>> >> >>>>>>> >>> >>> > >>
'29.el6_5.9',
>>> >> >>>>>>> >>> >>> > >>
'buildtime':
>>> >> >>>>>>> >>> >>> > >>
1402404612L, 'version': '0.10.2'}, 'glusterfs':
>>> >> >>>>>>> >>> >>> > >>
{'release':
>>> >> >>>>>>> >>> >>> > >>
'1.el6',
>>> >> >>>>>>> >>> >>> > >>
'buildtime': 1403622628L, 'version': '3.5.1'},
>>> 'mom':
>>> >> >>>>>>> >>> >>> > >>
{'release':
>>> >> >>>>>>> >>> >>> > >>
'2.el6',
>>> >> >>>>>>> >>> >>> > >>
'buildtime': 1403794344L, 'version': '0.4.1'},
>>> >> >>>>>>> >>> >>> > >>
'glusterfs-server':
>>> >> >>>>>>> >>> >>> > >>
{'release': '1.el6', 'buildtime': 1403622628L,
>>> >> >>>>>>> >>> >>> > >>
'version':
>>> >> >>>>>>> >>> >>> > >>
'3.5.1'}},
>>> >> >>>>>>> >>> >>> > >>
'numaNodeDistance': {'1': [20, 10], '0': [10,
>>> 20]},
>>> >> >>>>>>> >>> >>> > >>
'cpuModel':
>>> >> >>>>>>> >>> >>> > >>
'Intel(R)
>>> >> >>>>>>> >>> >>> > >>
Xeon(R) CPU X5650 @ 2.67GHz',
>>> 'liveMerge':
>>> >> >>>>>>> >>> >>> > >>
'false',
>>> >> >>>>>>> >>> >>> > >>
'hooks':
>>> >> >>>>>>> >>> >>> > >>
{},
>>> >> >>>>>>> >>> >>> > >>
'cpuSockets': '2', 'vmTypes': ['kvm'],
>>> 'selinux':
>>> >> >>>>>>> >>> >>> > >>
{'mode': '1'},
>>> >> >>>>>>> >>> >>> > >>
'kdumpStatus': 0, 'supportedProtocols': ['2.2',
>>> >> >>>>>>> >>> >>> > >>
'2.3'],
>>> >> >>>>>>> >>> >>> > >>
'networks':
>>> >> >>>>>>> >>> >>> > >>
{'ovirtmgmt': {'iface': u'bond0.10', 'addr':
>>> >> >>>>>>> >>> >>> > >>
'43.252.176.16',
>>> >> >>>>>>> >>> >>> > >>
'bridged':
>>> >> >>>>>>> >>> >>> > >>
False, 'ipv6addrs':
>>> ['fe80::62eb:69ff:fe20:b46c/64'],
>>> >> >>>>>>> >>> >>> > >>
'mtu':
>>> >> >>>>>>> >>> >>> > >>
'1500',
>>> >> >>>>>>> >>> >>> > >>
'bootproto4': 'none', 'netmask':
>>> '255.255.255.0',
>>> >> >>>>>>> >>> >>> > >>
'ipv4addrs':
>>> >> >>>>>>> >>> >>> > >>
['
>>> >> >>>>>>> >>> >>> > >>
43.252.176.16/24' <
http://43.252.176.16/24%27
>>> >],
>>> >> >>>>>>> >>> >>> > >>
'interface':
>>> >> >>>>>>> >>> >>> > >>
u'bond0.10', 'ipv6gateway': '::', 'gateway':
>>> >> >>>>>>> >>> >>> > >>
'43.25.17.1'},
>>> >> >>>>>>> >>> >>> > >>
'Internal':
>>> >> >>>>>>> >>> >>> > >>
{'iface': 'Internal', 'addr': '', 'cfg':
>>> {'DEFROUTE':
>>> >> >>>>>>> >>> >>> > >>
'no',
>>> >> >>>>>>> >>> >>> > >>
'HOTPLUG':
>>> >> >>>>>>> >>> >>> > >>
'no', 'MTU': '9000', 'DELAY': '0',
>>> 'NM_CONTROLLED':
>>> >> >>>>>>> >>> >>> > >>
'no',
>>> >> >>>>>>> >>> >>> > >>
'BOOTPROTO':
>>> >> >>>>>>> >>> >>> > >>
'none', 'STP': 'off', 'DEVICE': 'Internal',
>>> 'TYPE':
>>> >> >>>>>>> >>> >>> > >>
'Bridge',
>>> >> >>>>>>> >>> >>> > >>
'ONBOOT':
>>> >> >>>>>>> >>> >>> > >>
'no'}, 'bridged': True, 'ipv6addrs':
>>> >> >>>>>>> >>> >>> > >>
['fe80::210:18ff:fecd:daac/64'],
>>> >> >>>>>>> >>> >>> > >>
'gateway': '', 'bootproto4': 'none', 'netmask':
>>> '',
>>> >> >>>>>>> >>> >>> > >>
'stp':
>>> >> >>>>>>> >>> >>> > >>
'off',
>>> >> >>>>>>> >>> >>> > >>
'ipv4addrs': [], 'mtu': '9000', 'ipv6gateway':
>>> '::',
>>> >> >>>>>>> >>> >>> > >>
'ports':
>>> >> >>>>>>> >>> >>> > >>
['bond1.100']}, 'storage': {'iface': u'bond1',
>>> >> >>>>>>> >>> >>> > >>
'addr':
>>> >> >>>>>>> >>> >>> > >>
'10.10.10.6',
>>> >> >>>>>>> >>> >>> > >>
'bridged': False, 'ipv6addrs':
>>> >> >>>>>>> >>> >>> > >>
['fe80::210:18ff:fecd:daac/64'],
>>> >> >>>>>>> >>> >>> > >>
'mtu':
>>> >> >>>>>>> >>> >>> > >>
'9000', 'bootproto4': 'none', 'netmask':
>>> >> >>>>>>> >>> >>> > >>
'255.255.255.0',
>>> >> >>>>>>> >>> >>> > >>
'ipv4addrs': ['
>>> >> >>>>>>> >>> >>> > >>
10.10.10.6/24' <
http://10.10.10.6/24%27>],
>>> >> >>>>>>> >>> >>> > >>
'interface':
>>> >> >>>>>>> >>> >>> > >>
u'bond1',
>>> >> >>>>>>> >>> >>> > >>
'ipv6gateway': '::', 'gateway': ''},
>>> 'VMNetwork':
>>> >> >>>>>>> >>> >>> > >>
{'iface':
>>> >> >>>>>>> >>> >>> > >>
'VMNetwork',
>>> >> >>>>>>> >>> >>> > >>
'addr': '', 'cfg': {'DEFROUTE': 'no',
'HOTPLUG':
>>> >> >>>>>>> >>> >>> > >>
'no',
>>> >> >>>>>>> >>> >>> > >>
'MTU':
>>> >> >>>>>>> >>> >>> > >>
'1500',
>>> >> >>>>>>> >>> >>> > >>
'DELAY': '0', 'NM_CONTROLLED': 'no',
>>> 'BOOTPROTO':
>>> >> >>>>>>> >>> >>> > >>
'none',
>>> >> >>>>>>> >>> >>> > >>
'STP':
>>> >> >>>>>>> >>> >>> > >>
'off',
>>> >> >>>>>>> >>> >>> > >>
'DEVICE': 'VMNetwork', 'TYPE': 'Bridge',
>>> 'ONBOOT':
>>> >> >>>>>>> >>> >>> > >>
'no'},
>>> >> >>>>>>> >>> >>> > >>
'bridged':
>>> >> >>>>>>> >>> >>> > >>
True,
>>> >> >>>>>>> >>> >>> > >>
'ipv6addrs': ['fe80::62eb:69ff:fe20:b46c/64'],
>>> >> >>>>>>> >>> >>> > >>
'gateway':
>>> >> >>>>>>> >>> >>> > >>
'',
>>> >> >>>>>>> >>> >>> > >>
'bootproto4':
>>> >> >>>>>>> >>> >>> > >>
'none', 'netmask': '', 'stp': 'off',
>>> 'ipv4addrs': [],
>>> >> >>>>>>> >>> >>> > >>
'mtu':
>>> >> >>>>>>> >>> >>> > >>
'1500',
>>> >> >>>>>>> >>> >>> > >>
'ipv6gateway': '::', 'ports': ['bond0.36']}},
>>> >> >>>>>>> >>> >>> > >>
'bridges':
>>> >> >>>>>>> >>> >>> > >>
{'Internal':
>>> >> >>>>>>> >>> >>> > >>
{'addr': '', 'cfg': {'DEFROUTE': 'no',
>>> 'HOTPLUG':
>>> >> >>>>>>> >>> >>> > >>
'no',
>>> >> >>>>>>> >>> >>> > >>
'MTU':
>>> >> >>>>>>> >>> >>> > >>
'9000',
>>> >> >>>>>>> >>> >>> > >>
'DELAY': '0', 'NM_CONTROLLED': 'no',
>>> 'BOOTPROTO':
>>> >> >>>>>>> >>> >>> > >>
'none',
>>> >> >>>>>>> >>> >>> > >>
'STP':
>>> >> >>>>>>> >>> >>> > >>
'off',
>>> >> >>>>>>> >>> >>> > >>
'DEVICE': 'Internal', 'TYPE': 'Bridge',
>>> 'ONBOOT':
>>> >> >>>>>>> >>> >>> > >>
'no'},
>>> >> >>>>>>> >>> >>> > >>
'ipv6addrs':
>>> >> >>>>>>> >>> >>> > >>
['fe80::210:18ff:fecd:daac/64'], 'mtu': '9000',
>>> >> >>>>>>> >>> >>> > >>
'netmask': '',
>>> >> >>>>>>> >>> >>> > >>
'stp':
>>> >> >>>>>>> >>> >>> > >>
'off', 'ipv4addrs': [], 'ipv6gateway': '::',
>>> >> >>>>>>> >>> >>> > >>
'gateway':
>>> >> >>>>>>> >>> >>> > >>
'',
>>> >> >>>>>>> >>> >>> > >>
'opts':
>>> >> >>>>>>> >>> >>> > >>
{'topology_change_detected': '0',
>>> >> >>>>>>> >>> >>> > >>
'multicast_last_member_count':
>>> >> >>>>>>> >>> >>> > >>
'2',
>>> >> >>>>>>> >>> >>> > >>
'hash_elasticity': '4',
>>> >> >>>>>>> >>> >>> > >>
'multicast_query_response_interval':
>>> >> >>>>>>> >>> >>> > >>
'999',
>>> >> >>>>>>> >>> >>> > >>
'multicast_snooping': '1',
>>> >> >>>>>>> >>> >>> > >>
'multicast_startup_query_interval':
>>> >> >>>>>>> >>> >>> > >>
'3124',
>>> >> >>>>>>> >>> >>> > >>
'hello_timer': '31',
>>> 'multicast_querier_interval':
>>> >> >>>>>>> >>> >>> > >>
'25496',
>>> >> >>>>>>> >>> >>> > >>
'max_age':
>>> >> >>>>>>> >>> >>> > >>
'1999', 'hash_max': '512', 'stp_state': '0',
>>> >> >>>>>>> >>> >>> > >>
'root_id':
>>> >> >>>>>>> >>> >>> > >>
'8000.001018cddaac', 'priority': '32768',
>>> >> >>>>>>> >>> >>> > >>
'multicast_membership_interval':
>>> >> >>>>>>> >>> >>> > >>
'25996', 'root_path_cost': '0', 'root_port':
>>> '0',
>>> >> >>>>>>> >>> >>> > >>
'multicast_querier':
>>> >> >>>>>>> >>> >>> > >>
'0',
>>> >> >>>>>>> >>> >>> > >>
'multicast_startup_query_count': '2',
>>> 'hello_time':
>>> >> >>>>>>> >>> >>> > >>
'199',
>>> >> >>>>>>> >>> >>> > >>
'topology_change': '0', 'bridge_id':
>>> >> >>>>>>> >>> >>> > >>
'8000.001018cddaac',
>>> >> >>>>>>> >>> >>> > >>
'topology_change_timer': '0', 'ageing_time':
>>> '29995',
>>> >> >>>>>>> >>> >>> > >>
'gc_timer':
>>> >> >>>>>>> >>> >>> > >>
'31',
>>> >> >>>>>>> >>> >>> > >>
'group_addr': '1:80:c2:0:0:0', 'tcn_timer': '0',
>>> >> >>>>>>> >>> >>> > >>
'multicast_query_interval': '12498',
>>> >> >>>>>>> >>> >>> > >>
'multicast_last_member_interval':
>>> >> >>>>>>> >>> >>> > >>
'99', 'multicast_router': '1', 'forward_delay':
>>> '0'},
>>> >> >>>>>>> >>> >>> > >>
'ports':
>>> >> >>>>>>> >>> >>> > >>
['bond1.100']}, 'VMNetwork': {'addr': '', 'cfg':
>>> >> >>>>>>> >>> >>> > >>
{'DEFROUTE':
>>> >> >>>>>>> >>> >>> > >>
'no',
>>> >> >>>>>>> >>> >>> > >>
'HOTPLUG': 'no', 'MTU': '1500', 'DELAY':
'0',
>>> >> >>>>>>> >>> >>> > >>
'NM_CONTROLLED':
>>> >> >>>>>>> >>> >>> > >>
'no',
>>> >> >>>>>>> >>> >>> > >>
'BOOTPROTO': 'none', 'STP': 'off', 'DEVICE':
>>> >> >>>>>>> >>> >>> > >>
'VMNetwork',
>>> >> >>>>>>> >>> >>> > >>
'TYPE':
>>> >> >>>>>>> >>> >>> > >>
'Bridge',
>>> >> >>>>>>> >>> >>> > >>
'ONBOOT': 'no'}, 'ipv6addrs':
>>> >> >>>>>>> >>> >>> > >>
['fe80::62eb:69ff:fe20:b46c/64'],
>>> >> >>>>>>> >>> >>> > >>
'mtu':
>>> >> >>>>>>> >>> >>> > >>
'1500', 'netmask': '', 'stp': 'off',
>>> 'ipv4addrs': [],
>>> >> >>>>>>> >>> >>> > >>
'ipv6gateway':
>>> >> >>>>>>> >>> >>> > >>
'::',
>>> >> >>>>>>> >>> >>> > >>
'gateway': '', 'opts':
>>> {'topology_change_detected':
>>> >> >>>>>>> >>> >>> > >>
'0',
>>> >> >>>>>>> >>> >>> > >>
'multicast_last_member_count': '2',
>>> >> >>>>>>> >>> >>> > >>
'hash_elasticity':
>>> >> >>>>>>> >>> >>> > >>
'4',
>>> >> >>>>>>> >>> >>> > >>
'multicast_query_response_interval': '999',
>>> >> >>>>>>> >>> >>> > >>
'multicast_snooping':
>>> >> >>>>>>> >>> >>> > >>
'1',
>>> >> >>>>>>> >>> >>> > >>
'multicast_startup_query_interval': '3124',
>>> >> >>>>>>> >>> >>> > >>
'hello_timer':
>>> >> >>>>>>> >>> >>> > >>
'131',
>>> >> >>>>>>> >>> >>> > >>
'multicast_querier_interval': '25496',
>>> 'max_age':
>>> >> >>>>>>> >>> >>> > >>
'1999',
>>> >> >>>>>>> >>> >>> > >>
'hash_max':
>>> >> >>>>>>> >>> >>> > >>
'512', 'stp_state': '0', 'root_id':
>>> >> >>>>>>> >>> >>> > >>
'8000.60eb6920b46c',
>>> >> >>>>>>> >>> >>> > >>
'priority':
>>> >> >>>>>>> >>> >>> > >>
'32768', 'multicast_membership_interval':
>>> '25996',
>>> >> >>>>>>> >>> >>> > >>
'root_path_cost':
>>> >> >>>>>>> >>> >>> > >>
'0',
>>> >> >>>>>>> >>> >>> > >>
'root_port': '0', 'multicast_querier': '0',
>>> >> >>>>>>> >>> >>> > >>
'multicast_startup_query_count': '2',
>>> 'hello_time':
>>> >> >>>>>>> >>> >>> > >>
'199',
>>> >> >>>>>>> >>> >>> > >>
'topology_change': '0', 'bridge_id':
>>> >> >>>>>>> >>> >>> > >>
'8000.60eb6920b46c',
>>> >> >>>>>>> >>> >>> > >>
'topology_change_timer': '0', 'ageing_time':
>>> '29995',
>>> >> >>>>>>> >>> >>> > >>
'gc_timer':
>>> >> >>>>>>> >>> >>> > >>
'31',
>>> >> >>>>>>> >>> >>> > >>
'group_addr': '1:80:c2:0:0:0', 'tcn_timer': '0',
>>> >> >>>>>>> >>> >>> > >>
'multicast_query_interval': '12498',
>>> >> >>>>>>> >>> >>> > >>
'multicast_last_member_interval':
>>> >> >>>>>>> >>> >>> > >>
'99', 'multicast_router': '1', 'forward_delay':
>>> '0'},
>>> >> >>>>>>> >>> >>> > >>
'ports':
>>> >> >>>>>>> >>> >>> > >>
['bond0.36']}}, 'uuid':
>>> >> >>>>>>> >>> >>> > >>
'44454C4C-4C00-1057-8053-B7C04F504E31',
>>> >> >>>>>>> >>> >>> > >>
'lastClientIface': 'bond1', 'nics': {'eth3':
>>> >> >>>>>>> >>> >>> > >>
{'permhwaddr':
>>> >> >>>>>>> >>> >>> > >>
'00:10:18:cd:da:ae', 'addr': '', 'cfg':
>>> {'SLAVE':
>>> >> >>>>>>> >>> >>> > >>
'yes',
>>> >> >>>>>>> >>> >>> > >>
'NM_CONTROLLED':
>>> >> >>>>>>> >>> >>> > >>
'no', 'MTU': '9000', 'HWADDR':
>>> '00:10:18:cd:da:ae',
>>> >> >>>>>>> >>> >>> > >>
'MASTER':
>>> >> >>>>>>> >>> >>> > >>
'bond1',
>>> >> >>>>>>> >>> >>> > >>
'DEVICE': 'eth3', 'ONBOOT': 'no'}, 'ipv6addrs':
>>> [],
>>> >> >>>>>>> >>> >>> > >>
'mtu':
>>> >> >>>>>>> >>> >>> > >>
'9000',
>>> >> >>>>>>> >>> >>> > >>
'netmask': '', 'ipv4addrs': [], 'hwaddr':
>>> >> >>>>>>> >>> >>> > >>
'00:10:18:cd:da:ac',
>>> >> >>>>>>> >>> >>> > >>
'speed':
>>> >> >>>>>>> >>> >>> > >>
1000}, 'eth2': {'permhwaddr':
>>> '00:10:18:cd:da:ac',
>>> >> >>>>>>> >>> >>> > >>
'addr': '',
>>> >> >>>>>>> >>> >>> > >>
'cfg':
>>> >> >>>>>>> >>> >>> > >>
{'SLAVE': 'yes', 'NM_CONTROLLED': 'no', 'MTU':
>>> >> >>>>>>> >>> >>> > >>
'9000',
>>> >> >>>>>>> >>> >>> > >>
'HWADDR':
>>> >> >>>>>>> >>> >>> > >>
'00:10:18:cd:da:ac', 'MASTER': 'bond1',
>>> 'DEVICE':
>>> >> >>>>>>> >>> >>> > >>
'eth2',
>>> >> >>>>>>> >>> >>> > >>
'ONBOOT':
>>> >> >>>>>>> >>> >>> > >>
'no'},
>>> >> >>>>>>> >>> >>> > >>
'ipv6addrs': [], 'mtu': '9000', 'netmask': '',
>>> >> >>>>>>> >>> >>> > >>
'ipv4addrs': [],
>>> >> >>>>>>> >>> >>> > >>
'hwaddr':
>>> >> >>>>>>> >>> >>> > >>
'00:10:18:cd:da:ac', 'speed': 1000}, 'eth1':
>>> >> >>>>>>> >>> >>> > >>
{'permhwaddr':
>>> >> >>>>>>> >>> >>> > >>
'60:eb:69:20:b4:6d', 'addr': '', 'cfg':
>>> {'SLAVE':
>>> >> >>>>>>> >>> >>> > >>
'yes',
>>> >> >>>>>>> >>> >>> > >>
'NM_CONTROLLED':
>>> >> >>>>>>> >>> >>> > >>
'no', 'MTU': '1500', 'HWADDR':
>>> '60:eb:69:20:b4:6d',
>>> >> >>>>>>> >>> >>> > >>
'MASTER':
>>> >> >>>>>>> >>> >>> > >>
'bond0',
>>> >> >>>>>>> >>> >>> > >>
'DEVICE': 'eth1', 'ONBOOT': 'yes'},
>>> 'ipv6addrs': [],
>>> >> >>>>>>> >>> >>> > >>
'mtu':
>>> >> >>>>>>> >>> >>> > >>
'1500',
>>> >> >>>>>>> >>> >>> > >>
'netmask': '', 'ipv4addrs': [], 'hwaddr':
>>> >> >>>>>>> >>> >>> > >>
'60:eb:69:20:b4:6c',
>>> >> >>>>>>> >>> >>> > >>
'speed':
>>> >> >>>>>>> >>> >>> > >>
1000}, 'eth0': {'permhwaddr':
>>> '60:eb:69:20:b4:6c',
>>> >> >>>>>>> >>> >>> > >>
'addr': '',
>>> >> >>>>>>> >>> >>> > >>
'cfg':
>>> >> >>>>>>> >>> >>> > >>
{'SLAVE': 'yes', 'NM_CONTROLLED': 'no', 'MTU':
>>> >> >>>>>>> >>> >>> > >>
'1500',
>>> >> >>>>>>> >>> >>> > >>
'HWADDR':
>>> >> >>>>>>> >>> >>> > >>
'60:eb:69:20:b4:6c', 'MASTER': 'bond0',
>>> 'DEVICE':
>>> >> >>>>>>> >>> >>> > >>
'eth0',
>>> >> >>>>>>> >>> >>> > >>
'ONBOOT':
>>> >> >>>>>>> >>> >>> > >>
'yes'},
>>> >> >>>>>>> >>> >>> > >>
'ipv6addrs': [], 'mtu': '1500', 'netmask': '',
>>> >> >>>>>>> >>> >>> > >>
'ipv4addrs': [],
>>> >> >>>>>>> >>> >>> > >>
'hwaddr':
>>> >> >>>>>>> >>> >>> > >>
'60:eb:69:20:b4:6c', 'speed': 1000}},
>>> >> >>>>>>> >>> >>> > >>
'software_revision': '1',
>>> >> >>>>>>> >>> >>> > >>
'clusterLevels': ['3.0', '3.1', '3.2', '3.3',
>>> '3.4',
>>> >> >>>>>>> >>> >>> > >>
'3.5'],
>>> >> >>>>>>> >>> >>> > >>
'cpuFlags':
>>> >> >>>>>>> >>> >>> >
>>
>>> >> >>>>>>> >>> >>> >
>>
>>> >> >>>>>>> >>> >>> >
>>
>>> >> >>>>>>> >>> >>> >
>>
>>> >> >>>>>>> >>> >>> >
>>
>>>
u'fpu,vme,de,pse,tsc,msr,pae,mce,cx8,apic,sep,mtrr,pge,mca,cmov,pat,pse36,clflush,dts,acpi,mmx,fxsr,sse,sse2,ss,ht,tm,pbe,syscall,nx,pdpe1gb,rdtscp,lm,constant_tsc,arch_perfmon,pebs,bts,rep_good,xtopology,nonstop_tsc,pni,pclmulqdq,dtes64,monitor,ds_cpl,vmx,smx,est,tm2,ssse3,cx16,xtpr,pdcm,pcid,dca,sse4_1,sse4_2,popcnt,aes,lahf_lm,tpr_shadow,vnmi,flexpriority,ept,vpid,model_Nehalem,model_Conroe,model_coreduo,model_core2duo,model_Penryn,model_Westmere,model_n270',
>>> >> >>>>>>> >>> >>> > >>
'ISCSIInitiatorName':
>>> >> >>>>>>> >>> >>> > >>
'iqn.1994-05.com.redhat:32151ce183c8',
>>> >> >>>>>>> >>> >>> > >>
'netConfigDirty': 'False', 'supportedENGINEs':
>>> >> >>>>>>> >>> >>> > >>
['3.0',
>>> >> >>>>>>> >>> >>> > >>
'3.1',
>>> >> >>>>>>> >>> >>> > >>
'3.2',
>>> >> >>>>>>> >>> >>> > >>
'3.3',
>>> >> >>>>>>> >>> >>> > >>
'3.4', '3.5'], 'autoNumaBalancing': 2,
>>> 'reservedMem':
>>> >> >>>>>>> >>> >>> > >>
'321',
>>> >> >>>>>>> >>> >>> > >>
'bondings':
>>> >> >>>>>>> >>> >>> > >>
{'bond4': {'addr': '', 'cfg': {}, 'mtu':
'1500',
>>> >> >>>>>>> >>> >>> > >>
'netmask': '',
>>> >> >>>>>>> >>> >>> > >>
'slaves':
>>> >> >>>>>>> >>> >>> > >>
[], 'hwaddr': '00:00:00:00:00:00'}, 'bond0':
>>> {'addr':
>>> >> >>>>>>> >>> >>> > >>
'',
>>> >> >>>>>>> >>> >>> > >>
'cfg':
>>> >> >>>>>>> >>> >>> > >>
{'HOTPLUG': 'no', 'MTU': '1500',
>>> 'NM_CONTROLLED':
>>> >> >>>>>>> >>> >>> > >>
'no',
>>> >> >>>>>>> >>> >>> > >>
'BONDING_OPTS':
>>> >> >>>>>>> >>> >>> > >>
'mode=4 miimon=100', 'DEVICE': 'bond0',
>>> 'ONBOOT':
>>> >> >>>>>>> >>> >>> > >>
'yes'},
>>> >> >>>>>>> >>> >>> > >>
'ipv6addrs':
>>> >> >>>>>>> >>> >>> > >>
['fe80::62eb:69ff:fe20:b46c/64'], 'mtu': '1500',
>>> >> >>>>>>> >>> >>> > >>
'netmask': '',
>>> >> >>>>>>> >>> >>> > >>
'ipv4addrs': [], 'hwaddr': '60:eb:69:20:b4:6c',
>>> >> >>>>>>> >>> >>> > >>
'slaves':
>>> >> >>>>>>> >>> >>> > >>
['eth0',
>>> >> >>>>>>> >>> >>> > >>
'eth1'],
>>> >> >>>>>>> >>> >>> > >>
'opts': {'miimon': '100', 'mode': '4'}},
>>> 'bond1':
>>> >> >>>>>>> >>> >>> > >>
{'addr':
>>> >> >>>>>>> >>> >>> > >>
'10.10.10.6',
>>> >> >>>>>>> >>> >>> > >>
'cfg': {'DEFROUTE': 'no', 'IPADDR':
>>> '10.10.10.6',
>>> >> >>>>>>> >>> >>> > >>
'HOTPLUG':
>>> >> >>>>>>> >>> >>> > >>
'no',
>>> >> >>>>>>> >>> >>> > >>
'MTU':
>>> >> >>>>>>> >>> >>> > >>
'9000', 'NM_CONTROLLED': 'no', 'NETMASK':
>>> >> >>>>>>> >>> >>> > >>
'255.255.255.0',
>>> >> >>>>>>> >>> >>> > >>
'BOOTPROTO':
>>> >> >>>>>>> >>> >>> > >>
'none', 'BONDING_OPTS': 'mode=4 miimon=100',
>>> >> >>>>>>> >>> >>> > >>
'DEVICE':
>>> >> >>>>>>> >>> >>> > >>
'bond1',
>>> >> >>>>>>> >>> >>> > >>
'ONBOOT':
>>> >> >>>>>>> >>> >>> > >>
'no'}, 'ipv6addrs':
>>> ['fe80::210:18ff:fecd:daac/64'],
>>> >> >>>>>>> >>> >>> > >>
'mtu':
>>> >> >>>>>>> >>> >>> > >>
'9000',
>>> >> >>>>>>> >>> >>> > >>
'netmask': '255.255.255.0', 'ipv4addrs':
>>> >> >>>>>>> >>> >>> > >>
['10.10.10.6/24'
>>> >> >>>>>>> >>> >>> > >>
<
http://10.10.10.6/24%27>], 'hwaddr':
>>> >> >>>>>>> >>> >>> > >>
'00:10:18:cd:da:ac',
>>> >> >>>>>>> >>> >>> > >>
'slaves':
>>> >> >>>>>>> >>> >>> > >>
['eth2', 'eth3'], 'opts': {'miimon': '100',
>>> 'mode':
>>> >> >>>>>>> >>> >>> > >>
'4'}},
>>> >> >>>>>>> >>> >>> > >>
'bond2':
>>> >> >>>>>>> >>> >>> > >>
{'addr': '', 'cfg': {}, 'mtu': '1500',
>>> 'netmask': '',
>>> >> >>>>>>> >>> >>> > >>
'slaves':
>>> >> >>>>>>> >>> >>> > >>
[],
>>> >> >>>>>>> >>> >>> > >>
'hwaddr': '00:00:00:00:00:00'}, 'bond3':
>>> {'addr': '',
>>> >> >>>>>>> >>> >>> > >>
'cfg': {},
>>> >> >>>>>>> >>> >>> > >>
'mtu':
>>> >> >>>>>>> >>> >>> > >>
'1500', 'netmask': '', 'slaves': [], 'hwaddr':
>>> >> >>>>>>> >>> >>> > >>
'00:00:00:00:00:00'}},
>>> >> >>>>>>> >>> >>> > >>
'software_version': '4.16', 'memSize': '24019',
>>> >> >>>>>>> >>> >>> > >>
'cpuSpeed':
>>> >> >>>>>>> >>> >>> > >>
'2667.000',
>>> >> >>>>>>> >>> >>> > >>
'numaNodes': {u'1': {'totalMemory': '12288',
>>> 'cpus':
>>> >> >>>>>>> >>> >>> > >>
[6,
>>> >> >>>>>>> >>> >>> > >>
7, 8,
>>> >> >>>>>>> >>> >>> > >>
9,
>>> >> >>>>>>> >>> >>> > >>
10, 11,
>>> >> >>>>>>> >>> >>> > >>
18, 19, 20, 21, 22, 23]}, u'0': {'totalMemory':
>>> >> >>>>>>> >>> >>> > >>
'12278',
>>> >> >>>>>>> >>> >>> > >>
'cpus':
>>> >> >>>>>>> >>> >>> > >>
[0,
>>> >> >>>>>>> >>> >>> > >>
1, 2,
>>> >> >>>>>>> >>> >>> > >>
3, 4, 5, 12, 13, 14, 15, 16, 17]}},
>>> 'version_name':
>>> >> >>>>>>> >>> >>> > >>
'Snow
>>> >> >>>>>>> >>> >>> > >>
Man',
>>> >> >>>>>>> >>> >>> > >>
'vlans':
>>> >> >>>>>>> >>> >>> > >>
{'bond0.10': {'iface': 'bond0', 'addr':
>>> >> >>>>>>> >>> >>> > >>
'43.25.17.16',
>>> >> >>>>>>> >>> >>> > >>
'cfg':
>>> >> >>>>>>> >>> >>> > >>
{'DEFROUTE':
>>> >> >>>>>>> >>> >>> > >>
'yes', 'VLAN': 'yes', 'IPADDR': '43.25.17.16',
>>> >> >>>>>>> >>> >>> > >>
'HOTPLUG':
>>> >> >>>>>>> >>> >>> > >>
'no',
>>> >> >>>>>>> >>> >>> > >>
'GATEWAY':
>>> >> >>>>>>> >>> >>> > >>
'43.25.17.1', 'NM_CONTROLLED': 'no', 'NETMASK':
>>> >> >>>>>>> >>> >>> > >>
'255.255.255.0',
>>> >> >>>>>>> >>> >>> > >>
'BOOTPROTO': 'none', 'DEVICE': 'bond0.10',
>>> 'MTU':
>>> >> >>>>>>> >>> >>> > >>
'1500',
>>> >> >>>>>>> >>> >>> > >>
'ONBOOT':
>>> >> >>>>>>> >>> >>> > >>
'yes'},
>>> >> >>>>>>> >>> >>> > >>
'ipv6addrs': ['fe80::62eb:69ff:fe20:b46c/64'],
>>> >> >>>>>>> >>> >>> > >>
'vlanid':
>>> >> >>>>>>> >>> >>> > >>
10,
>>> >> >>>>>>> >>> >>> > >>
'mtu':
>>> >> >>>>>>> >>> >>> > >>
'1500',
>>> >> >>>>>>> >>> >>> > >>
'netmask': '255.255.255.0', 'ipv4addrs':
>>> >> >>>>>>> >>> >>> > >>
['43.25.17.16/24']
>>> >> >>>>>>> >>> >>> > >>
<
http://43.25.17.16/24%27%5D>}, 'bond0.36':
>>> {'iface':
>>> >> >>>>>>> >>> >>> > >>
'bond0',
>>> >> >>>>>>> >>> >>> > >>
'addr':
>>> >> >>>>>>> >>> >>> > >>
'', 'cfg': {'BRIDGE': 'VMNetwork', 'VLAN':
>>> 'yes',
>>> >> >>>>>>> >>> >>> > >>
'HOTPLUG':
>>> >> >>>>>>> >>> >>> > >>
'no',
>>> >> >>>>>>> >>> >>> > >>
'MTU':
>>> >> >>>>>>> >>> >>> > >>
'1500', 'NM_CONTROLLED': 'no', 'DEVICE':
>>> 'bond0.36',
>>> >> >>>>>>> >>> >>> > >>
'ONBOOT':
>>> >> >>>>>>> >>> >>> > >>
'no'},
>>> >> >>>>>>> >>> >>> > >>
'ipv6addrs': ['fe80::62eb:69ff:fe20:b46c/64'],
>>> >> >>>>>>> >>> >>> > >>
'vlanid':
>>> >> >>>>>>> >>> >>> > >>
36,
>>> >> >>>>>>> >>> >>> > >>
'mtu':
>>> >> >>>>>>> >>> >>> > >>
'1500',
>>> >> >>>>>>> >>> >>> > >>
'netmask': '', 'ipv4addrs': []}, 'bond1.100':
>>> >> >>>>>>> >>> >>> > >>
{'iface':
>>> >> >>>>>>> >>> >>> > >>
'bond1',
>>> >> >>>>>>> >>> >>> > >>
'addr':
>>> >> >>>>>>> >>> >>> > >>
'', 'cfg': {'BRIDGE': 'Internal', 'VLAN':
'yes',
>>> >> >>>>>>> >>> >>> > >>
'HOTPLUG':
>>> >> >>>>>>> >>> >>> > >>
'no',
>>> >> >>>>>>> >>> >>> > >>
'MTU':
>>> >> >>>>>>> >>> >>> > >>
'9000', 'NM_CONTROLLED': 'no', 'DEVICE':
>>> 'bond1.100',
>>> >> >>>>>>> >>> >>> > >>
'ONBOOT':
>>> >> >>>>>>> >>> >>> > >>
'no'},
>>> >> >>>>>>> >>> >>> > >>
'ipv6addrs': ['fe80::210:18ff:fecd:daac/64'],
>>> >> >>>>>>> >>> >>> > >>
'vlanid':
>>> >> >>>>>>> >>> >>> > >>
100,
>>> >> >>>>>>> >>> >>> > >>
'mtu':
>>> >> >>>>>>> >>> >>> > >>
'9000',
>>> >> >>>>>>> >>> >>> > >>
'netmask': '', 'ipv4addrs': []}}, 'cpuCores':
>>> '12',
>>> >> >>>>>>> >>> >>> > >>
'kvmEnabled':
>>> >> >>>>>>> >>> >>> > >>
'true',
>>> >> >>>>>>> >>> >>> > >>
'guestOverhead': '65', 'cpuThreads': '24',
>>> >> >>>>>>> >>> >>> > >>
'emulatedMachines':
>>> >> >>>>>>> >>> >>> > >>
[u'rhel6.5.0', u'pc', u'rhel6.4.0',
>>> u'rhel6.3.0',
>>> >> >>>>>>> >>> >>> > >>
u'rhel6.2.0',
>>> >> >>>>>>> >>> >>> > >>
u'rhel6.1.0', u'rhel6.0.0', u'rhel5.5.0',
>>> >> >>>>>>> >>> >>> > >>
u'rhel5.4.4',
>>> >> >>>>>>> >>> >>> > >>
u'rhel5.4.0'],
>>> >> >>>>>>> >>> >>> > >>
'operatingSystem': {'release':
>>> '5.el6.centos.11.1',
>>> >> >>>>>>> >>> >>> > >>
'version':
>>> >> >>>>>>> >>> >>> > >>
'6',
>>> >> >>>>>>> >>> >>> > >>
'name':
>>> >> >>>>>>> >>> >>> > >>
'RHEL'}, 'lastClient': '10.10.10.2'}}
>>> >> >>>>>>> >>> >>> > >>
Thread-13::DEBUG::2014-11-24
>>> >> >>>>>>> >>> >>> >
>>
>>> 21:41:41,620::BindingXMLRPC::1132::vds::(wrapper)
>>> >> >>>>>>> >>> >>> > >>
client
>>> >> >>>>>>> >>> >>> > >>
[10.10.10.2]::call
>>> >> >>>>>>> >>> >>> > >>
getHardwareInfo with () {}
>>> >> >>>>>>> >>> >>> > >>
Thread-13::DEBUG::2014-11-24
>>> >> >>>>>>> >>> >>> >
>>
>>> 21:41:41,621::BindingXMLRPC::1139::vds::(wrapper)
>>> >> >>>>>>> >>> >>> > >>
return
>>> >> >>>>>>> >>> >>> > >>
getHardwareInfo
>>> >> >>>>>>> >>> >>> > >>
with {'status': {'message': 'Done', 'code': 0},
>>> >> >>>>>>> >>> >>> > >>
'info':
>>> >> >>>>>>> >>> >>> > >>
{'systemProductName': 'CS24-TY',
>>> >> >>>>>>> >>> >>> > >>
'systemSerialNumber':
>>> >> >>>>>>> >>> >>> > >>
'7LWSPN1',
>>> >> >>>>>>> >>> >>> > >>
'systemFamily': 'Server', 'systemVersion':
>>> 'A00',
>>> >> >>>>>>> >>> >>> > >>
'systemUUID':
>>> >> >>>>>>> >>> >>> > >>
'44454c4c-4c00-1057-8053-b7c04f504e31',
>>> >> >>>>>>> >>> >>> > >>
'systemManufacturer':
>>> >> >>>>>>> >>> >>> > >>
'Dell'}}
>>> >> >>>>>>> >>> >>> > >>
Thread-13::DEBUG::2014-11-24
>>> >> >>>>>>> >>> >>> >
>>
>>> 21:41:41,733::BindingXMLRPC::1132::vds::(wrapper)
>>> >> >>>>>>> >>> >>> > >>
client
>>> >> >>>>>>> >>> >>> > >>
[10.10.10.2]::call
>>> >> >>>>>>> >>> >>> > >>
hostsList with () {} flowID [222e8036]
>>> >> >>>>>>> >>> >>> > >>
Thread-13::ERROR::2014-11-24
>>> >> >>>>>>> >>> >>> >
>>
>>> 21:41:44,753::BindingXMLRPC::1148::vds::(wrapper)
>>> >> >>>>>>> >>> >>> > >>
vdsm
>>> >> >>>>>>> >>> >>> > >>
exception
>>> >> >>>>>>> >>> >>> > >>
occured
>>> >> >>>>>>> >>> >>> > >>
Traceback (most recent call last):
>>> >> >>>>>>> >>> >>> > >>
File "/usr/share/vdsm/rpc/BindingXMLRPC.py",
>>> line
>>> >> >>>>>>> >>> >>> > >>
1135,
>>> >> >>>>>>> >>> >>> > >>
in
>>> >> >>>>>>> >>> >>> > >>
wrapper
>>> >> >>>>>>> >>> >>> > >>
res = f(*args, **kwargs)
>>> >> >>>>>>> >>> >>> > >>
File "/usr/share/vdsm/gluster/api.py", line
>>> 54, in
>>> >> >>>>>>> >>> >>> > >>
wrapper
>>> >> >>>>>>> >>> >>> > >>
rv = func(*args, **kwargs)
>>> >> >>>>>>> >>> >>> > >>
File "/usr/share/vdsm/gluster/api.py", line
>>> 251, in
>>> >> >>>>>>> >>> >>> > >>
hostsList
>>> >> >>>>>>> >>> >>> > >>
return {'hosts':
>>> >> >>>>>>> >>> >>> > >>
self.svdsmProxy.glusterPeerStatus()}
>>> >> >>>>>>> >>> >>> > >>
File "/usr/share/vdsm/supervdsm.py", line 50,
>>> in
>>> >> >>>>>>> >>> >>> > >>
__call__
>>> >> >>>>>>> >>> >>> > >>
return callMethod()
>>> >> >>>>>>> >>> >>> > >>
File "/usr/share/vdsm/supervdsm.py", line 48,
>>> in
>>> >> >>>>>>> >>> >>> > >>
<lambda>
>>> >> >>>>>>> >>> >>> > >>
**kwargs)
>>> >> >>>>>>> >>> >>> > >>
File "<string>", line 2, in glusterPeerStatus
>>> >> >>>>>>> >>> >>> > >>
File
>>> >> >>>>>>> >>> >>> >
>>
>>> "/usr/lib64/python2.6/multiprocessing/managers.py",
>>> >> >>>>>>> >>> >>> > >>
line
>>> >> >>>>>>> >>> >>> > >>
740,
>>> >> >>>>>>> >>> >>> > >>
in
>>> >> >>>>>>> >>> >>> > >>
_callmethod
>>> >> >>>>>>> >>> >>> > >>
raise convert_to_error(kind, result)
>>> >> >>>>>>> >>> >>> > >>
GlusterCmdExecFailedException: Command execution
>>> >> >>>>>>> >>> >>> > >>
failed
>>> >> >>>>>>> >>> >>> > >>
error: Connection failed. Please check if
>>> gluster
>>> >> >>>>>>> >>> >>> > >>
daemon
>>> >> >>>>>>> >>> >>> > >>
is
>>> >> >>>>>>> >>> >>> > >>
operational.
>>> >> >>>>>>> >>> >>> > >>
return code: 1
>>> >> >>>>>>> >>> >>> > >>
Thread-13::DEBUG::2014-11-24
>>> >> >>>>>>> >>> >>> >
>>
>>> >> >>>>>>> >>> >>> >
>>
>>> >> >>>>>>> >>> >>> >
>>
>>> >> >>>>>>> >>> >>> >
>>
>>> 21:41:50,949::task::595::Storage.TaskManager.Task::(_updateState)
>>> >> >>>>>>> >>> >>> >
>>
>>> Task=`c9042986-c978-4b08-adb2-616f5299e115`::moving
>>> >> >>>>>>> >>> >>> > >>
from
>>> >> >>>>>>> >>> >>> > >>
state
>>> >> >>>>>>> >>> >>> > >>
init
>>> >> >>>>>>> >>> >>> > >>
->
>>> >> >>>>>>> >>> >>> > >>
state preparing
>>> >> >>>>>>> >>> >>> > >>
Thread-13::INFO::2014-11-24
>>> >> >>>>>>> >>> >>> >
>>
>>> 21:41:50,950::logUtils::44::dispatcher::(wrapper) Run
>>> >> >>>>>>> >>> >>> > >>
and
>>> >> >>>>>>> >>> >>> > >>
protect:
>>> >> >>>>>>> >>> >>> > >>
repoStats(options=None)
>>> >> >>>>>>> >>> >>> > >>
Thread-13::INFO::2014-11-24
>>> >> >>>>>>> >>> >>> >
>>
>>> 21:41:50,950::logUtils::47::dispatcher::(wrapper) Run
>>> >> >>>>>>> >>> >>> > >>
and
>>> >> >>>>>>> >>> >>> > >>
protect:
>>> >> >>>>>>> >>> >>> > >>
repoStats, Return response: {}
>>> >> >>>>>>> >>> >>> > >>
Thread-13::DEBUG::2014-11-24
>>> >> >>>>>>> >>> >>> >
>>
>>> >> >>>>>>> >>> >>> >
>>
>>> >> >>>>>>> >>> >>> >
>>
>>> 21:41:50,950::task::1191::Storage.TaskManager.Task::(prepare)
>>> >> >>>>>>> >>> >>> >
>>
>>> >> >>>>>>> >>> >>> >
>>
>>> Task=`c9042986-c978-4b08-adb2-616f5299e115`::finished: {}
>>> >> >>>>>>> >>> >>> > >>
Thread-13::DEBUG::2014-11-24
>>> >> >>>>>>> >>> >>> >
>>
>>> >> >>>>>>> >>> >>> >
>>
>>> >> >>>>>>> >>> >>> >
>>
>>> >> >>>>>>> >>> >>> >
>>
>>> 21:41:50,950::task::595::Storage.TaskManager.Task::(_updateState)
>>> >> >>>>>>> >>> >>> >
>>
>>> Task=`c9042986-c978-4b08-adb2-616f5299e115`::moving
>>> >> >>>>>>> >>> >>> > >>
from
>>> >> >>>>>>> >>> >>> > >>
state
>>> >> >>>>>>> >>> >>> > >>
preparing
>>> >> >>>>>>> >>> >>> > >>
->
>>> >> >>>>>>> >>> >>> > >>
state finished
>>> >> >>>>>>> >>> >>> > >>
Thread-13::DEBUG::2014-11-24
>>> >> >>>>>>> >>> >>> >
>>
>>> >> >>>>>>> >>> >>> >
>>
>>> >> >>>>>>> >>> >>> >
>>
>>> >> >>>>>>> >>> >>> >
>>
>>> >> >>>>>>> >>> >>> >
>>
>>>
21:41:50,951::resourceManager::940::Storage.ResourceManager.Owner::(releaseAll)
>>> >> >>>>>>> >>> >>> > >>
Owner.releaseAll requests {} resources {}
>>> >> >>>>>>> >>> >>> > >>
Thread-13::DEBUG::2014-11-24
>>> >> >>>>>>> >>> >>> >
>>
>>> >> >>>>>>> >>> >>> >
>>
>>> >> >>>>>>> >>> >>> >
>>
>>> >> >>>>>>> >>> >>> >
>>
>>> >> >>>>>>> >>> >>> >
>>
>>>
21:41:50,951::resourceManager::977::Storage.ResourceManager.Owner::(cancelAll)
>>> >> >>>>>>> >>> >>> > >>
Owner.cancelAll requests {}
>>> >> >>>>>>> >>> >>> > >>
Thread-13::DEBUG::2014-11-24
>>> >> >>>>>>> >>> >>> >
>>
>>> >> >>>>>>> >>> >>> >
>>
>>> >> >>>>>>> >>> >>> >
>>
>>> 21:41:50,951::task::993::Storage.TaskManager.Task::(_decref)
>>> >> >>>>>>> >>> >>> >
>>
>>> Task=`c9042986-c978-4b08-adb2-616f5299e115`::ref 0
>>> >> >>>>>>> >>> >>> > >>
aborting
>>> >> >>>>>>> >>> >>> > >>
False
>>> >> >>>>>>> >>> >>> > >>
-------------------------------
>>> >> >>>>>>> >>> >>> >
>>
>>> >> >>>>>>> >>> >>> > >>
[root@compute4 ~]# service glusterd status
>>> >> >>>>>>> >>> >>> > >>
glusterd is stopped
>>> >> >>>>>>> >>> >>> > >>
[root@compute4 ~]# chkconfig --list | grep
>>> glusterd
>>> >> >>>>>>> >>> >>> > >>
glusterd 0:off 1:off 2:on 3:on
>>> 4:on
>>> >> >>>>>>> >>> >>> > >>
5:on
>>> >> >>>>>>> >>> >>> > >>
6:off
>>> >> >>>>>>> >>> >>> > >>
[root@compute4 ~]#
>>> >> >>>>>>> >>> >>> >
>>
>>> >> >>>>>>> >>> >>> > >>
Thanks,
>>> >> >>>>>>> >>> >>> > >>
Punit
>>> >> >>>>>>> >>> >>> >
>>
>>> >> >>>>>>> >>> >>> > >>
On Mon, Nov 24, 2014 at 6:36 PM, Kanagaraj
>>> >> >>>>>>> >>> >>> > >>
<kmayilsa(a)redhat.com>
>>> >> >>>>>>> >>> >>> > >>
wrote:
>>> >> >>>>>>> >>> >>> >
>>
>>> >> >>>>>>> >>> >>> >
>>> Can you send the corresponding error in
>>> vdsm.log
>>> >> >>>>>>> >>> >>> >
>>> from
>>> >> >>>>>>> >>> >>> >
>>> the
>>> >> >>>>>>> >>> >>> >
>>> host?
>>> >> >>>>>>> >>> >>> >
>>>
>>> >> >>>>>>> >>> >>> >
>>> Also check if glusterd service is running.
>>> >> >>>>>>> >>> >>> >
>>>
>>> >> >>>>>>> >>> >>> >
>>> Thanks,
>>> >> >>>>>>> >>> >>> >
>>> Kanagaraj
>>> >> >>>>>>> >>> >>> >
>>>
>>> >> >>>>>>> >>> >>> >
>>>
>>> >> >>>>>>> >>> >>> >
>>> On 11/24/2014 03:39 PM, Punit Dambiwal wrote:
>>> >> >>>>>>> >>> >>> >
>>>
>>> >> >>>>>>> >>> >>> >
>>> Hi,
>>> >> >>>>>>> >>> >>> >
>>>
>>> >> >>>>>>> >>> >>> >
>>> After reboot my Hypervisior host can not
>>> activate
>>> >> >>>>>>> >>> >>> >
>>> again
>>> >> >>>>>>> >>> >>> >
>>> in the
>>> >> >>>>>>> >>> >>> >
>>> cluster
>>> >> >>>>>>> >>> >>> >
>>> and failed with the following error :-
>>> >> >>>>>>> >>> >>> >
>>>
>>> >> >>>>>>> >>> >>> >
>>> Gluster command [<UNKNOWN>] failed on
>>> server...
>>> >> >>>>>>> >>> >>> >
>>>
>>> >> >>>>>>> >>> >>> >
>>> Engine logs :-
>>> >> >>>>>>> >>> >>> >
>>>
>>> >> >>>>>>> >>> >>> >
>>> 2014-11-24 18:05:28,397 INFO
>>> >> >>>>>>> >>> >>> >
>>>
>>> >> >>>>>>> >>> >>> >
>>>
>>> >> >>>>>>> >>> >>> >
>>>
>>> >> >>>>>>> >>> >>> >
>>>
>>> >> >>>>>>> >>> >>> >
>>>
>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand]
>>> >> >>>>>>> >>> >>> >
>>> (DefaultQuartzScheduler_Worker-64) START,
>>> >> >>>>>>> >>> >>> >
>>> GlusterVolumesListVDSCommand(HostName =
>>> Compute4,
>>> >> >>>>>>> >>> >>> >
>>> HostId
>>> >> >>>>>>> >>> >>> >
>>> =
>>> >> >>>>>>> >>> >>> >
>>> 33648a90-200c-45ca-89d5-1ce305d79a6a), log id:
>>> >> >>>>>>> >>> >>> >
>>> 5f251c90
>>> >> >>>>>>> >>> >>> >
>>> 2014-11-24 18:05:30,609 INFO
>>> >> >>>>>>> >>> >>> >
>>>
>>> >> >>>>>>> >>> >>> >
>>>
>>> >> >>>>>>> >>> >>> >
>>>
>>> >> >>>>>>> >>> >>> >
>>>
>>> >> >>>>>>> >>> >>> >
>>>
>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand]
>>> >> >>>>>>> >>> >>> >
>>> (DefaultQuartzScheduler_Worker-64) FINISH,
>>> >> >>>>>>> >>> >>> >
>>> GlusterVolumesListVDSCommand,
>>> >> >>>>>>> >>> >>> >
>>> return:
>>> >> >>>>>>> >>> >>> >
>>>
>>> >> >>>>>>> >>> >>> >
>>>
>>> >> >>>>>>> >>> >>> >
>>>
>>> >> >>>>>>> >>> >>> >
>>>
>>> >> >>>>>>> >>> >>> >
>>>
>>>
{26ae1672-ee09-4a38-8fd2-72dd9974cc2b=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@d95203e0
>>> },
>>> >> >>>>>>> >>> >>> >
>>> log id: 5f251c90
>>> >> >>>>>>> >>> >>> >
>>> 2014-11-24 18:05:33,768 INFO
>>> >> >>>>>>> >>> >>> >
>>> [org.ovirt.engine.core.bll.ActivateVdsCommand]
>>> >> >>>>>>> >>> >>> >
>>> (ajp--127.0.0.1-8702-8)
>>> >> >>>>>>> >>> >>> >
>>> [287d570d] Lock Acquired to object EngineLock
>>> >> >>>>>>> >>> >>> >
>>> [exclusiveLocks=
>>> >> >>>>>>> >>> >>> >
>>> key:
>>> >> >>>>>>> >>> >>> >
>>> 0bf6b00f-7947-4411-b55a-cc5eea2b381a value: VDS
>>> >> >>>>>>> >>> >>> >
>>> , sharedLocks= ]
>>> >> >>>>>>> >>> >>> >
>>> 2014-11-24 18:05:33,795 INFO
>>> >> >>>>>>> >>> >>> >
>>> [org.ovirt.engine.core.bll.ActivateVdsCommand]
>>> >> >>>>>>> >>> >>> >
>>> (org.ovirt.thread.pool-8-thread-45) [287d570d]
>>> >> >>>>>>> >>> >>> >
>>> Running
>>> >> >>>>>>> >>> >>> >
>>> command:
>>> >> >>>>>>> >>> >>> >
>>> ActivateVdsCommand internal: false. Entities
>>> >> >>>>>>> >>> >>> >
>>> affected :
>>> >> >>>>>>> >>> >>> >
>>> ID:
>>> >> >>>>>>> >>> >>> >
>>> 0bf6b00f-7947-4411-b55a-cc5eea2b381a Type:
>>> VDSAction
>>> >> >>>>>>> >>> >>> >
>>> group
>>> >> >>>>>>> >>> >>> >
>>> MANIPULATE_HOST
>>> >> >>>>>>> >>> >>> >
>>> with role type ADMIN
>>> >> >>>>>>> >>> >>> >
>>> 2014-11-24 18:05:33,796 INFO
>>> >> >>>>>>> >>> >>> >
>>> [org.ovirt.engine.core.bll.ActivateVdsCommand]
>>> >> >>>>>>> >>> >>> >
>>> (org.ovirt.thread.pool-8-thread-45) [287d570d]
>>> >> >>>>>>> >>> >>> >
>>> Before
>>> >> >>>>>>> >>> >>> >
>>> acquiring
>>> >> >>>>>>> >>> >>> >
>>> lock in
>>> >> >>>>>>> >>> >>> >
>>> order to prevent monitoring for host Compute5
>>> from
>>> >> >>>>>>> >>> >>> >
>>> data-center
>>> >> >>>>>>> >>> >>> >
>>> SV_WTC
>>> >> >>>>>>> >>> >>> >
>>> 2014-11-24 18:05:33,797 INFO
>>> >> >>>>>>> >>> >>> >
>>> [org.ovirt.engine.core.bll.ActivateVdsCommand]
>>> >> >>>>>>> >>> >>> >
>>> (org.ovirt.thread.pool-8-thread-45) [287d570d]
>>> Lock
>>> >> >>>>>>> >>> >>> >
>>> acquired,
>>> >> >>>>>>> >>> >>> >
>>> from
>>> >> >>>>>>> >>> >>> >
>>> now a
>>> >> >>>>>>> >>> >>> >
>>> monitoring of host will be skipped for host
>>> Compute5
>>> >> >>>>>>> >>> >>> >
>>> from
>>> >> >>>>>>> >>> >>> >
>>> data-center
>>> >> >>>>>>> >>> >>> >
>>> SV_WTC
>>> >> >>>>>>> >>> >>> >
>>> 2014-11-24 18:05:33,817 INFO
>>> >> >>>>>>> >>> >>> >
>>>
>>> >> >>>>>>> >>> >>> >
>>>
>>> >> >>>>>>> >>> >>> >
>>>
>>> [org.ovirt.engine.core.vdsbroker.SetVdsStatusVDSCommand]
>>> >> >>>>>>> >>> >>> >
>>> (org.ovirt.thread.pool-8-thread-45) [287d570d]
>>> >> >>>>>>> >>> >>> >
>>> START,
>>> >> >>>>>>> >>> >>> >
>>> SetVdsStatusVDSCommand(HostName = Compute5,
>>> HostId =
>>> >> >>>>>>> >>> >>> >
>>> 0bf6b00f-7947-4411-b55a-cc5eea2b381a,
>>> >> >>>>>>> >>> >>> >
>>> status=Unassigned,
>>> >> >>>>>>> >>> >>> >
>>> nonOperationalReason=NONE,
>>> >> >>>>>>> >>> >>> >
>>> stopSpmFailureLogged=false),
>>> >> >>>>>>> >>> >>> >
>>> log id:
>>> >> >>>>>>> >>> >>> >
>>> 1cbc7311
>>> >> >>>>>>> >>> >>> >
>>> 2014-11-24 18:05:33,820 INFO
>>> >> >>>>>>> >>> >>> >
>>>
>>> >> >>>>>>> >>> >>> >
>>>
>>> >> >>>>>>> >>> >>> >
>>>
>>> [org.ovirt.engine.core.vdsbroker.SetVdsStatusVDSCommand]
>>> >> >>>>>>> >>> >>> >
>>> (org.ovirt.thread.pool-8-thread-45) [287d570d]
>>> >> >>>>>>> >>> >>> >
>>> FINISH,
>>> >> >>>>>>> >>> >>> >
>>> SetVdsStatusVDSCommand, log id: 1cbc7311
>>> >> >>>>>>> >>> >>> >
>>> 2014-11-24 18:05:34,086 INFO
>>> >> >>>>>>> >>> >>> >
>>> [org.ovirt.engine.core.bll.ActivateVdsCommand]
>>> >> >>>>>>> >>> >>> >
>>> (org.ovirt.thread.pool-8-thread-45) Activate
>>> >> >>>>>>> >>> >>> >
>>> finished.
>>> >> >>>>>>> >>> >>> >
>>> Lock
>>> >> >>>>>>> >>> >>> >
>>> released.
>>> >> >>>>>>> >>> >>> >
>>> Monitoring can run now for host Compute5 from
>>> >> >>>>>>> >>> >>> >
>>> data-center
>>> >> >>>>>>> >>> >>> >
>>> SV_WTC
>>> >> >>>>>>> >>> >>> >
>>> 2014-11-24 18:05:34,088 INFO
>>> >> >>>>>>> >>> >>> >
>>>
>>> >> >>>>>>> >>> >>> >
>>>
>>> >> >>>>>>> >>> >>> >
>>>
>>> >> >>>>>>> >>> >>> >
>>>
>>> >> >>>>>>> >>> >>> >
>>>
>>> [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
>>> >> >>>>>>> >>> >>> >
>>> (org.ovirt.thread.pool-8-thread-45)
>>> Correlation ID:
>>> >> >>>>>>> >>> >>> >
>>> 287d570d,
>>> >> >>>>>>> >>> >>> >
>>> Job
>>> >> >>>>>>> >>> >>> >
>>> ID:
>>> >> >>>>>>> >>> >>> >
>>> 5ef8e4d6-b2bc-469e-8e81-7ef74b2a001a, Call
>>> Stack:
>>> >> >>>>>>> >>> >>> >
>>> null,
>>> >> >>>>>>> >>> >>> >
>>> Custom
>>> >> >>>>>>> >>> >>> >
>>> Event ID:
>>> >> >>>>>>> >>> >>> >
>>> -1, Message: Host Compute5 was activated by
>>> admin.
>>> >> >>>>>>> >>> >>> >
>>> 2014-11-24 18:05:34,090 INFO
>>> >> >>>>>>> >>> >>> >
>>> [org.ovirt.engine.core.bll.ActivateVdsCommand]
>>> >> >>>>>>> >>> >>> >
>>> (org.ovirt.thread.pool-8-thread-45) Lock freed
>>> to
>>> >> >>>>>>> >>> >>> >
>>> object
>>> >> >>>>>>> >>> >>> >
>>> EngineLock
>>> >> >>>>>>> >>> >>> >
>>> [exclusiveLocks= key:
>>> >> >>>>>>> >>> >>> >
>>> 0bf6b00f-7947-4411-b55a-cc5eea2b381a
>>> >> >>>>>>> >>> >>> >
>>> value:
>>> >> >>>>>>> >>> >>> >
>>> VDS
>>> >> >>>>>>> >>> >>> >
>>> , sharedLocks= ]
>>> >> >>>>>>> >>> >>> >
>>> 2014-11-24 18:05:35,792 INFO
>>> >> >>>>>>> >>> >>> >
>>>
>>> >> >>>>>>> >>> >>> >
>>>
>>> >> >>>>>>> >>> >>> >
>>>
>>> >> >>>>>>> >>> >>> >
>>>
>>> >> >>>>>>> >>> >>> >
>>>
>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand]
>>> >> >>>>>>> >>> >>> >
>>> (DefaultQuartzScheduler_Worker-55) [3706e836]
>>> START,
>>> >> >>>>>>> >>> >>> >
>>> GlusterVolumesListVDSCommand(HostName =
>>> Compute4,
>>> >> >>>>>>> >>> >>> >
>>> HostId
>>> >> >>>>>>> >>> >>> >
>>> =
>>> >> >>>>>>> >>> >>> >
>>> 33648a90-200c-45ca-89d5-1ce305d79a6a), log id:
>>> >> >>>>>>> >>> >>> >
>>> 48a0c832
>>> >> >>>>>>> >>> >>> >
>>> 2014-11-24 18:05:37,064 INFO
>>> >> >>>>>>> >>> >>> >
>>>
>>> >> >>>>>>> >>> >>> >
>>>
>>> >> >>>>>>> >>> >>> >
>>>
>>> >> >>>>>>> >>> >>> >
>>>
>>> >> >>>>>>> >>> >>> >
>>>
>>> [org.ovirt.engine.core.vdsbroker.vdsbroker.GetHardwareInfoVDSCommand]
>>> >> >>>>>>> >>> >>> >
>>> (DefaultQuartzScheduler_Worker-69) START,
>>> >> >>>>>>> >>> >>> >
>>> GetHardwareInfoVDSCommand(HostName = Compute5,
>>> >> >>>>>>> >>> >>> >
>>> HostId =
>>> >> >>>>>>> >>> >>> >
>>> 0bf6b00f-7947-4411-b55a-cc5eea2b381a,
>>> >> >>>>>>> >>> >>> >
>>>
>>> >> >>>>>>> >>> >>> >
>>>
>>> >> >>>>>>> >>> >>> >
>>>
>>> vds=Host[Compute5,0bf6b00f-7947-4411-b55a-cc5eea2b381a]), log
>>> >> >>>>>>> >>> >>> >
>>> id:
>>> >> >>>>>>> >>> >>> >
>>> 6d560cc2
>>> >> >>>>>>> >>> >>> >
>>> 2014-11-24 18:05:37,074 INFO
>>> >> >>>>>>> >>> >>> >
>>>
>>> >> >>>>>>> >>> >>> >
>>>
>>> >> >>>>>>> >>> >>> >
>>>
>>> >> >>>>>>> >>> >>> >
>>>
>>> >> >>>>>>> >>> >>> >
>>>
>>> [org.ovirt.engine.core.vdsbroker.vdsbroker.GetHardwareInfoVDSCommand]
>>> >> >>>>>>> >>> >>> >
>>> (DefaultQuartzScheduler_Worker-69) FINISH,
>>> >> >>>>>>> >>> >>> >
>>> GetHardwareInfoVDSCommand, log
>>> >> >>>>>>> >>> >>> >
>>> id: 6d560cc2
>>> >> >>>>>>> >>> >>> >
>>> 2014-11-24 18:05:37,093 WARN
>>> >> >>>>>>> >>> >>> >
>>> [org.ovirt.engine.core.vdsbroker.VdsManager]
>>> >> >>>>>>> >>> >>> >
>>> (DefaultQuartzScheduler_Worker-69) Host
>>> Compute5 is
>>> >> >>>>>>> >>> >>> >
>>> running
>>> >> >>>>>>> >>> >>> >
>>> with
>>> >> >>>>>>> >>> >>> >
>>> disabled
>>> >> >>>>>>> >>> >>> >
>>> SELinux.
>>> >> >>>>>>> >>> >>> >
>>> 2014-11-24 18:05:37,127 INFO
>>> >> >>>>>>> >>> >>> >
>>>
>>> >> >>>>>>> >>> >>> >
>>>
>>> >> >>>>>>> >>> >>> >
>>>
>>> >> >>>>>>> >>> >>> >
>>>
>>> >> >>>>>>> >>> >>> >
>>>
>>> [org.ovirt.engine.core.bll.HandleVdsCpuFlagsOrClusterChangedCommand]
>>> >> >>>>>>> >>> >>> >
>>> (DefaultQuartzScheduler_Worker-69) [2b4a51cf]
>>> >> >>>>>>> >>> >>> >
>>> Running
>>> >> >>>>>>> >>> >>> >
>>> command:
>>> >> >>>>>>> >>> >>> >
>>> HandleVdsCpuFlagsOrClusterChangedCommand
>>> internal:
>>> >> >>>>>>> >>> >>> >
>>> true.
>>> >> >>>>>>> >>> >>> >
>>> Entities
>>> >> >>>>>>> >>> >>> >
>>> affected
>>> >> >>>>>>> >>> >>> >
>>> : ID: 0bf6b00f-7947-4411-b55a-cc5eea2b381a
>>> Type:
>>> >> >>>>>>> >>> >>> >
>>> VDS
>>> >> >>>>>>> >>> >>> >
>>> 2014-11-24 18:05:37,147 INFO
>>> >> >>>>>>> >>> >>> >
>>>
>>> >> >>>>>>> >>> >>> >
>>>
>>> >> >>>>>>> >>> >>> >
>>>
>>> >> >>>>>>> >>> >>> >
>>>
>>> >> >>>>>>> >>> >>> >
>>>
>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand]
>>> >> >>>>>>> >>> >>> >
>>> (DefaultQuartzScheduler_Worker-69) [2b4a51cf]
>>> START,
>>> >> >>>>>>> >>> >>> >
>>> GlusterServersListVDSCommand(HostName =
>>> Compute5,
>>> >> >>>>>>> >>> >>> >
>>> HostId
>>> >> >>>>>>> >>> >>> >
>>> =
>>> >> >>>>>>> >>> >>> >
>>> 0bf6b00f-7947-4411-b55a-cc5eea2b381a), log id:
>>> >> >>>>>>> >>> >>> >
>>> 4faed87
>>> >> >>>>>>> >>> >>> >
>>> 2014-11-24 18:05:37,164 INFO
>>> >> >>>>>>> >>> >>> >
>>>
>>> >> >>>>>>> >>> >>> >
>>>
>>> >> >>>>>>> >>> >>> >
>>>
>>> >> >>>>>>> >>> >>> >
>>>
>>> >> >>>>>>> >>> >>> >
>>>
>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand]
>>> >> >>>>>>> >>> >>> >
>>> (DefaultQuartzScheduler_Worker-69) [2b4a51cf]
>>> >> >>>>>>> >>> >>> >
>>> FINISH,
>>> >> >>>>>>> >>> >>> >
>>> GlusterServersListVDSCommand, log id: 4faed87
>>> >> >>>>>>> >>> >>> >
>>> 2014-11-24 18:05:37,189 INFO
>>> >> >>>>>>> >>> >>> >
>>>
>>> >> >>>>>>> >>> >>> >
>>>
>>> [org.ovirt.engine.core.bll.SetNonOperationalVdsCommand]
>>> >> >>>>>>> >>> >>> >
>>> (DefaultQuartzScheduler_Worker-69) [4a84c4e5]
>>> >> >>>>>>> >>> >>> >
>>> Running
>>> >> >>>>>>> >>> >>> >
>>> command:
>>> >> >>>>>>> >>> >>> >
>>> SetNonOperationalVdsCommand internal: true.
>>> Entities
>>> >> >>>>>>> >>> >>> >
>>> affected :
>>> >> >>>>>>> >>> >>> >
>>> ID:
>>> >> >>>>>>> >>> >>> >
>>> 0bf6b00f-7947-4411-b55a-cc5eea2b381a Type: VDS
>>> >> >>>>>>> >>> >>> >
>>> 2014-11-24 18:05:37,206 INFO
>>> >> >>>>>>> >>> >>> >
>>>
>>> >> >>>>>>> >>> >>> >
>>>
>>> >> >>>>>>> >>> >>> >
>>>
>>> [org.ovirt.engine.core.vdsbroker.SetVdsStatusVDSCommand]
>>> >> >>>>>>> >>> >>> >
>>> (DefaultQuartzScheduler_Worker-69) [4a84c4e5]
>>> START,
>>> >> >>>>>>> >>> >>> >
>>> SetVdsStatusVDSCommand(HostName = Compute5,
>>> HostId =
>>> >> >>>>>>> >>> >>> >
>>> 0bf6b00f-7947-4411-b55a-cc5eea2b381a,
>>> >> >>>>>>> >>> >>> >
>>> status=NonOperational,
>>> >> >>>>>>> >>> >>> >
>>> nonOperationalReason=GLUSTER_COMMAND_FAILED,
>>> >> >>>>>>> >>> >>> >
>>> stopSpmFailureLogged=false),
>>> >> >>>>>>> >>> >>> >
>>> log id: fed5617
>>> >> >>>>>>> >>> >>> >
>>> 2014-11-24 18:05:37,209 INFO
>>> >> >>>>>>> >>> >>> >
>>>
>>> >> >>>>>>> >>> >>> >
>>>
>>> >> >>>>>>> >>> >>> >
>>>
>>> [org.ovirt.engine.core.vdsbroker.SetVdsStatusVDSCommand]
>>> >> >>>>>>> >>> >>> >
>>> (DefaultQuartzScheduler_Worker-69) [4a84c4e5]
>>> >> >>>>>>> >>> >>> >
>>> FINISH,
>>> >> >>>>>>> >>> >>> >
>>> SetVdsStatusVDSCommand, log id: fed5617
>>> >> >>>>>>> >>> >>> >
>>> 2014-11-24 18:05:37,223 ERROR
>>> >> >>>>>>> >>> >>> >
>>>
>>> >> >>>>>>> >>> >>> >
>>>
>>> >> >>>>>>> >>> >>> >
>>>
>>> >> >>>>>>> >>> >>> >
>>>
>>> >> >>>>>>> >>> >>> >
>>>
>>> [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
>>> >> >>>>>>> >>> >>> >
>>> (DefaultQuartzScheduler_Worker-69) [4a84c4e5]
>>> >> >>>>>>> >>> >>> >
>>> Correlation ID:
>>> >> >>>>>>> >>> >>> >
>>> 4a84c4e5,
>>> >> >>>>>>> >>> >>> >
>>> Job
>>> >> >>>>>>> >>> >>> >
>>> ID: 4bfd4a6d-c3ef-468f-a40e-a3a6ca13011b, Call
>>> >> >>>>>>> >>> >>> >
>>> Stack:
>>> >> >>>>>>> >>> >>> >
>>> null,
>>> >> >>>>>>> >>> >>> >
>>> Custom
>>> >> >>>>>>> >>> >>> >
>>> Event
>>> >> >>>>>>> >>> >>> >
>>> ID: -1, Message: Gluster command [<UNKNOWN>]
>>> failed
>>> >> >>>>>>> >>> >>> >
>>> on
>>> >> >>>>>>> >>> >>> >
>>> server
>>> >> >>>>>>> >>> >>> >
>>> Compute5.
>>> >> >>>>>>> >>> >>> >
>>> 2014-11-24 18:05:37,243 INFO
>>> >> >>>>>>> >>> >>> >
>>>
>>> >> >>>>>>> >>> >>> >
>>>
>>> >> >>>>>>> >>> >>> >
>>>
>>> >> >>>>>>> >>> >>> >
>>>
>>> >> >>>>>>> >>> >>> >
>>>
>>> [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
>>> >> >>>>>>> >>> >>> >
>>> (DefaultQuartzScheduler_Worker-69) [4a84c4e5]
>>> >> >>>>>>> >>> >>> >
>>> Correlation ID:
>>> >> >>>>>>> >>> >>> >
>>> null,
>>> >> >>>>>>> >>> >>> >
>>> Call
>>> >> >>>>>>> >>> >>> >
>>> Stack: null, Custom Event ID: -1, Message:
>>> Status of
>>> >> >>>>>>> >>> >>> >
>>> host
>>> >> >>>>>>> >>> >>> >
>>> Compute5
>>> >> >>>>>>> >>> >>> >
>>> was
>>> >> >>>>>>> >>> >>> >
>>> set
>>> >> >>>>>>> >>> >>> >
>>> to NonOperational.
>>> >> >>>>>>> >>> >>> >
>>> 2014-11-24 18:05:37,272 INFO
>>> >> >>>>>>> >>> >>> >
>>>
>>> [org.ovirt.engine.core.bll.HandleVdsVersionCommand]
>>> >> >>>>>>> >>> >>> >
>>> (DefaultQuartzScheduler_Worker-69) [a0c8a7f]
>>> Running
>>> >> >>>>>>> >>> >>> >
>>> command:
>>> >> >>>>>>> >>> >>> >
>>> HandleVdsVersionCommand internal: true.
>>> Entities
>>> >> >>>>>>> >>> >>> >
>>> affected :
>>> >> >>>>>>> >>> >>> >
>>> ID:
>>> >> >>>>>>> >>> >>> >
>>> 0bf6b00f-7947-4411-b55a-cc5eea2b381a Type: VDS
>>> >> >>>>>>> >>> >>> >
>>> 2014-11-24 18:05:37,274 INFO
>>> >> >>>>>>> >>> >>> >
>>>
>>> >> >>>>>>> >>> >>> >
>>>
>>> [org.ovirt.engine.core.vdsbroker.VdsUpdateRunTimeInfo]
>>> >> >>>>>>> >>> >>> >
>>> (DefaultQuartzScheduler_Worker-69) [a0c8a7f]
>>> Host
>>> >> >>>>>>> >>> >>> >
>>> 0bf6b00f-7947-4411-b55a-cc5eea2b381a :
>>> Compute5 is
>>> >> >>>>>>> >>> >>> >
>>> already in
>>> >> >>>>>>> >>> >>> >
>>> NonOperational status for reason
>>> >> >>>>>>> >>> >>> >
>>> GLUSTER_COMMAND_FAILED.
>>> >> >>>>>>> >>> >>> >
>>> SetNonOperationalVds command is skipped.
>>> >> >>>>>>> >>> >>> >
>>> 2014-11-24 18:05:38,065 INFO
>>> >> >>>>>>> >>> >>> >
>>>
>>> >> >>>>>>> >>> >>> >
>>>
>>> >> >>>>>>> >>> >>> >
>>>
>>> >> >>>>>>> >>> >>> >
>>>
>>> >> >>>>>>> >>> >>> >
>>>
>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand]
>>> >> >>>>>>> >>> >>> >
>>> (DefaultQuartzScheduler_Worker-55) [3706e836]
>>> >> >>>>>>> >>> >>> >
>>> FINISH,
>>> >> >>>>>>> >>> >>> >
>>> GlusterVolumesListVDSCommand, return:
>>> >> >>>>>>> >>> >>> >
>>>
>>> >> >>>>>>> >>> >>> >
>>>
>>> >> >>>>>>> >>> >>> >
>>>
>>> >> >>>>>>> >>> >>> >
>>>
>>> >> >>>>>>> >>> >>> >
>>>
>>>
{26ae1672-ee09-4a38-8fd2-72dd9974cc2b=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@4e72a1b1
>>> },
>>> >> >>>>>>> >>> >>> >
>>> log id: 48a0c832
>>> >> >>>>>>> >>> >>> >
>>> 2014-11-24 18:05:43,243 INFO
>>> >> >>>>>>> >>> >>> >
>>>
>>> >> >>>>>>> >>> >>> >
>>>
>>> >> >>>>>>> >>> >>> >
>>>
>>> >> >>>>>>> >>> >>> >
>>>
>>> >> >>>>>>> >>> >>> >
>>>
>>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand]
>>> >> >>>>>>> >>> >>> >
>>> (DefaultQuartzScheduler_Worker-35) START,
>>> >> >>>>>>> >>> >>> >
>>> GlusterVolumesListVDSCommand(HostName =
>>> Compute4,
>>> >> >>>>>>> >>> >>> >
>>> HostId
>>> >> >>>>>>> >>> >>> >
>>> =
>>> >> >>>>>>> >>> >>> >
>>> 33648a90-200c-45ca-89d5-1ce305d79a6a), log id:
>>> >> >>>>>>> >>> >>> >
>>> 3ce13ebc
>>> >> >>>>>>> >>> >>> >
>>> ^C
>>> >> >>>>>>> >>> >>> >
>>> [root@ccr01 ~]#
>>> >> >>>>>>> >>> >>> >
>>>
>>> >> >>>>>>> >>> >>> >
>>> Thanks,
>>> >> >>>>>>> >>> >>> >
>>> Punit
>>> >> >>>>>>> >>> >>> >
>>>
>>> >> >>>>>>> >>> >>> >
>>>
>>> >> >>>>>>> >>> >>> >
>>>
>>> _______________________________________________
>>> >> >>>>>>> >>> >>> >
>>> Users mailing
>>> >> >>>>>>> >>> >>> >
>>>
>>> >> >>>>>>> >>> >>> >
>>>
>>> >> >>>>>>> >>> >>> >
>>>
>>> >> >>>>>>> >>> >>> >
>>> listUsers@ovirt.orghttp://
>>>
lists.ovirt.org/mailman/listinfo/users
>>> >> >>>>>>> >>> >>> >
>>>
>>> >> >>>>>>> >>> >>> >
>>>
>>> >> >>>>>>> >>> >>> >
>>>
>>> >> >>>>>>> >>> >>> >
>>
>>> >> >>>>>>> >>> >>> >
>>
>>> >> >>>>>>> >>> >>> > >
>>> >> >>>>>>> >>> >>> > >
>>> >> >>>>>>> >>> >>> >
>>> >> >>>>>>> >>> >>
>>> >> >>>>>>> >>> >>
>>> >> >>>>>>> >>> >
>>> >> >>>>>>> >>> >
>>> >> >>>>>>> >>> >
_______________________________________________
>>> >> >>>>>>> >>> > Gluster-users mailing
list
>>> >> >>>>>>> >>> >
Gluster-users(a)gluster.org
>>> >> >>>>>>> >>> >
>>> >> >>>>>>> >>> >
>>>
http://supercolony.gluster.org/mailman/listinfo/gluster-users
>>> >> >>>>>>> >>
>>> >> >>>>>>> >>
>>> >> >>>>>>> >
>>> >> >>>>>>
>>> >> >>>>>>
>>> >> >>>>
>>> >> >>
>>> >
>>> >
>>>
>>
>>
>
_______________________________________________
Users mailing list
Users(a)ovirt.org
http://lists.ovirt.org/mailman/listinfo/users