I post a bug in centos about a problem with
NetworkManager-wait-online.service, this service is executed before
network-online.target
I had a problem with mount drbd in a centos7 at boot time. I think it could
be similar to the problem to mount gluster in boot time in centos7.
I modified the service adding one line that wait to NetworkManager to set
up the ip address in the interface that is used to connect de drbd, the the
ip address in eth0 is 10.88.0.1 and the line I added is:
ExecStart=/bin/sh -c 'until ip a s eth0 | grep -q 10.88; do usleep
100000; done'
# cat /usr/lib/systemd/system/NetworkManager-wait-online.service
[Unit]
Description=Network Manager Wait Online
Requisite=NetworkManager.service
After=NetworkManager.service
Wants=network.target
Before=network.target network-online.target
[Service]
Type=oneshot
ExecStart=/usr/bin/nm-online -q --timeout=30
ExecStart=/bin/sh -c 'until ip a s eth0 | grep -q 10.88; do usleep 100000;
done'
Alberto
2014-12-05 14:50 GMT+01:00 Kaushal M <kshlmster(a)gmail.com>:
Can you replace 'Before=network-online.target' with
'Wants=network-online.target' and try the boot again? This should
force the network to be online before starting GlusterD.
If even that fails, you could try adding an entry into /etc/hosts with
the hostname of the system. This should prevent any more failures.
I still don't believe it's a problem with Gluster. Gluster uses apis
provided by the system to perform name resolution. These definitely
work correctly because you can start GlusterD later. Since the
resolution failure only happens during boot, it points to system or
network setup issues during boot. To me it seems like the network
isn't completely setup at that point of time.
~kaushal
On Fri, Dec 5, 2014 at 12:47 PM, Punit Dambiwal <hypunit(a)gmail.com> wrote:
> Hi Kaushal,
>
> It seems it's bug in glusterfs 3.6....even i manage my systemd to start
the
> network service before glusterd...but it's still fail...
>
> ---------------
> [Unit]
> Description=GlusterFS, a clustered file-system server
> After=network.target rpcbind.service
> Before=network-online.target
>
> [Service]
> Type=forking
> PIDFile=/var/run/glusterd.pid
> LimitNOFILE=65536
> ExecStartPre=/etc/rc.d/init.d/network start
> ExecStart=/usr/sbin/glusterd -p /var/run/glusterd.pid
> KillMode=process
>
> [Install]
> WantedBy=multi-user.target
> ----------------
>
> Thanks,
> Punit
>
> On Wed, Dec 3, 2014 at 8:56 PM, Kaushal M <kshlmster(a)gmail.com> wrote:
>>
>> I just remembered this.
>>
>> There was another user having a similar issue of GlusterD failing to
>> start on the mailing list a while back. The cause of his problem was
>> the way his network was brought up.
>> IIRC, he was using a static network configuration. The problem
>> vanished when he began using dhcp. Or it might have been he was using
>> dhcp.service and it got solved after switching to NetworkManager.
>>
>> This could be one more thing you could look at.
>>
>> I'll try to find the mail thread to see if it was the same problem as
you.
>>
>> ~kaushal
>>
>> On Wed, Dec 3, 2014 at 6:22 PM, Kaushal M <kshlmster(a)gmail.com> wrote:
>> > I don't know much about how the network target is brought up in
>> > CentOS7, but I'll try as much as I can.
>> >
>> > It seems to me that, after the network has been brought up and by the
>> > time GlusterD is started,
>> > a. The machine hasn't yet recieved it's hostname, or
>> > b. It hasn't yet registered with the name server.
>> >
>> > This is causing name resolution failures.
>> >
>> > I don't know if the network target could come up without the machine
>> > getting its hostname, so I'm pretty sure it's not a.
>> >
>> > So it seems to be b. But these kind of signing in happens only in DDNS
>> > systems, which doesn't seem to be the case for you.
>> >
>> > Both of these reasons might be wrong (most likely wrong). You'd do
>> > good if you could ask for help from someone with more experience in
>> > systemd + networking.
>> >
>> > ~kaushal
>> >
>> > On Wed, Dec 3, 2014 at 10:54 AM, Punit Dambiwal <hypunit(a)gmail.com>
>> > wrote:
>> >> Hi Kaushal,
>> >>
>> >> This is the host...which i rebooted...would you mind to let me know
how
>> >> i
>> >> can make the glusterd sevice come up after network...i am using
>> >> centos7...if
>> >> network is the issue...
>> >>
>> >> On Wed, Dec 3, 2014 at 11:54 AM, Kaushal M <kshlmster(a)gmail.com>
wrote:
>> >>>
>> >>> This peer cannot be identified.
>> >>>
>> >>> " [2014-12-03 02:29:25.998153] D
>> >>> [glusterd-peer-utils.c:121:glusterd_peerinfo_find_by_hostname]
>> >>> 0-management:
>> >>> Unable to find friend: cpu05.zne01.hkg1.ovt.36stack.com"
>> >>>
>> >>> I don't know why this address is not being resolved during
boot
time.
>> >>> If
>> >>> this is a valid peer, the the only reason I can think of this that
the
>> >>> network is not up.
>> >>>
>> >>> If you had previously detached the peer forcefully, the that could
>> >>> have
>> >>> left stale entries in some volumes. In this case as well, GlusterD
>> >>> will fail
>> >>> to identify the peer.
>> >>>
>> >>> Do either of these reasons seem a possibility to you?
>> >>>
>> >>> On Dec 3, 2014 8:07 AM, "Punit Dambiwal"
<hypunit(a)gmail.com> wrote:
>> >>>>
>> >>>> Hi Kaushal,
>> >>>>
>> >>>> Please find the logs here :-
http://ur1.ca/iyoe5 and
>> >>>>
http://ur1.ca/iyoed
>> >>>>
>> >>>> On Tue, Dec 2, 2014 at 10:43 PM, Kaushal M
<kshlmster(a)gmail.com>
>> >>>> wrote:
>> >>>>>
>> >>>>> Hey Punit,
>> >>>>> In the logs you've provided, GlusterD appears to be
running
>> >>>>> correctly.
>> >>>>> Could you provide the logs for the time period when
GlusterD
>> >>>>> attempts to
>> >>>>> start but fails.
>> >>>>>
>> >>>>> ~kaushal
>> >>>>>
>> >>>>> On Dec 2, 2014 8:03 PM, "Punit Dambiwal"
<hypunit(a)gmail.com>
wrote:
>> >>>>>>
>> >>>>>> Hi Kaushal,
>> >>>>>>
>> >>>>>> Please find the logs here :-
http://ur1.ca/iyhs5 and
>> >>>>>>
http://ur1.ca/iyhue
>> >>>>>>
>> >>>>>> Thanks,
>> >>>>>> punit
>> >>>>>>
>> >>>>>>
>> >>>>>> On Tue, Dec 2, 2014 at 12:00 PM, Kaushal M
<kshlmster(a)gmail.com>
>> >>>>>> wrote:
>> >>>>>>>
>> >>>>>>> Hey Punit,
>> >>>>>>> Could you start Glusterd in debug mode and provide
the logs
here?
>> >>>>>>> To start it in debug mode, append '-LDEBUG'
to the ExecStart
line
>> >>>>>>> in
>> >>>>>>> the service file.
>> >>>>>>>
>> >>>>>>> ~kaushal
>> >>>>>>>
>> >>>>>>> On Mon, Dec 1, 2014 at 9:05 AM, Punit Dambiwal
<
hypunit(a)gmail.com>
>> >>>>>>> wrote:
>> >>>>>>> > Hi,
>> >>>>>>> >
>> >>>>>>> > Can Any body help me on this ??
>> >>>>>>> >
>> >>>>>>> > On Thu, Nov 27, 2014 at 9:29 AM, Punit
Dambiwal
>> >>>>>>> > <hypunit(a)gmail.com>
>> >>>>>>> > wrote:
>> >>>>>>> >>
>> >>>>>>> >> Hi Kaushal,
>> >>>>>>> >>
>> >>>>>>> >> Thanks for the detailed reply....let me
explain my setup
first
>> >>>>>>> >> :-
>> >>>>>>> >>
>> >>>>>>> >> 1. Ovirt Engine
>> >>>>>>> >> 2. 4* host as well as storage machine
(Host and gluster
>> >>>>>>> >> combined)
>> >>>>>>> >> 3. Every host has 24 bricks...
>> >>>>>>> >>
>> >>>>>>> >> Now whenever the host machine reboot...it
can come up but can
>> >>>>>>> >> not
>> >>>>>>> >> join the
>> >>>>>>> >> cluster again and through the following
error "Gluster
command
>> >>>>>>> >> [<UNKNOWN>]
>> >>>>>>> >> failed on server.."
>> >>>>>>> >>
>> >>>>>>> >> Please check my comment in line :-
>> >>>>>>> >>
>> >>>>>>> >> 1. Use the same string for doing the peer
probe and for the
>> >>>>>>> >> brick
>> >>>>>>> >> address
>> >>>>>>> >> during volume create/add-brick. Ideally,
we suggest you use
>> >>>>>>> >> properly
>> >>>>>>> >> resolvable FQDNs everywhere. If that is
not possible, then
use
>> >>>>>>> >> only
>> >>>>>>> >> IP
>> >>>>>>> >> addresses. Try to avoid short names.
>> >>>>>>> >> ---------------
>> >>>>>>> >> [root@cpu05 ~]# gluster peer status
>> >>>>>>> >> Number of Peers: 3
>> >>>>>>> >>
>> >>>>>>> >> Hostname:
cpu03.stack.com
>> >>>>>>> >> Uuid:
5729b8c4-e80d-4353-b456-6f467bddbdfb
>> >>>>>>> >> State: Peer in Cluster (Connected)
>> >>>>>>> >>
>> >>>>>>> >> Hostname:
cpu04.stack.com
>> >>>>>>> >> Uuid:
d272b790-c4b2-4bed-ba68-793656e6d7b0
>> >>>>>>> >> State: Peer in Cluster (Connected)
>> >>>>>>> >> Other names:
>> >>>>>>> >> 10.10.0.8
>> >>>>>>> >>
>> >>>>>>> >> Hostname:
cpu02.stack.com
>> >>>>>>> >> Uuid:
8d8a7041-950e-40d0-85f9-58d14340ca25
>> >>>>>>> >> State: Peer in Cluster (Connected)
>> >>>>>>> >> [root@cpu05 ~]#
>> >>>>>>> >> ----------------
>> >>>>>>> >> 2. During boot up, make sure to launch
glusterd only after
the
>> >>>>>>> >> network is
>> >>>>>>> >> up. This will allow the new peer
identification mechanism to
do
>> >>>>>>> >> its
>> >>>>>>> >> job correctly.
>> >>>>>>> >> >> I think the service itself doing
the same job....
>> >>>>>>> >>
>> >>>>>>> >> [root@cpu05 ~]# cat
/usr/lib/systemd/system/glusterd.service
>> >>>>>>> >> [Unit]
>> >>>>>>> >> Description=GlusterFS, a clustered
file-system server
>> >>>>>>> >> After=network.target rpcbind.service
>> >>>>>>> >> Before=network-online.target
>> >>>>>>> >>
>> >>>>>>> >> [Service]
>> >>>>>>> >> Type=forking
>> >>>>>>> >> PIDFile=/var/run/glusterd.pid
>> >>>>>>> >> LimitNOFILE=65536
>> >>>>>>> >> ExecStart=/usr/sbin/glusterd -p
/var/run/glusterd.pid
>> >>>>>>> >> KillMode=process
>> >>>>>>> >>
>> >>>>>>> >> [Install]
>> >>>>>>> >> WantedBy=multi-user.target
>> >>>>>>> >> [root@cpu05 ~]#
>> >>>>>>> >> --------------------
>> >>>>>>> >>
>> >>>>>>> >> gluster logs :-
>> >>>>>>> >>
>> >>>>>>> >> [2014-11-24 09:22:22.147471] I [MSGID:
100030]
>> >>>>>>> >> [glusterfsd.c:2018:main]
>> >>>>>>> >> 0-/usr/sbin/glusterd: Started running
/usr/sbin/glusterd
>> >>>>>>> >> version
>> >>>>>>> >> 3.6.1
>> >>>>>>> >> (args: /usr/sbin/glusterd -p
/var/run/glusterd.pid)
>> >>>>>>> >> [2014-11-24 09:22:22.151565] I
[glusterd.c:1214:init]
>> >>>>>>> >> 0-management:
>> >>>>>>> >> Maximum allowed open file descriptors set
to 65536
>> >>>>>>> >> [2014-11-24 09:22:22.151599] I
[glusterd.c:1259:init]
>> >>>>>>> >> 0-management:
>> >>>>>>> >> Using
>> >>>>>>> >> /var/lib/glusterd as working directory
>> >>>>>>> >> [2014-11-24 09:22:22.155216] W
>> >>>>>>> >> [rdma.c:4195:__gf_rdma_ctx_create]
>> >>>>>>> >> 0-rpc-transport/rdma: rdma_cm event
channel creation failed
(No
>> >>>>>>> >> such device)
>> >>>>>>> >> [2014-11-24 09:22:22.155264] E
[rdma.c:4483:init]
>> >>>>>>> >> 0-rdma.management:
>> >>>>>>> >> Failed to initialize IB Device
>> >>>>>>> >> [2014-11-24 09:22:22.155285] E
>> >>>>>>> >> [rpc-transport.c:333:rpc_transport_load]
>> >>>>>>> >> 0-rpc-transport: 'rdma'
initialization failed
>> >>>>>>> >> [2014-11-24 09:22:22.155354] W
>> >>>>>>> >> [rpcsvc.c:1524:rpcsvc_transport_create]
>> >>>>>>> >> 0-rpc-service: cannot create listener,
initing the transport
>> >>>>>>> >> failed
>> >>>>>>> >> [2014-11-24 09:22:22.156290] I
>> >>>>>>> >>
[glusterd.c:413:glusterd_check_gsync_present] 0-glusterd:
>> >>>>>>> >> geo-replication
>> >>>>>>> >> module not installed in the system
>> >>>>>>> >> [2014-11-24 09:22:22.161318] I
>> >>>>>>> >>
[glusterd-store.c:2043:glusterd_restore_op_version]
0-glusterd:
>> >>>>>>> >> retrieved
>> >>>>>>> >> op-version: 30600
>> >>>>>>> >> [2014-11-24 09:22:22.821800] I
>> >>>>>>> >>
[glusterd-handler.c:3146:glusterd_friend_add_from_peerinfo]
>> >>>>>>> >> 0-management:
>> >>>>>>> >> connect returned 0
>> >>>>>>> >> [2014-11-24 09:22:22.825810] I
>> >>>>>>> >>
[glusterd-handler.c:3146:glusterd_friend_add_from_peerinfo]
>> >>>>>>> >> 0-management:
>> >>>>>>> >> connect returned 0
>> >>>>>>> >> [2014-11-24 09:22:22.828705] I
>> >>>>>>> >>
[glusterd-handler.c:3146:glusterd_friend_add_from_peerinfo]
>> >>>>>>> >> 0-management:
>> >>>>>>> >> connect returned 0
>> >>>>>>> >> [2014-11-24 09:22:22.828771] I
>> >>>>>>> >> [rpc-clnt.c:969:rpc_clnt_connection_init]
>> >>>>>>> >> 0-management: setting frame-timeout to
600
>> >>>>>>> >> [2014-11-24 09:22:22.832670] I
>> >>>>>>> >> [rpc-clnt.c:969:rpc_clnt_connection_init]
>> >>>>>>> >> 0-management: setting frame-timeout to
600
>> >>>>>>> >> [2014-11-24 09:22:22.835919] I
>> >>>>>>> >> [rpc-clnt.c:969:rpc_clnt_connection_init]
>> >>>>>>> >> 0-management: setting frame-timeout to
600
>> >>>>>>> >> [2014-11-24 09:22:22.840209] E
>> >>>>>>> >>
[glusterd-store.c:4248:glusterd_resolve_all_bricks]
0-glusterd:
>> >>>>>>> >> resolve
>> >>>>>>> >> brick failed in restore
>> >>>>>>> >> [2014-11-24 09:22:22.840233] E
[xlator.c:425:xlator_init]
>> >>>>>>> >> 0-management:
>> >>>>>>> >> Initialization of volume
'management' failed, review your
>> >>>>>>> >> volfile
>> >>>>>>> >> again
>> >>>>>>> >> [2014-11-24 09:22:22.840245] E
>> >>>>>>> >> [graph.c:322:glusterfs_graph_init]
>> >>>>>>> >> 0-management: initializing translator
failed
>> >>>>>>> >> [2014-11-24 09:22:22.840264] E
>> >>>>>>> >> [graph.c:525:glusterfs_graph_activate]
>> >>>>>>> >> 0-graph: init failed
>> >>>>>>> >> [2014-11-24 09:22:22.840754] W
>> >>>>>>> >> [glusterfsd.c:1194:cleanup_and_exit]
>> >>>>>>> >> (-->
>> >>>>>>> >> 0-: received signum (0), shutting down
>> >>>>>>> >>
>> >>>>>>> >> Thanks,
>> >>>>>>> >> Punit
>> >>>>>>> >>
>> >>>>>>> >>
>> >>>>>>> >>
>> >>>>>>> >>
>> >>>>>>> >> On Wed, Nov 26, 2014 at 7:14 PM, Kaushal
M
>> >>>>>>> >> <kshlmster(a)gmail.com>
>> >>>>>>> >> wrote:
>> >>>>>>> >>>
>> >>>>>>> >>> Based on the logs I can guess that
glusterd is being started
>> >>>>>>> >>> before
>> >>>>>>> >>> the network has come up and that the
addresses given to
bricks
>> >>>>>>> >>> do
>> >>>>>>> >>> not
>> >>>>>>> >>> directly match the addresses used in
during peer probe.
>> >>>>>>> >>>
>> >>>>>>> >>> The gluster_after_reboot log has the
line "[2014-11-25
>> >>>>>>> >>> 06:46:09.972113] E
>> >>>>>>> >>>
[glusterd-store.c:2632:glusterd_resolve_all_bricks]
>> >>>>>>> >>> 0-glusterd: resolve brick failed in
restore".
>> >>>>>>> >>>
>> >>>>>>> >>> Brick resolution fails when glusterd
cannot match the
address
>> >>>>>>> >>> for
>> >>>>>>> >>> the
>> >>>>>>> >>> brick, with one of the peers. Brick
resolution happens in
two
>> >>>>>>> >>> phases,
>> >>>>>>> >>> 1. We first try to identify the peer
by performing string
>> >>>>>>> >>> comparisions
>> >>>>>>> >>> with the brick address and the peer
addresses (The peer
names
>> >>>>>>> >>> will
>> >>>>>>> >>> be
>> >>>>>>> >>> the names/addresses that were given
when the peer was
probed).
>> >>>>>>> >>> 2. If we don't find a match from
step 1, we will then
resolve
>> >>>>>>> >>> all
>> >>>>>>> >>> the
>> >>>>>>> >>> brick address and the peer addresses
into addrinfo structs,
>> >>>>>>> >>> and
>> >>>>>>> >>> then
>> >>>>>>> >>> compare these structs to find a match.
This process should
>> >>>>>>> >>> generally
>> >>>>>>> >>> find a match if available. This will
fail only if the
network
>> >>>>>>> >>> is
>> >>>>>>> >>> not
>> >>>>>>> >>> up yet as we cannot resolve
addresses.
>> >>>>>>> >>>
>> >>>>>>> >>> The above steps are applicable only to
glusterfs versions
>> >>>>>>> >>> >=3.6.
>> >>>>>>> >>> They
>> >>>>>>> >>> were introduced to reduce problems
with peer identification,
>> >>>>>>> >>> like
>> >>>>>>> >>> the
>> >>>>>>> >>> one you encountered
>> >>>>>>> >>>
>> >>>>>>> >>> Since both of the steps failed to find
a match in one run,
but
>> >>>>>>> >>> succeeded later, we can come to the
conclusion that,
>> >>>>>>> >>> a) the bricks don't have the exact
same string used in peer
>> >>>>>>> >>> probe
>> >>>>>>> >>> for
>> >>>>>>> >>> their addresses as step 1 failed, and
>> >>>>>>> >>> b) the network was not up in the
initial run, as step 2
failed
>> >>>>>>> >>> during
>> >>>>>>> >>> the initial run, but passed in the
second run.
>> >>>>>>> >>>
>> >>>>>>> >>> Please let me know if my conclusion is
correct.
>> >>>>>>> >>>
>> >>>>>>> >>> If it is, you can solve your problem
in two ways.
>> >>>>>>> >>> 1. Use the same string for doing the
peer probe and for the
>> >>>>>>> >>> brick
>> >>>>>>> >>> address during volume
create/add-brick. Ideally, we suggest
>> >>>>>>> >>> you
>> >>>>>>> >>> use
>> >>>>>>> >>> properly resolvable FQDNs everywhere.
If that is not
possible,
>> >>>>>>> >>> then
>> >>>>>>> >>> use only IP addresses. Try to avoid
short names.
>> >>>>>>> >>> 2. During boot up, make sure to launch
glusterd only after
the
>> >>>>>>> >>> network
>> >>>>>>> >>> is up. This will allow the new peer
identification mechanism
>> >>>>>>> >>> to do
>> >>>>>>> >>> its
>> >>>>>>> >>> job correctly.
>> >>>>>>> >>>
>> >>>>>>> >>>
>> >>>>>>> >>> If you have already followed these
steps and yet still hit
the
>> >>>>>>> >>> problem, then please provide more
information (setup, logs,
>> >>>>>>> >>> etc.).
>> >>>>>>> >>> It
>> >>>>>>> >>> could be much different problem that
you are facing.
>> >>>>>>> >>>
>> >>>>>>> >>> ~kaushal
>> >>>>>>> >>>
>> >>>>>>> >>> On Wed, Nov 26, 2014 at 4:01 PM, Punit
Dambiwal
>> >>>>>>> >>> <hypunit(a)gmail.com>
>> >>>>>>> >>> wrote:
>> >>>>>>> >>> > Is there any one can help on this
??
>> >>>>>>> >>> >
>> >>>>>>> >>> > Thanks,
>> >>>>>>> >>> > punit
>> >>>>>>> >>> >
>> >>>>>>> >>> > On Wed, Nov 26, 2014 at 9:42 AM,
Punit Dambiwal
>> >>>>>>> >>> > <hypunit(a)gmail.com>
>> >>>>>>> >>> > wrote:
>> >>>>>>> >>> >>
>> >>>>>>> >>> >> Hi,
>> >>>>>>> >>> >>
>> >>>>>>> >>> >> My Glusterfs version is :-
glusterfs-3.6.1-1.el7
>> >>>>>>> >>> >>
>> >>>>>>> >>> >> On Wed, Nov 26, 2014 at 1:59
AM, Kanagaraj Mayilsamy
>> >>>>>>> >>> >> <kmayilsa(a)redhat.com>
>> >>>>>>> >>> >> wrote:
>> >>>>>>> >>> >>>
>> >>>>>>> >>> >>>
[+Gluster-users(a)gluster.org]
>> >>>>>>> >>> >>>
>> >>>>>>> >>> >>> "Initialization of
volume 'management' failed, review
your
>> >>>>>>> >>> >>> volfile
>> >>>>>>> >>> >>> again", glusterd
throws this error when the service is
>> >>>>>>> >>> >>> started
>> >>>>>>> >>> >>> automatically
>> >>>>>>> >>> >>> after the reboot. But the
service is successfully
started
>> >>>>>>> >>> >>> later
>> >>>>>>> >>> >>> manually by
>> >>>>>>> >>> >>> the user.
>> >>>>>>> >>> >>>
>> >>>>>>> >>> >>> can somebody from
gluster-users please help on this?
>> >>>>>>> >>> >>>
>> >>>>>>> >>> >>> glusterfs version: 3.5.1
>> >>>>>>> >>> >>>
>> >>>>>>> >>> >>> Thanks,
>> >>>>>>> >>> >>> Kanagaraj
>> >>>>>>> >>> >>>
>> >>>>>>> >>> >>> ----- Original Message
-----
>> >>>>>>> >>> >>> > From: "Punit
Dambiwal" <hypunit(a)gmail.com>
>> >>>>>>> >>> >>> > To:
"Kanagaraj" <kmayilsa(a)redhat.com>
>> >>>>>>> >>> >>> > Cc: users(a)ovirt.org
>> >>>>>>> >>> >>> > Sent: Tuesday,
November 25, 2014 7:24:45 PM
>> >>>>>>> >>> >>> > Subject: Re:
[ovirt-users] Gluster command [<UNKNOWN>]
>> >>>>>>> >>> >>> > failed on
>> >>>>>>> >>> >>> > server...
>> >>>>>>> >>> >>> >
>> >>>>>>> >>> >>> > Hi Kanagraj,
>> >>>>>>> >>> >>> >
>> >>>>>>> >>> >>> > Please check the
attached log files....i didn't find
any
>> >>>>>>> >>> >>> > thing
>> >>>>>>> >>> >>> > special....
>> >>>>>>> >>> >>> >
>> >>>>>>> >>> >>> > On Tue, Nov 25, 2014
at 12:12 PM, Kanagaraj
>> >>>>>>> >>> >>> >
<kmayilsa(a)redhat.com>
>> >>>>>>> >>> >>> > wrote:
>> >>>>>>> >>> >>> >
>> >>>>>>> >>> >>> > > Do you see any
errors in
>> >>>>>>> >>> >>> > >
/var/log/glusterfs/etc-glusterfs-glusterd.vol.log or
>> >>>>>>> >>> >>> > > vdsm.log
>> >>>>>>> >>> >>> > > when
>> >>>>>>> >>> >>> > > the
>> >>>>>>> >>> >>> > > service is
trying to start automatically after the
>> >>>>>>> >>> >>> > > reboot?
>> >>>>>>> >>> >>> > >
>> >>>>>>> >>> >>> > > Thanks,
>> >>>>>>> >>> >>> > > Kanagaraj
>> >>>>>>> >>> >>> > >
>> >>>>>>> >>> >>> > >
>> >>>>>>> >>> >>> > > On 11/24/2014
08:13 PM, Punit Dambiwal wrote:
>> >>>>>>> >>> >>> > >
>> >>>>>>> >>> >>> > > Hi Kanagaraj,
>> >>>>>>> >>> >>> > >
>> >>>>>>> >>> >>> > > Yes...once i
will start the gluster service and
then
>> >>>>>>> >>> >>> > > vdsmd
>> >>>>>>> >>> >>> > > ...the
>> >>>>>>> >>> >>> > > host
>> >>>>>>> >>> >>> > > can connect to
cluster...but the question is why
it's
>> >>>>>>> >>> >>> > > not
>> >>>>>>> >>> >>> > > started
>> >>>>>>> >>> >>> > > even it
>> >>>>>>> >>> >>> > > has chkconfig
enabled...
>> >>>>>>> >>> >>> > >
>> >>>>>>> >>> >>> > > I have tested
it in two host cluster
>> >>>>>>> >>> >>> > >
environment...(Centos 6.6
>> >>>>>>> >>> >>> > > and
>> >>>>>>> >>> >>> > > centos 7.0) on
both hypervisior cluster..it's failed
>> >>>>>>> >>> >>> > > to
>> >>>>>>> >>> >>> > > reconnect
>> >>>>>>> >>> >>> > > in
>> >>>>>>> >>> >>> > > to
>> >>>>>>> >>> >>> > > cluster after
reboot....
>> >>>>>>> >>> >>> > >
>> >>>>>>> >>> >>> > > In both the
environment glusterd enabled for next
>> >>>>>>> >>> >>> > > boot....but
>> >>>>>>> >>> >>> > > it's
>> >>>>>>> >>> >>> > > failed with the
same error....seems it's bug in
either
>> >>>>>>> >>> >>> > > gluster or
>> >>>>>>> >>> >>> > > Ovirt ??
>> >>>>>>> >>> >>> > >
>> >>>>>>> >>> >>> > > Please help me
to find the workaround here if can
not
>> >>>>>>> >>> >>> > > resolve
>> >>>>>>> >>> >>> > > it...as
>> >>>>>>> >>> >>> > > without this
the Host machine can not connect after
>> >>>>>>> >>> >>> > > reboot....that
>> >>>>>>> >>> >>> > > means
>> >>>>>>> >>> >>> > > engine will
consider it as down and every time need
to
>> >>>>>>> >>> >>> > > manually
>> >>>>>>> >>> >>> > > start
>> >>>>>>> >>> >>> > > the
>> >>>>>>> >>> >>> > > gluster service
and vdsmd... ??
>> >>>>>>> >>> >>> > >
>> >>>>>>> >>> >>> > > Thanks,
>> >>>>>>> >>> >>> > > Punit
>> >>>>>>> >>> >>> > >
>> >>>>>>> >>> >>> > > On Mon, Nov 24,
2014 at 10:20 PM, Kanagaraj
>> >>>>>>> >>> >>> > >
<kmayilsa(a)redhat.com>
>> >>>>>>> >>> >>> > > wrote:
>> >>>>>>> >>> >>> > >
>> >>>>>>> >>> >>> > >> From
vdsm.log "error: Connection failed. Please
>> >>>>>>> >>> >>> > >> check if
>> >>>>>>> >>> >>> > >> gluster
>> >>>>>>> >>> >>> > >> daemon
>> >>>>>>> >>> >>> > >> is
operational."
>> >>>>>>> >>> >>> > >>
>> >>>>>>> >>> >>> > >> Starting
glusterd service should fix this issue.
>> >>>>>>> >>> >>> > >>
'service
>> >>>>>>> >>> >>> > >> glusterd
>> >>>>>>> >>> >>> > >> start'
>> >>>>>>> >>> >>> > >> But i am
wondering why the glusterd was not started
>> >>>>>>> >>> >>> > >>
automatically
>> >>>>>>> >>> >>> > >> after
>> >>>>>>> >>> >>> > >> the
reboot.
>> >>>>>>> >>> >>> > >>
>> >>>>>>> >>> >>> > >> Thanks,
>> >>>>>>> >>> >>> > >> Kanagaraj
>> >>>>>>> >>> >>> > >>
>> >>>>>>> >>> >>> > >>
>> >>>>>>> >>> >>> > >>
>> >>>>>>> >>> >>> > >> On
11/24/2014 07:18 PM, Punit Dambiwal wrote:
>> >>>>>>> >>> >>> > >>
>> >>>>>>> >>> >>> > >> Hi
Kanagaraj,
>> >>>>>>> >>> >>> > >>
>> >>>>>>> >>> >>> > >> Please
find the attached VDSM logs :-
>> >>>>>>> >>> >>> > >>
>> >>>>>>> >>> >>> > >>
----------------
>> >>>>>>> >>> >>> > >>
Thread-13::DEBUG::2014-11-24
>> >>>>>>> >>> >>> > >>
>> >>>>>>> >>> >>> > >>
>> >>>>>>> >>> >>> > >>
>> >>>>>>> >>> >>> > >>
>> >>>>>>> >>> >>> > >>
21:41:17,182::resourceManager::977::Storage.ResourceManager.Owner::(cancelAll)
>> >>>>>>> >>> >>> > >>
Owner.cancelAll requests {}
>> >>>>>>> >>> >>> > >>
Thread-13::DEBUG::2014-11-24
>> >>>>>>> >>> >>> > >>
>> >>>>>>> >>> >>> > >>
>> >>>>>>> >>> >>> > >>
21:41:17,182::task::993::Storage.TaskManager.Task::(_decref)
>> >>>>>>> >>> >>> > >>
Task=`1691d409-9b27-4585-8281-5ec26154367a`::ref 0
>> >>>>>>> >>> >>> > >> aborting
>> >>>>>>> >>> >>> > >> False
>> >>>>>>> >>> >>> > >>
Thread-13::DEBUG::2014-11-24
>> >>>>>>> >>> >>> > >>
>> >>>>>>> >>> >>> > >>
>> >>>>>>> >>> >>> > >>
>> >>>>>>> >>> >>> > >>
21:41:32,393::task::595::Storage.TaskManager.Task::(_updateState)
>> >>>>>>> >>> >>> > >>
Task=`994c7bc3-a236-4d03-a732-e068c7ed9ed4`::moving
>> >>>>>>> >>> >>> > >> from
>> >>>>>>> >>> >>> > >> state
>> >>>>>>> >>> >>> > >> init
>> >>>>>>> >>> >>> > >> ->
>> >>>>>>> >>> >>> > >> state
preparing
>> >>>>>>> >>> >>> > >>
Thread-13::INFO::2014-11-24
>> >>>>>>> >>> >>> > >>
21:41:32,393::logUtils::44::dispatcher::(wrapper)
Run
>> >>>>>>> >>> >>> > >> and
>> >>>>>>> >>> >>> > >> protect:
>> >>>>>>> >>> >>> > >>
repoStats(options=None)
>> >>>>>>> >>> >>> > >>
Thread-13::INFO::2014-11-24
>> >>>>>>> >>> >>> > >>
21:41:32,393::logUtils::47::dispatcher::(wrapper)
Run
>> >>>>>>> >>> >>> > >> and
>> >>>>>>> >>> >>> > >> protect:
>> >>>>>>> >>> >>> > >> repoStats,
Return response: {}
>> >>>>>>> >>> >>> > >>
Thread-13::DEBUG::2014-11-24
>> >>>>>>> >>> >>> > >>
>> >>>>>>> >>> >>> > >>
>> >>>>>>> >>> >>> > >>
21:41:32,393::task::1191::Storage.TaskManager.Task::(prepare)
>> >>>>>>> >>> >>> > >>
>> >>>>>>> >>> >>> > >>
Task=`994c7bc3-a236-4d03-a732-e068c7ed9ed4`::finished: {}
>> >>>>>>> >>> >>> > >>
Thread-13::DEBUG::2014-11-24
>> >>>>>>> >>> >>> > >>
>> >>>>>>> >>> >>> > >>
>> >>>>>>> >>> >>> > >>
>> >>>>>>> >>> >>> > >>
21:41:32,394::task::595::Storage.TaskManager.Task::(_updateState)
>> >>>>>>> >>> >>> > >>
Task=`994c7bc3-a236-4d03-a732-e068c7ed9ed4`::moving
>> >>>>>>> >>> >>> > >> from
>> >>>>>>> >>> >>> > >> state
>> >>>>>>> >>> >>> > >> preparing
>> >>>>>>> >>> >>> > >> ->
>> >>>>>>> >>> >>> > >> state
finished
>> >>>>>>> >>> >>> > >>
Thread-13::DEBUG::2014-11-24
>> >>>>>>> >>> >>> > >>
>> >>>>>>> >>> >>> > >>
>> >>>>>>> >>> >>> > >>
>> >>>>>>> >>> >>> > >>
>> >>>>>>> >>> >>> > >>
21:41:32,394::resourceManager::940::Storage.ResourceManager.Owner::(releaseAll)
>> >>>>>>> >>> >>> > >>
Owner.releaseAll requests {} resources {}
>> >>>>>>> >>> >>> > >>
Thread-13::DEBUG::2014-11-24
>> >>>>>>> >>> >>> > >>
>> >>>>>>> >>> >>> > >>
>> >>>>>>> >>> >>> > >>
>> >>>>>>> >>> >>> > >>
>> >>>>>>> >>> >>> > >>
21:41:32,394::resourceManager::977::Storage.ResourceManager.Owner::(cancelAll)
>> >>>>>>> >>> >>> > >>
Owner.cancelAll requests {}
>> >>>>>>> >>> >>> > >>
Thread-13::DEBUG::2014-11-24
>> >>>>>>> >>> >>> > >>
>> >>>>>>> >>> >>> > >>
>> >>>>>>> >>> >>> > >>
21:41:32,394::task::993::Storage.TaskManager.Task::(_decref)
>> >>>>>>> >>> >>> > >>
Task=`994c7bc3-a236-4d03-a732-e068c7ed9ed4`::ref 0
>> >>>>>>> >>> >>> > >> aborting
>> >>>>>>> >>> >>> > >> False
>> >>>>>>> >>> >>> > >>
Thread-13::DEBUG::2014-11-24
>> >>>>>>> >>> >>> > >>
21:41:41,550::BindingXMLRPC::1132::vds::(wrapper)
>> >>>>>>> >>> >>> > >> client
>> >>>>>>> >>> >>> > >>
[10.10.10.2]::call
>> >>>>>>> >>> >>> > >>
getCapabilities with () {}
>> >>>>>>> >>> >>> > >>
Thread-13::DEBUG::2014-11-24
>> >>>>>>> >>> >>> > >>
21:41:41,553::utils::738::root::(execCmd)
>> >>>>>>> >>> >>> > >> /sbin/ip
route show to 0.0.0.0/0 table all (cwd
None)
>> >>>>>>> >>> >>> > >>
Thread-13::DEBUG::2014-11-24
>> >>>>>>> >>> >>> > >>
21:41:41,560::utils::758::root::(execCmd)
>> >>>>>>> >>> >>> > >> SUCCESS:
<err> = ''; <rc> = 0
>> >>>>>>> >>> >>> > >>
Thread-13::DEBUG::2014-11-24
>> >>>>>>> >>> >>> > >>
21:41:41,588::caps::728::root::(_getKeyPackages)
rpm
>> >>>>>>> >>> >>> > >> package
>> >>>>>>> >>> >>> > >>
('gluster-swift',) not found
>> >>>>>>> >>> >>> > >>
Thread-13::DEBUG::2014-11-24
>> >>>>>>> >>> >>> > >>
21:41:41,592::caps::728::root::(_getKeyPackages)
rpm
>> >>>>>>> >>> >>> > >> package
>> >>>>>>> >>> >>> > >>
('gluster-swift-object',) not found
>> >>>>>>> >>> >>> > >>
Thread-13::DEBUG::2014-11-24
>> >>>>>>> >>> >>> > >>
21:41:41,593::caps::728::root::(_getKeyPackages)
rpm
>> >>>>>>> >>> >>> > >> package
>> >>>>>>> >>> >>> > >>
('gluster-swift-plugin',) not found
>> >>>>>>> >>> >>> > >>
Thread-13::DEBUG::2014-11-24
>> >>>>>>> >>> >>> > >>
21:41:41,598::caps::728::root::(_getKeyPackages)
rpm
>> >>>>>>> >>> >>> > >> package
>> >>>>>>> >>> >>> > >>
('gluster-swift-account',) not found
>> >>>>>>> >>> >>> > >>
Thread-13::DEBUG::2014-11-24
>> >>>>>>> >>> >>> > >>
21:41:41,598::caps::728::root::(_getKeyPackages)
rpm
>> >>>>>>> >>> >>> > >> package
>> >>>>>>> >>> >>> > >>
('gluster-swift-proxy',) not found
>> >>>>>>> >>> >>> > >>
Thread-13::DEBUG::2014-11-24
>> >>>>>>> >>> >>> > >>
21:41:41,598::caps::728::root::(_getKeyPackages)
rpm
>> >>>>>>> >>> >>> > >> package
>> >>>>>>> >>> >>> > >>
('gluster-swift-doc',) not found
>> >>>>>>> >>> >>> > >>
Thread-13::DEBUG::2014-11-24
>> >>>>>>> >>> >>> > >>
21:41:41,599::caps::728::root::(_getKeyPackages)
rpm
>> >>>>>>> >>> >>> > >> package
>> >>>>>>> >>> >>> > >>
('gluster-swift-container',) not found
>> >>>>>>> >>> >>> > >>
Thread-13::DEBUG::2014-11-24
>> >>>>>>> >>> >>> > >>
21:41:41,599::caps::728::root::(_getKeyPackages)
rpm
>> >>>>>>> >>> >>> > >> package
>> >>>>>>> >>> >>> > >>
('glusterfs-geo-replication',) not found
>> >>>>>>> >>> >>> > >>
Thread-13::DEBUG::2014-11-24
>> >>>>>>> >>> >>> > >>
21:41:41,600::caps::646::root::(get)
>> >>>>>>> >>> >>> > >> VirtioRNG
DISABLED: libvirt version
0.10.2-29.el6_5.9
>> >>>>>>> >>> >>> > >> required
>> >>>>>>> >>> >>> > >> >=
>> >>>>>>> >>> >>> > >> 0.10.2-31
>> >>>>>>> >>> >>> > >>
Thread-13::DEBUG::2014-11-24
>> >>>>>>> >>> >>> > >>
21:41:41,603::BindingXMLRPC::1139::vds::(wrapper)
>> >>>>>>> >>> >>> > >> return
>> >>>>>>> >>> >>> > >>
getCapabilities
>> >>>>>>> >>> >>> > >> with
{'status': {'message': 'Done', 'code': 0},
>> >>>>>>> >>> >>> > >>
'info':
>> >>>>>>> >>> >>> > >>
{'HBAInventory':
>> >>>>>>> >>> >>> > >>
{'iSCSI': [{'InitiatorName':
>> >>>>>>> >>> >>> > >>
'iqn.1994-05.com.redhat:32151ce183c8'}],
>> >>>>>>> >>> >>> > >>
'FC':
>> >>>>>>> >>> >>> > >> []},
'packages2': {'kernel': {'release':
>> >>>>>>> >>> >>> > >>
'431.el6.x86_64',
>> >>>>>>> >>> >>> > >>
'buildtime':
>> >>>>>>> >>> >>> > >>
1385061309.0, 'version': '2.6.32'},
'glusterfs-rdma':
>> >>>>>>> >>> >>> > >>
{'release':
>> >>>>>>> >>> >>> > >>
'1.el6',
>> >>>>>>> >>> >>> > >>
'buildtime': 1403622628L, 'version': '3.5.1'},
>> >>>>>>> >>> >>> > >>
'glusterfs-fuse':
>> >>>>>>> >>> >>> > >>
{'release': '1.el6', 'buildtime': 1403622628L,
>> >>>>>>> >>> >>> > >>
'version':
>> >>>>>>> >>> >>> > >>
'3.5.1'},
>> >>>>>>> >>> >>> > >>
'spice-server': {'release': '6.el6_5.2',
'buildtime':
>> >>>>>>> >>> >>> > >>
1402324637L,
>> >>>>>>> >>> >>> > >>
'version': '0.12.4'}, 'vdsm': {'release':
>> >>>>>>> >>> >>> > >>
'1.gitdb83943.el6',
>> >>>>>>> >>> >>> > >>
'buildtime':
>> >>>>>>> >>> >>> > >>
1412784567L, 'version': '4.16.7'}, 'qemu-kvm':
>> >>>>>>> >>> >>> > >>
{'release':
>> >>>>>>> >>> >>> > >>
'2.415.el6_5.10', 'buildtime': 1402435700L,
>> >>>>>>> >>> >>> > >>
'version':
>> >>>>>>> >>> >>> > >>
'0.12.1.2'},
>> >>>>>>> >>> >>> > >>
'qemu-img': {'release': '2.415.el6_5.10',
>> >>>>>>> >>> >>> > >>
'buildtime':
>> >>>>>>> >>> >>> > >>
1402435700L,
>> >>>>>>> >>> >>> > >>
'version': '0.12.1.2'}, 'libvirt': {'release':
>> >>>>>>> >>> >>> > >>
'29.el6_5.9',
>> >>>>>>> >>> >>> > >>
'buildtime':
>> >>>>>>> >>> >>> > >>
1402404612L, 'version': '0.10.2'}, 'glusterfs':
>> >>>>>>> >>> >>> > >>
{'release':
>> >>>>>>> >>> >>> > >>
'1.el6',
>> >>>>>>> >>> >>> > >>
'buildtime': 1403622628L, 'version': '3.5.1'},
'mom':
>> >>>>>>> >>> >>> > >>
{'release':
>> >>>>>>> >>> >>> > >>
'2.el6',
>> >>>>>>> >>> >>> > >>
'buildtime': 1403794344L, 'version': '0.4.1'},
>> >>>>>>> >>> >>> > >>
'glusterfs-server':
>> >>>>>>> >>> >>> > >>
{'release': '1.el6', 'buildtime': 1403622628L,
>> >>>>>>> >>> >>> > >>
'version':
>> >>>>>>> >>> >>> > >>
'3.5.1'}},
>> >>>>>>> >>> >>> > >>
'numaNodeDistance': {'1': [20, 10], '0': [10, 20]},
>> >>>>>>> >>> >>> > >>
'cpuModel':
>> >>>>>>> >>> >>> > >>
'Intel(R)
>> >>>>>>> >>> >>> > >> Xeon(R) CPU
X5650 @ 2.67GHz',
'liveMerge':
>> >>>>>>> >>> >>> > >>
'false',
>> >>>>>>> >>> >>> > >>
'hooks':
>> >>>>>>> >>> >>> > >> {},
>> >>>>>>> >>> >>> > >>
'cpuSockets': '2', 'vmTypes': ['kvm'], 'selinux':
>> >>>>>>> >>> >>> > >>
{'mode': '1'},
>> >>>>>>> >>> >>> > >>
'kdumpStatus': 0, 'supportedProtocols': ['2.2',
>> >>>>>>> >>> >>> > >>
'2.3'],
>> >>>>>>> >>> >>> > >>
'networks':
>> >>>>>>> >>> >>> > >>
{'ovirtmgmt': {'iface': u'bond0.10', 'addr':
>> >>>>>>> >>> >>> > >>
'43.252.176.16',
>> >>>>>>> >>> >>> > >>
'bridged':
>> >>>>>>> >>> >>> > >> False,
'ipv6addrs':
['fe80::62eb:69ff:fe20:b46c/64'],
>> >>>>>>> >>> >>> > >>
'mtu':
>> >>>>>>> >>> >>> > >>
'1500',
>> >>>>>>> >>> >>> > >>
'bootproto4': 'none', 'netmask': '255.255.255.0',
>> >>>>>>> >>> >>> > >>
'ipv4addrs':
>> >>>>>>> >>> >>> > >> ['
>> >>>>>>> >>> >>> > >>
43.252.176.16/24' <
http://43.252.176.16/24%27>],
>> >>>>>>> >>> >>> > >>
'interface':
>> >>>>>>> >>> >>> > >>
u'bond0.10', 'ipv6gateway': '::', 'gateway':
>> >>>>>>> >>> >>> > >>
'43.25.17.1'},
>> >>>>>>> >>> >>> > >>
'Internal':
>> >>>>>>> >>> >>> > >>
{'iface': 'Internal', 'addr': '', 'cfg':
{'DEFROUTE':
>> >>>>>>> >>> >>> > >>
'no',
>> >>>>>>> >>> >>> > >>
'HOTPLUG':
>> >>>>>>> >>> >>> > >>
'no', 'MTU': '9000', 'DELAY': '0',
'NM_CONTROLLED':
>> >>>>>>> >>> >>> > >>
'no',
>> >>>>>>> >>> >>> > >>
'BOOTPROTO':
>> >>>>>>> >>> >>> > >>
'none', 'STP': 'off', 'DEVICE': 'Internal',
'TYPE':
>> >>>>>>> >>> >>> > >>
'Bridge',
>> >>>>>>> >>> >>> > >>
'ONBOOT':
>> >>>>>>> >>> >>> > >>
'no'}, 'bridged': True, 'ipv6addrs':
>> >>>>>>> >>> >>> > >>
['fe80::210:18ff:fecd:daac/64'],
>> >>>>>>> >>> >>> > >>
'gateway': '', 'bootproto4': 'none', 'netmask':
'',
>> >>>>>>> >>> >>> > >>
'stp':
>> >>>>>>> >>> >>> > >>
'off',
>> >>>>>>> >>> >>> > >>
'ipv4addrs': [], 'mtu': '9000', 'ipv6gateway':
'::',
>> >>>>>>> >>> >>> > >>
'ports':
>> >>>>>>> >>> >>> > >>
['bond1.100']}, 'storage': {'iface': u'bond1',
>> >>>>>>> >>> >>> > >>
'addr':
>> >>>>>>> >>> >>> > >>
'10.10.10.6',
>> >>>>>>> >>> >>> > >>
'bridged': False, 'ipv6addrs':
>> >>>>>>> >>> >>> > >>
['fe80::210:18ff:fecd:daac/64'],
>> >>>>>>> >>> >>> > >>
'mtu':
>> >>>>>>> >>> >>> > >>
'9000', 'bootproto4': 'none', 'netmask':
>> >>>>>>> >>> >>> > >>
'255.255.255.0',
>> >>>>>>> >>> >>> > >>
'ipv4addrs': ['
>> >>>>>>> >>> >>> > >>
10.10.10.6/24' <
http://10.10.10.6/24%27>],
>> >>>>>>> >>> >>> > >>
'interface':
>> >>>>>>> >>> >>> > >>
u'bond1',
>> >>>>>>> >>> >>> > >>
'ipv6gateway': '::', 'gateway': ''}, 'VMNetwork':
>> >>>>>>> >>> >>> > >>
{'iface':
>> >>>>>>> >>> >>> > >>
'VMNetwork',
>> >>>>>>> >>> >>> > >>
'addr': '', 'cfg': {'DEFROUTE': 'no',
'HOTPLUG':
>> >>>>>>> >>> >>> > >>
'no',
>> >>>>>>> >>> >>> > >>
'MTU':
>> >>>>>>> >>> >>> > >>
'1500',
>> >>>>>>> >>> >>> > >>
'DELAY': '0', 'NM_CONTROLLED': 'no', 'BOOTPROTO':
>> >>>>>>> >>> >>> > >>
'none',
>> >>>>>>> >>> >>> > >>
'STP':
>> >>>>>>> >>> >>> > >>
'off',
>> >>>>>>> >>> >>> > >>
'DEVICE': 'VMNetwork', 'TYPE': 'Bridge',
'ONBOOT':
>> >>>>>>> >>> >>> > >>
'no'},
>> >>>>>>> >>> >>> > >>
'bridged':
>> >>>>>>> >>> >>> > >> True,
>> >>>>>>> >>> >>> > >>
'ipv6addrs': ['fe80::62eb:69ff:fe20:b46c/64'],
>> >>>>>>> >>> >>> > >>
'gateway':
>> >>>>>>> >>> >>> > >>
'',
>> >>>>>>> >>> >>> > >>
'bootproto4':
>> >>>>>>> >>> >>> > >>
'none', 'netmask': '', 'stp': 'off',
'ipv4addrs':
[],
>> >>>>>>> >>> >>> > >>
'mtu':
>> >>>>>>> >>> >>> > >>
'1500',
>> >>>>>>> >>> >>> > >>
'ipv6gateway': '::', 'ports': ['bond0.36']}},
>> >>>>>>> >>> >>> > >>
'bridges':
>> >>>>>>> >>> >>> > >>
{'Internal':
>> >>>>>>> >>> >>> > >>
{'addr': '', 'cfg': {'DEFROUTE': 'no',
'HOTPLUG':
>> >>>>>>> >>> >>> > >>
'no',
>> >>>>>>> >>> >>> > >>
'MTU':
>> >>>>>>> >>> >>> > >>
'9000',
>> >>>>>>> >>> >>> > >>
'DELAY': '0', 'NM_CONTROLLED': 'no', 'BOOTPROTO':
>> >>>>>>> >>> >>> > >>
'none',
>> >>>>>>> >>> >>> > >>
'STP':
>> >>>>>>> >>> >>> > >>
'off',
>> >>>>>>> >>> >>> > >>
'DEVICE': 'Internal', 'TYPE': 'Bridge', 'ONBOOT':
>> >>>>>>> >>> >>> > >>
'no'},
>> >>>>>>> >>> >>> > >>
'ipv6addrs':
>> >>>>>>> >>> >>> > >>
['fe80::210:18ff:fecd:daac/64'], 'mtu': '9000',
>> >>>>>>> >>> >>> > >>
'netmask': '',
>> >>>>>>> >>> >>> > >>
'stp':
>> >>>>>>> >>> >>> > >>
'off', 'ipv4addrs': [], 'ipv6gateway': '::',
>> >>>>>>> >>> >>> > >>
'gateway':
>> >>>>>>> >>> >>> > >>
'',
>> >>>>>>> >>> >>> > >>
'opts':
>> >>>>>>> >>> >>> > >>
{'topology_change_detected': '0',
>> >>>>>>> >>> >>> > >>
'multicast_last_member_count':
>> >>>>>>> >>> >>> > >>
'2',
>> >>>>>>> >>> >>> > >>
'hash_elasticity': '4',
>> >>>>>>> >>> >>> > >>
'multicast_query_response_interval':
>> >>>>>>> >>> >>> > >>
'999',
>> >>>>>>> >>> >>> > >>
'multicast_snooping': '1',
>> >>>>>>> >>> >>> > >>
'multicast_startup_query_interval':
>> >>>>>>> >>> >>> > >>
'3124',
>> >>>>>>> >>> >>> > >>
'hello_timer': '31', 'multicast_querier_interval':
>> >>>>>>> >>> >>> > >>
'25496',
>> >>>>>>> >>> >>> > >>
'max_age':
>> >>>>>>> >>> >>> > >>
'1999', 'hash_max': '512', 'stp_state': '0',
>> >>>>>>> >>> >>> > >>
'root_id':
>> >>>>>>> >>> >>> > >>
'8000.001018cddaac', 'priority': '32768',
>> >>>>>>> >>> >>> > >>
'multicast_membership_interval':
>> >>>>>>> >>> >>> > >>
'25996', 'root_path_cost': '0', 'root_port': '0',
>> >>>>>>> >>> >>> > >>
'multicast_querier':
>> >>>>>>> >>> >>> > >>
'0',
>> >>>>>>> >>> >>> > >>
'multicast_startup_query_count': '2', 'hello_time':
>> >>>>>>> >>> >>> > >>
'199',
>> >>>>>>> >>> >>> > >>
'topology_change': '0', 'bridge_id':
>> >>>>>>> >>> >>> > >>
'8000.001018cddaac',
>> >>>>>>> >>> >>> > >>
'topology_change_timer': '0', 'ageing_time':
'29995',
>> >>>>>>> >>> >>> > >>
'gc_timer':
>> >>>>>>> >>> >>> > >>
'31',
>> >>>>>>> >>> >>> > >>
'group_addr': '1:80:c2:0:0:0', 'tcn_timer': '0',
>> >>>>>>> >>> >>> > >>
'multicast_query_interval': '12498',
>> >>>>>>> >>> >>> > >>
'multicast_last_member_interval':
>> >>>>>>> >>> >>> > >>
'99', 'multicast_router': '1', 'forward_delay':
'0'},
>> >>>>>>> >>> >>> > >>
'ports':
>> >>>>>>> >>> >>> > >>
['bond1.100']}, 'VMNetwork': {'addr': '', 'cfg':
>> >>>>>>> >>> >>> > >>
{'DEFROUTE':
>> >>>>>>> >>> >>> > >>
'no',
>> >>>>>>> >>> >>> > >>
'HOTPLUG': 'no', 'MTU': '1500', 'DELAY':
'0',
>> >>>>>>> >>> >>> > >>
'NM_CONTROLLED':
>> >>>>>>> >>> >>> > >>
'no',
>> >>>>>>> >>> >>> > >>
'BOOTPROTO': 'none', 'STP': 'off', 'DEVICE':
>> >>>>>>> >>> >>> > >>
'VMNetwork',
>> >>>>>>> >>> >>> > >>
'TYPE':
>> >>>>>>> >>> >>> > >>
'Bridge',
>> >>>>>>> >>> >>> > >>
'ONBOOT': 'no'}, 'ipv6addrs':
>> >>>>>>> >>> >>> > >>
['fe80::62eb:69ff:fe20:b46c/64'],
>> >>>>>>> >>> >>> > >>
'mtu':
>> >>>>>>> >>> >>> > >>
'1500', 'netmask': '', 'stp': 'off',
'ipv4addrs':
[],
>> >>>>>>> >>> >>> > >>
'ipv6gateway':
>> >>>>>>> >>> >>> > >>
'::',
>> >>>>>>> >>> >>> > >>
'gateway': '', 'opts': {'topology_change_detected':
>> >>>>>>> >>> >>> > >>
'0',
>> >>>>>>> >>> >>> > >>
'multicast_last_member_count': '2',
>> >>>>>>> >>> >>> > >>
'hash_elasticity':
>> >>>>>>> >>> >>> > >>
'4',
>> >>>>>>> >>> >>> > >>
'multicast_query_response_interval': '999',
>> >>>>>>> >>> >>> > >>
'multicast_snooping':
>> >>>>>>> >>> >>> > >>
'1',
>> >>>>>>> >>> >>> > >>
'multicast_startup_query_interval': '3124',
>> >>>>>>> >>> >>> > >>
'hello_timer':
>> >>>>>>> >>> >>> > >>
'131',
>> >>>>>>> >>> >>> > >>
'multicast_querier_interval': '25496', 'max_age':
>> >>>>>>> >>> >>> > >>
'1999',
>> >>>>>>> >>> >>> > >>
'hash_max':
>> >>>>>>> >>> >>> > >>
'512', 'stp_state': '0', 'root_id':
>> >>>>>>> >>> >>> > >>
'8000.60eb6920b46c',
>> >>>>>>> >>> >>> > >>
'priority':
>> >>>>>>> >>> >>> > >>
'32768', 'multicast_membership_interval': '25996',
>> >>>>>>> >>> >>> > >>
'root_path_cost':
>> >>>>>>> >>> >>> > >>
'0',
>> >>>>>>> >>> >>> > >>
'root_port': '0', 'multicast_querier': '0',
>> >>>>>>> >>> >>> > >>
'multicast_startup_query_count': '2', 'hello_time':
>> >>>>>>> >>> >>> > >>
'199',
>> >>>>>>> >>> >>> > >>
'topology_change': '0', 'bridge_id':
>> >>>>>>> >>> >>> > >>
'8000.60eb6920b46c',
>> >>>>>>> >>> >>> > >>
'topology_change_timer': '0', 'ageing_time':
'29995',
>> >>>>>>> >>> >>> > >>
'gc_timer':
>> >>>>>>> >>> >>> > >>
'31',
>> >>>>>>> >>> >>> > >>
'group_addr': '1:80:c2:0:0:0', 'tcn_timer': '0',
>> >>>>>>> >>> >>> > >>
'multicast_query_interval': '12498',
>> >>>>>>> >>> >>> > >>
'multicast_last_member_interval':
>> >>>>>>> >>> >>> > >>
'99', 'multicast_router': '1', 'forward_delay':
'0'},
>> >>>>>>> >>> >>> > >>
'ports':
>> >>>>>>> >>> >>> > >>
['bond0.36']}}, 'uuid':
>> >>>>>>> >>> >>> > >>
'44454C4C-4C00-1057-8053-B7C04F504E31',
>> >>>>>>> >>> >>> > >>
'lastClientIface': 'bond1', 'nics': {'eth3':
>> >>>>>>> >>> >>> > >>
{'permhwaddr':
>> >>>>>>> >>> >>> > >>
'00:10:18:cd:da:ae', 'addr': '', 'cfg': {'SLAVE':
>> >>>>>>> >>> >>> > >>
'yes',
>> >>>>>>> >>> >>> > >>
'NM_CONTROLLED':
>> >>>>>>> >>> >>> > >>
'no', 'MTU': '9000', 'HWADDR':
'00:10:18:cd:da:ae',
>> >>>>>>> >>> >>> > >>
'MASTER':
>> >>>>>>> >>> >>> > >>
'bond1',
>> >>>>>>> >>> >>> > >>
'DEVICE': 'eth3', 'ONBOOT': 'no'}, 'ipv6addrs':
[],
>> >>>>>>> >>> >>> > >>
'mtu':
>> >>>>>>> >>> >>> > >>
'9000',
>> >>>>>>> >>> >>> > >>
'netmask': '', 'ipv4addrs': [], 'hwaddr':
>> >>>>>>> >>> >>> > >>
'00:10:18:cd:da:ac',
>> >>>>>>> >>> >>> > >>
'speed':
>> >>>>>>> >>> >>> > >> 1000},
'eth2': {'permhwaddr': '00:10:18:cd:da:ac',
>> >>>>>>> >>> >>> > >>
'addr': '',
>> >>>>>>> >>> >>> > >>
'cfg':
>> >>>>>>> >>> >>> > >>
{'SLAVE': 'yes', 'NM_CONTROLLED': 'no', 'MTU':
>> >>>>>>> >>> >>> > >>
'9000',
>> >>>>>>> >>> >>> > >>
'HWADDR':
>> >>>>>>> >>> >>> > >>
'00:10:18:cd:da:ac', 'MASTER': 'bond1', 'DEVICE':
>> >>>>>>> >>> >>> > >>
'eth2',
>> >>>>>>> >>> >>> > >>
'ONBOOT':
>> >>>>>>> >>> >>> > >>
'no'},
>> >>>>>>> >>> >>> > >>
'ipv6addrs': [], 'mtu': '9000', 'netmask': '',
>> >>>>>>> >>> >>> > >>
'ipv4addrs': [],
>> >>>>>>> >>> >>> > >>
'hwaddr':
>> >>>>>>> >>> >>> > >>
'00:10:18:cd:da:ac', 'speed': 1000}, 'eth1':
>> >>>>>>> >>> >>> > >>
{'permhwaddr':
>> >>>>>>> >>> >>> > >>
'60:eb:69:20:b4:6d', 'addr': '', 'cfg': {'SLAVE':
>> >>>>>>> >>> >>> > >>
'yes',
>> >>>>>>> >>> >>> > >>
'NM_CONTROLLED':
>> >>>>>>> >>> >>> > >>
'no', 'MTU': '1500', 'HWADDR':
'60:eb:69:20:b4:6d',
>> >>>>>>> >>> >>> > >>
'MASTER':
>> >>>>>>> >>> >>> > >>
'bond0',
>> >>>>>>> >>> >>> > >>
'DEVICE': 'eth1', 'ONBOOT': 'yes'}, 'ipv6addrs':
[],
>> >>>>>>> >>> >>> > >>
'mtu':
>> >>>>>>> >>> >>> > >>
'1500',
>> >>>>>>> >>> >>> > >>
'netmask': '', 'ipv4addrs': [], 'hwaddr':
>> >>>>>>> >>> >>> > >>
'60:eb:69:20:b4:6c',
>> >>>>>>> >>> >>> > >>
'speed':
>> >>>>>>> >>> >>> > >> 1000},
'eth0': {'permhwaddr': '60:eb:69:20:b4:6c',
>> >>>>>>> >>> >>> > >>
'addr': '',
>> >>>>>>> >>> >>> > >>
'cfg':
>> >>>>>>> >>> >>> > >>
{'SLAVE': 'yes', 'NM_CONTROLLED': 'no', 'MTU':
>> >>>>>>> >>> >>> > >>
'1500',
>> >>>>>>> >>> >>> > >>
'HWADDR':
>> >>>>>>> >>> >>> > >>
'60:eb:69:20:b4:6c', 'MASTER': 'bond0', 'DEVICE':
>> >>>>>>> >>> >>> > >>
'eth0',
>> >>>>>>> >>> >>> > >>
'ONBOOT':
>> >>>>>>> >>> >>> > >>
'yes'},
>> >>>>>>> >>> >>> > >>
'ipv6addrs': [], 'mtu': '1500', 'netmask': '',
>> >>>>>>> >>> >>> > >>
'ipv4addrs': [],
>> >>>>>>> >>> >>> > >>
'hwaddr':
>> >>>>>>> >>> >>> > >>
'60:eb:69:20:b4:6c', 'speed': 1000}},
>> >>>>>>> >>> >>> > >>
'software_revision': '1',
>> >>>>>>> >>> >>> > >>
'clusterLevels': ['3.0', '3.1', '3.2', '3.3',
'3.4',
>> >>>>>>> >>> >>> > >>
'3.5'],
>> >>>>>>> >>> >>> > >>
'cpuFlags':
>> >>>>>>> >>> >>> > >>
>> >>>>>>> >>> >>> > >>
>> >>>>>>> >>> >>> > >>
>> >>>>>>> >>> >>> > >>
>> >>>>>>> >>> >>> > >>
u'fpu,vme,de,pse,tsc,msr,pae,mce,cx8,apic,sep,mtrr,pge,mca,cmov,pat,pse36,clflush,dts,acpi,mmx,fxsr,sse,sse2,ss,ht,tm,pbe,syscall,nx,pdpe1gb,rdtscp,lm,constant_tsc,arch_perfmon,pebs,bts,rep_good,xtopology,nonstop_tsc,pni,pclmulqdq,dtes64,monitor,ds_cpl,vmx,smx,est,tm2,ssse3,cx16,xtpr,pdcm,pcid,dca,sse4_1,sse4_2,popcnt,aes,lahf_lm,tpr_shadow,vnmi,flexpriority,ept,vpid,model_Nehalem,model_Conroe,model_coreduo,model_core2duo,model_Penryn,model_Westmere,model_n270',
>> >>>>>>> >>> >>> > >>
'ISCSIInitiatorName':
>> >>>>>>> >>> >>> > >>
'iqn.1994-05.com.redhat:32151ce183c8',
>> >>>>>>> >>> >>> > >>
'netConfigDirty': 'False', 'supportedENGINEs':
>> >>>>>>> >>> >>> > >>
['3.0',
>> >>>>>>> >>> >>> > >>
'3.1',
>> >>>>>>> >>> >>> > >>
'3.2',
>> >>>>>>> >>> >>> > >>
'3.3',
>> >>>>>>> >>> >>> > >>
'3.4', '3.5'], 'autoNumaBalancing': 2,
'reservedMem':
>> >>>>>>> >>> >>> > >>
'321',
>> >>>>>>> >>> >>> > >>
'bondings':
>> >>>>>>> >>> >>> > >>
{'bond4': {'addr': '', 'cfg': {}, 'mtu':
'1500',
>> >>>>>>> >>> >>> > >>
'netmask': '',
>> >>>>>>> >>> >>> > >>
'slaves':
>> >>>>>>> >>> >>> > >> [],
'hwaddr': '00:00:00:00:00:00'}, 'bond0':
{'addr':
>> >>>>>>> >>> >>> > >>
'',
>> >>>>>>> >>> >>> > >>
'cfg':
>> >>>>>>> >>> >>> > >>
{'HOTPLUG': 'no', 'MTU': '1500', 'NM_CONTROLLED':
>> >>>>>>> >>> >>> > >>
'no',
>> >>>>>>> >>> >>> > >>
'BONDING_OPTS':
>> >>>>>>> >>> >>> > >> 'mode=4
miimon=100', 'DEVICE': 'bond0', 'ONBOOT':
>> >>>>>>> >>> >>> > >>
'yes'},
>> >>>>>>> >>> >>> > >>
'ipv6addrs':
>> >>>>>>> >>> >>> > >>
['fe80::62eb:69ff:fe20:b46c/64'], 'mtu': '1500',
>> >>>>>>> >>> >>> > >>
'netmask': '',
>> >>>>>>> >>> >>> > >>
'ipv4addrs': [], 'hwaddr': '60:eb:69:20:b4:6c',
>> >>>>>>> >>> >>> > >>
'slaves':
>> >>>>>>> >>> >>> > >>
['eth0',
>> >>>>>>> >>> >>> > >>
'eth1'],
>> >>>>>>> >>> >>> > >>
'opts': {'miimon': '100', 'mode': '4'}},
'bond1':
>> >>>>>>> >>> >>> > >>
{'addr':
>> >>>>>>> >>> >>> > >>
'10.10.10.6',
>> >>>>>>> >>> >>> > >>
'cfg': {'DEFROUTE': 'no', 'IPADDR': '10.10.10.6',
>> >>>>>>> >>> >>> > >>
'HOTPLUG':
>> >>>>>>> >>> >>> > >>
'no',
>> >>>>>>> >>> >>> > >>
'MTU':
>> >>>>>>> >>> >>> > >>
'9000', 'NM_CONTROLLED': 'no', 'NETMASK':
>> >>>>>>> >>> >>> > >>
'255.255.255.0',
>> >>>>>>> >>> >>> > >>
'BOOTPROTO':
>> >>>>>>> >>> >>> > >>
'none', 'BONDING_OPTS': 'mode=4 miimon=100',
>> >>>>>>> >>> >>> > >>
'DEVICE':
>> >>>>>>> >>> >>> > >>
'bond1',
>> >>>>>>> >>> >>> > >>
'ONBOOT':
>> >>>>>>> >>> >>> > >>
'no'}, 'ipv6addrs':
['fe80::210:18ff:fecd:daac/64'],
>> >>>>>>> >>> >>> > >>
'mtu':
>> >>>>>>> >>> >>> > >>
'9000',
>> >>>>>>> >>> >>> > >>
'netmask': '255.255.255.0', 'ipv4addrs':
>> >>>>>>> >>> >>> > >>
['10.10.10.6/24'
>> >>>>>>> >>> >>> > >>
<
http://10.10.10.6/24%27>], 'hwaddr':
>> >>>>>>> >>> >>> > >>
'00:10:18:cd:da:ac',
>> >>>>>>> >>> >>> > >>
'slaves':
>> >>>>>>> >>> >>> > >>
['eth2', 'eth3'], 'opts': {'miimon': '100',
'mode':
>> >>>>>>> >>> >>> > >>
'4'}},
>> >>>>>>> >>> >>> > >>
'bond2':
>> >>>>>>> >>> >>> > >>
{'addr': '', 'cfg': {}, 'mtu': '1500',
'netmask':
'',
>> >>>>>>> >>> >>> > >>
'slaves':
>> >>>>>>> >>> >>> > >> [],
>> >>>>>>> >>> >>> > >>
'hwaddr': '00:00:00:00:00:00'}, 'bond3': {'addr':
'',
>> >>>>>>> >>> >>> > >>
'cfg': {},
>> >>>>>>> >>> >>> > >>
'mtu':
>> >>>>>>> >>> >>> > >>
'1500', 'netmask': '', 'slaves': [], 'hwaddr':
>> >>>>>>> >>> >>> > >>
'00:00:00:00:00:00'}},
>> >>>>>>> >>> >>> > >>
'software_version': '4.16', 'memSize': '24019',
>> >>>>>>> >>> >>> > >>
'cpuSpeed':
>> >>>>>>> >>> >>> > >>
'2667.000',
>> >>>>>>> >>> >>> > >>
'numaNodes': {u'1': {'totalMemory': '12288',
'cpus':
>> >>>>>>> >>> >>> > >> [6,
>> >>>>>>> >>> >>> > >> 7, 8,
>> >>>>>>> >>> >>> > >> 9,
>> >>>>>>> >>> >>> > >> 10, 11,
>> >>>>>>> >>> >>> > >> 18, 19, 20,
21, 22, 23]}, u'0': {'totalMemory':
>> >>>>>>> >>> >>> > >>
'12278',
>> >>>>>>> >>> >>> > >>
'cpus':
>> >>>>>>> >>> >>> > >> [0,
>> >>>>>>> >>> >>> > >> 1, 2,
>> >>>>>>> >>> >>> > >> 3, 4, 5,
12, 13, 14, 15, 16, 17]}}, 'version_name':
>> >>>>>>> >>> >>> > >> 'Snow
>> >>>>>>> >>> >>> > >> Man',
>> >>>>>>> >>> >>> > >>
'vlans':
>> >>>>>>> >>> >>> > >>
{'bond0.10': {'iface': 'bond0', 'addr':
>> >>>>>>> >>> >>> > >>
'43.25.17.16',
>> >>>>>>> >>> >>> > >>
'cfg':
>> >>>>>>> >>> >>> > >>
{'DEFROUTE':
>> >>>>>>> >>> >>> > >>
'yes', 'VLAN': 'yes', 'IPADDR': '43.25.17.16',
>> >>>>>>> >>> >>> > >>
'HOTPLUG':
>> >>>>>>> >>> >>> > >>
'no',
>> >>>>>>> >>> >>> > >>
'GATEWAY':
>> >>>>>>> >>> >>> > >>
'43.25.17.1', 'NM_CONTROLLED': 'no', 'NETMASK':
>> >>>>>>> >>> >>> > >>
'255.255.255.0',
>> >>>>>>> >>> >>> > >>
'BOOTPROTO': 'none', 'DEVICE': 'bond0.10', 'MTU':
>> >>>>>>> >>> >>> > >>
'1500',
>> >>>>>>> >>> >>> > >>
'ONBOOT':
>> >>>>>>> >>> >>> > >>
'yes'},
>> >>>>>>> >>> >>> > >>
'ipv6addrs': ['fe80::62eb:69ff:fe20:b46c/64'],
>> >>>>>>> >>> >>> > >>
'vlanid':
>> >>>>>>> >>> >>> > >> 10,
>> >>>>>>> >>> >>> > >>
'mtu':
>> >>>>>>> >>> >>> > >>
'1500',
>> >>>>>>> >>> >>> > >>
'netmask': '255.255.255.0', 'ipv4addrs':
>> >>>>>>> >>> >>> > >>
['43.25.17.16/24']
>> >>>>>>> >>> >>> > >>
<
http://43.25.17.16/24%27%5D>}, 'bond0.36':
{'iface':
>> >>>>>>> >>> >>> > >>
'bond0',
>> >>>>>>> >>> >>> > >>
'addr':
>> >>>>>>> >>> >>> > >> '',
'cfg': {'BRIDGE': 'VMNetwork', 'VLAN': 'yes',
>> >>>>>>> >>> >>> > >>
'HOTPLUG':
>> >>>>>>> >>> >>> > >>
'no',
>> >>>>>>> >>> >>> > >>
'MTU':
>> >>>>>>> >>> >>> > >>
'1500', 'NM_CONTROLLED': 'no', 'DEVICE':
'bond0.36',
>> >>>>>>> >>> >>> > >>
'ONBOOT':
>> >>>>>>> >>> >>> > >>
'no'},
>> >>>>>>> >>> >>> > >>
'ipv6addrs': ['fe80::62eb:69ff:fe20:b46c/64'],
>> >>>>>>> >>> >>> > >>
'vlanid':
>> >>>>>>> >>> >>> > >> 36,
>> >>>>>>> >>> >>> > >>
'mtu':
>> >>>>>>> >>> >>> > >>
'1500',
>> >>>>>>> >>> >>> > >>
'netmask': '', 'ipv4addrs': []}, 'bond1.100':
>> >>>>>>> >>> >>> > >>
{'iface':
>> >>>>>>> >>> >>> > >>
'bond1',
>> >>>>>>> >>> >>> > >>
'addr':
>> >>>>>>> >>> >>> > >> '',
'cfg': {'BRIDGE': 'Internal', 'VLAN': 'yes',
>> >>>>>>> >>> >>> > >>
'HOTPLUG':
>> >>>>>>> >>> >>> > >>
'no',
>> >>>>>>> >>> >>> > >>
'MTU':
>> >>>>>>> >>> >>> > >>
'9000', 'NM_CONTROLLED': 'no', 'DEVICE':
'bond1.100',
>> >>>>>>> >>> >>> > >>
'ONBOOT':
>> >>>>>>> >>> >>> > >>
'no'},
>> >>>>>>> >>> >>> > >>
'ipv6addrs': ['fe80::210:18ff:fecd:daac/64'],
>> >>>>>>> >>> >>> > >>
'vlanid':
>> >>>>>>> >>> >>> > >> 100,
>> >>>>>>> >>> >>> > >>
'mtu':
>> >>>>>>> >>> >>> > >>
'9000',
>> >>>>>>> >>> >>> > >>
'netmask': '', 'ipv4addrs': []}}, 'cpuCores':
'12',
>> >>>>>>> >>> >>> > >>
'kvmEnabled':
>> >>>>>>> >>> >>> > >>
'true',
>> >>>>>>> >>> >>> > >>
'guestOverhead': '65', 'cpuThreads': '24',
>> >>>>>>> >>> >>> > >>
'emulatedMachines':
>> >>>>>>> >>> >>> > >>
[u'rhel6.5.0', u'pc', u'rhel6.4.0', u'rhel6.3.0',
>> >>>>>>> >>> >>> > >>
u'rhel6.2.0',
>> >>>>>>> >>> >>> > >>
u'rhel6.1.0', u'rhel6.0.0', u'rhel5.5.0',
>> >>>>>>> >>> >>> > >>
u'rhel5.4.4',
>> >>>>>>> >>> >>> > >>
u'rhel5.4.0'],
>> >>>>>>> >>> >>> > >>
'operatingSystem': {'release': '5.el6.centos.11.1',
>> >>>>>>> >>> >>> > >>
'version':
>> >>>>>>> >>> >>> > >>
'6',
>> >>>>>>> >>> >>> > >>
'name':
>> >>>>>>> >>> >>> > >>
'RHEL'}, 'lastClient': '10.10.10.2'}}
>> >>>>>>> >>> >>> > >>
Thread-13::DEBUG::2014-11-24
>> >>>>>>> >>> >>> > >>
21:41:41,620::BindingXMLRPC::1132::vds::(wrapper)
>> >>>>>>> >>> >>> > >> client
>> >>>>>>> >>> >>> > >>
[10.10.10.2]::call
>> >>>>>>> >>> >>> > >>
getHardwareInfo with () {}
>> >>>>>>> >>> >>> > >>
Thread-13::DEBUG::2014-11-24
>> >>>>>>> >>> >>> > >>
21:41:41,621::BindingXMLRPC::1139::vds::(wrapper)
>> >>>>>>> >>> >>> > >> return
>> >>>>>>> >>> >>> > >>
getHardwareInfo
>> >>>>>>> >>> >>> > >> with
{'status': {'message': 'Done', 'code': 0},
>> >>>>>>> >>> >>> > >>
'info':
>> >>>>>>> >>> >>> > >>
{'systemProductName': 'CS24-TY',
>> >>>>>>> >>> >>> > >>
'systemSerialNumber':
>> >>>>>>> >>> >>> > >>
'7LWSPN1',
>> >>>>>>> >>> >>> > >>
'systemFamily': 'Server', 'systemVersion': 'A00',
>> >>>>>>> >>> >>> > >>
'systemUUID':
>> >>>>>>> >>> >>> > >>
'44454c4c-4c00-1057-8053-b7c04f504e31',
>> >>>>>>> >>> >>> > >>
'systemManufacturer':
>> >>>>>>> >>> >>> > >>
'Dell'}}
>> >>>>>>> >>> >>> > >>
Thread-13::DEBUG::2014-11-24
>> >>>>>>> >>> >>> > >>
21:41:41,733::BindingXMLRPC::1132::vds::(wrapper)
>> >>>>>>> >>> >>> > >> client
>> >>>>>>> >>> >>> > >>
[10.10.10.2]::call
>> >>>>>>> >>> >>> > >> hostsList
with () {} flowID [222e8036]
>> >>>>>>> >>> >>> > >>
Thread-13::ERROR::2014-11-24
>> >>>>>>> >>> >>> > >>
21:41:44,753::BindingXMLRPC::1148::vds::(wrapper)
>> >>>>>>> >>> >>> > >> vdsm
>> >>>>>>> >>> >>> > >> exception
>> >>>>>>> >>> >>> > >> occured
>> >>>>>>> >>> >>> > >> Traceback
(most recent call last):
>> >>>>>>> >>> >>> > >> File
"/usr/share/vdsm/rpc/BindingXMLRPC.py", line
>> >>>>>>> >>> >>> > >> 1135,
>> >>>>>>> >>> >>> > >> in
>> >>>>>>> >>> >>> > >> wrapper
>> >>>>>>> >>> >>> > >> res =
f(*args, **kwargs)
>> >>>>>>> >>> >>> > >> File
"/usr/share/vdsm/gluster/api.py", line 54,
in
>> >>>>>>> >>> >>> > >> wrapper
>> >>>>>>> >>> >>> > >> rv =
func(*args, **kwargs)
>> >>>>>>> >>> >>> > >> File
"/usr/share/vdsm/gluster/api.py", line 251,
in
>> >>>>>>> >>> >>> > >> hostsList
>> >>>>>>> >>> >>> > >> return
{'hosts':
>> >>>>>>> >>> >>> > >>
self.svdsmProxy.glusterPeerStatus()}
>> >>>>>>> >>> >>> > >> File
"/usr/share/vdsm/supervdsm.py", line 50, in
>> >>>>>>> >>> >>> > >> __call__
>> >>>>>>> >>> >>> > >> return
callMethod()
>> >>>>>>> >>> >>> > >> File
"/usr/share/vdsm/supervdsm.py", line 48, in
>> >>>>>>> >>> >>> > >>
<lambda>
>> >>>>>>> >>> >>> > >>
**kwargs)
>> >>>>>>> >>> >>> > >> File
"<string>", line 2, in glusterPeerStatus
>> >>>>>>> >>> >>> > >> File
>> >>>>>>> >>> >>> > >>
"/usr/lib64/python2.6/multiprocessing/managers.py",
>> >>>>>>> >>> >>> > >> line
>> >>>>>>> >>> >>> > >> 740,
>> >>>>>>> >>> >>> > >> in
>> >>>>>>> >>> >>> > >>
_callmethod
>> >>>>>>> >>> >>> > >> raise
convert_to_error(kind, result)
>> >>>>>>> >>> >>> > >>
GlusterCmdExecFailedException: Command execution
>> >>>>>>> >>> >>> > >> failed
>> >>>>>>> >>> >>> > >> error:
Connection failed. Please check if gluster
>> >>>>>>> >>> >>> > >> daemon
>> >>>>>>> >>> >>> > >> is
>> >>>>>>> >>> >>> > >>
operational.
>> >>>>>>> >>> >>> > >> return
code: 1
>> >>>>>>> >>> >>> > >>
Thread-13::DEBUG::2014-11-24
>> >>>>>>> >>> >>> > >>
>> >>>>>>> >>> >>> > >>
>> >>>>>>> >>> >>> > >>
>> >>>>>>> >>> >>> > >>
21:41:50,949::task::595::Storage.TaskManager.Task::(_updateState)
>> >>>>>>> >>> >>> > >>
Task=`c9042986-c978-4b08-adb2-616f5299e115`::moving
>> >>>>>>> >>> >>> > >> from
>> >>>>>>> >>> >>> > >> state
>> >>>>>>> >>> >>> > >> init
>> >>>>>>> >>> >>> > >> ->
>> >>>>>>> >>> >>> > >> state
preparing
>> >>>>>>> >>> >>> > >>
Thread-13::INFO::2014-11-24
>> >>>>>>> >>> >>> > >>
21:41:50,950::logUtils::44::dispatcher::(wrapper)
Run
>> >>>>>>> >>> >>> > >> and
>> >>>>>>> >>> >>> > >> protect:
>> >>>>>>> >>> >>> > >>
repoStats(options=None)
>> >>>>>>> >>> >>> > >>
Thread-13::INFO::2014-11-24
>> >>>>>>> >>> >>> > >>
21:41:50,950::logUtils::47::dispatcher::(wrapper)
Run
>> >>>>>>> >>> >>> > >> and
>> >>>>>>> >>> >>> > >> protect:
>> >>>>>>> >>> >>> > >> repoStats,
Return response: {}
>> >>>>>>> >>> >>> > >>
Thread-13::DEBUG::2014-11-24
>> >>>>>>> >>> >>> > >>
>> >>>>>>> >>> >>> > >>
>> >>>>>>> >>> >>> > >>
21:41:50,950::task::1191::Storage.TaskManager.Task::(prepare)
>> >>>>>>> >>> >>> > >>
>> >>>>>>> >>> >>> > >>
Task=`c9042986-c978-4b08-adb2-616f5299e115`::finished: {}
>> >>>>>>> >>> >>> > >>
Thread-13::DEBUG::2014-11-24
>> >>>>>>> >>> >>> > >>
>> >>>>>>> >>> >>> > >>
>> >>>>>>> >>> >>> > >>
>> >>>>>>> >>> >>> > >>
21:41:50,950::task::595::Storage.TaskManager.Task::(_updateState)
>> >>>>>>> >>> >>> > >>
Task=`c9042986-c978-4b08-adb2-616f5299e115`::moving
>> >>>>>>> >>> >>> > >> from
>> >>>>>>> >>> >>> > >> state
>> >>>>>>> >>> >>> > >> preparing
>> >>>>>>> >>> >>> > >> ->
>> >>>>>>> >>> >>> > >> state
finished
>> >>>>>>> >>> >>> > >>
Thread-13::DEBUG::2014-11-24
>> >>>>>>> >>> >>> > >>
>> >>>>>>> >>> >>> > >>
>> >>>>>>> >>> >>> > >>
>> >>>>>>> >>> >>> > >>
>> >>>>>>> >>> >>> > >>
21:41:50,951::resourceManager::940::Storage.ResourceManager.Owner::(releaseAll)
>> >>>>>>> >>> >>> > >>
Owner.releaseAll requests {} resources {}
>> >>>>>>> >>> >>> > >>
Thread-13::DEBUG::2014-11-24
>> >>>>>>> >>> >>> > >>
>> >>>>>>> >>> >>> > >>
>> >>>>>>> >>> >>> > >>
>> >>>>>>> >>> >>> > >>
>> >>>>>>> >>> >>> > >>
21:41:50,951::resourceManager::977::Storage.ResourceManager.Owner::(cancelAll)
>> >>>>>>> >>> >>> > >>
Owner.cancelAll requests {}
>> >>>>>>> >>> >>> > >>
Thread-13::DEBUG::2014-11-24
>> >>>>>>> >>> >>> > >>
>> >>>>>>> >>> >>> > >>
>> >>>>>>> >>> >>> > >>
21:41:50,951::task::993::Storage.TaskManager.Task::(_decref)
>> >>>>>>> >>> >>> > >>
Task=`c9042986-c978-4b08-adb2-616f5299e115`::ref 0
>> >>>>>>> >>> >>> > >> aborting
>> >>>>>>> >>> >>> > >> False
>> >>>>>>> >>> >>> > >>
-------------------------------
>> >>>>>>> >>> >>> > >>
>> >>>>>>> >>> >>> > >>
[root@compute4 ~]# service glusterd status
>> >>>>>>> >>> >>> > >> glusterd is
stopped
>> >>>>>>> >>> >>> > >>
[root@compute4 ~]# chkconfig --list | grep
glusterd
>> >>>>>>> >>> >>> > >> glusterd
0:off 1:off 2:on 3:on
4:on
>> >>>>>>> >>> >>> > >> 5:on
>> >>>>>>> >>> >>> > >> 6:off
>> >>>>>>> >>> >>> > >>
[root@compute4 ~]#
>> >>>>>>> >>> >>> > >>
>> >>>>>>> >>> >>> > >> Thanks,
>> >>>>>>> >>> >>> > >> Punit
>> >>>>>>> >>> >>> > >>
>> >>>>>>> >>> >>> > >> On Mon, Nov
24, 2014 at 6:36 PM, Kanagaraj
>> >>>>>>> >>> >>> > >>
<kmayilsa(a)redhat.com>
>> >>>>>>> >>> >>> > >> wrote:
>> >>>>>>> >>> >>> > >>
>> >>>>>>> >>> >>> > >>> Can
you send the corresponding error in vdsm.log
>> >>>>>>> >>> >>> > >>> from
>> >>>>>>> >>> >>> > >>> the
>> >>>>>>> >>> >>> > >>> host?
>> >>>>>>> >>> >>> > >>>
>> >>>>>>> >>> >>> > >>> Also
check if glusterd service is running.
>> >>>>>>> >>> >>> > >>>
>> >>>>>>> >>> >>> > >>>
Thanks,
>> >>>>>>> >>> >>> > >>>
Kanagaraj
>> >>>>>>> >>> >>> > >>>
>> >>>>>>> >>> >>> > >>>
>> >>>>>>> >>> >>> > >>> On
11/24/2014 03:39 PM, Punit Dambiwal wrote:
>> >>>>>>> >>> >>> > >>>
>> >>>>>>> >>> >>> > >>> Hi,
>> >>>>>>> >>> >>> > >>>
>> >>>>>>> >>> >>> > >>> After
reboot my Hypervisior host can not activate
>> >>>>>>> >>> >>> > >>> again
>> >>>>>>> >>> >>> > >>> in the
>> >>>>>>> >>> >>> > >>>
cluster
>> >>>>>>> >>> >>> > >>> and
failed with the following error :-
>> >>>>>>> >>> >>> > >>>
>> >>>>>>> >>> >>> > >>>
Gluster command [<UNKNOWN>] failed on server...
>> >>>>>>> >>> >>> > >>>
>> >>>>>>> >>> >>> > >>> Engine
logs :-
>> >>>>>>> >>> >>> > >>>
>> >>>>>>> >>> >>> > >>>
2014-11-24 18:05:28,397 INFO
>> >>>>>>> >>> >>> > >>>
>> >>>>>>> >>> >>> > >>>
>> >>>>>>> >>> >>> > >>>
>> >>>>>>> >>> >>> > >>>
>> >>>>>>> >>> >>> > >>>
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand]
>> >>>>>>> >>> >>> > >>>
(DefaultQuartzScheduler_Worker-64) START,
>> >>>>>>> >>> >>> > >>>
GlusterVolumesListVDSCommand(HostName = Compute4,
>> >>>>>>> >>> >>> > >>> HostId
>> >>>>>>> >>> >>> > >>> =
>> >>>>>>> >>> >>> > >>>
33648a90-200c-45ca-89d5-1ce305d79a6a), log id:
>> >>>>>>> >>> >>> > >>>
5f251c90
>> >>>>>>> >>> >>> > >>>
2014-11-24 18:05:30,609 INFO
>> >>>>>>> >>> >>> > >>>
>> >>>>>>> >>> >>> > >>>
>> >>>>>>> >>> >>> > >>>
>> >>>>>>> >>> >>> > >>>
>> >>>>>>> >>> >>> > >>>
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand]
>> >>>>>>> >>> >>> > >>>
(DefaultQuartzScheduler_Worker-64) FINISH,
>> >>>>>>> >>> >>> > >>>
GlusterVolumesListVDSCommand,
>> >>>>>>> >>> >>> > >>>
return:
>> >>>>>>> >>> >>> > >>>
>> >>>>>>> >>> >>> > >>>
>> >>>>>>> >>> >>> > >>>
>> >>>>>>> >>> >>> > >>>
>> >>>>>>> >>> >>> > >>>
{26ae1672-ee09-4a38-8fd2-72dd9974cc2b=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@d95203e0
},
>> >>>>>>> >>> >>> > >>> log id:
5f251c90
>> >>>>>>> >>> >>> > >>>
2014-11-24 18:05:33,768 INFO
>> >>>>>>> >>> >>> > >>>
[org.ovirt.engine.core.bll.ActivateVdsCommand]
>> >>>>>>> >>> >>> > >>>
(ajp--127.0.0.1-8702-8)
>> >>>>>>> >>> >>> > >>>
[287d570d] Lock Acquired to object EngineLock
>> >>>>>>> >>> >>> > >>>
[exclusiveLocks=
>> >>>>>>> >>> >>> > >>> key:
>> >>>>>>> >>> >>> > >>>
0bf6b00f-7947-4411-b55a-cc5eea2b381a value: VDS
>> >>>>>>> >>> >>> > >>> ,
sharedLocks= ]
>> >>>>>>> >>> >>> > >>>
2014-11-24 18:05:33,795 INFO
>> >>>>>>> >>> >>> > >>>
[org.ovirt.engine.core.bll.ActivateVdsCommand]
>> >>>>>>> >>> >>> > >>>
(org.ovirt.thread.pool-8-thread-45) [287d570d]
>> >>>>>>> >>> >>> > >>>
Running
>> >>>>>>> >>> >>> > >>>
command:
>> >>>>>>> >>> >>> > >>>
ActivateVdsCommand internal: false. Entities
>> >>>>>>> >>> >>> > >>>
affected :
>> >>>>>>> >>> >>> > >>> ID:
>> >>>>>>> >>> >>> > >>>
0bf6b00f-7947-4411-b55a-cc5eea2b381a Type:
VDSAction
>> >>>>>>> >>> >>> > >>> group
>> >>>>>>> >>> >>> > >>>
MANIPULATE_HOST
>> >>>>>>> >>> >>> > >>> with
role type ADMIN
>> >>>>>>> >>> >>> > >>>
2014-11-24 18:05:33,796 INFO
>> >>>>>>> >>> >>> > >>>
[org.ovirt.engine.core.bll.ActivateVdsCommand]
>> >>>>>>> >>> >>> > >>>
(org.ovirt.thread.pool-8-thread-45) [287d570d]
>> >>>>>>> >>> >>> > >>> Before
>> >>>>>>> >>> >>> > >>>
acquiring
>> >>>>>>> >>> >>> > >>> lock
in
>> >>>>>>> >>> >>> > >>> order
to prevent monitoring for host Compute5 from
>> >>>>>>> >>> >>> > >>>
data-center
>> >>>>>>> >>> >>> > >>> SV_WTC
>> >>>>>>> >>> >>> > >>>
2014-11-24 18:05:33,797 INFO
>> >>>>>>> >>> >>> > >>>
[org.ovirt.engine.core.bll.ActivateVdsCommand]
>> >>>>>>> >>> >>> > >>>
(org.ovirt.thread.pool-8-thread-45) [287d570d]
Lock
>> >>>>>>> >>> >>> > >>>
acquired,
>> >>>>>>> >>> >>> > >>> from
>> >>>>>>> >>> >>> > >>> now a
>> >>>>>>> >>> >>> > >>>
monitoring of host will be skipped for host
Compute5
>> >>>>>>> >>> >>> > >>> from
>> >>>>>>> >>> >>> > >>>
data-center
>> >>>>>>> >>> >>> > >>> SV_WTC
>> >>>>>>> >>> >>> > >>>
2014-11-24 18:05:33,817 INFO
>> >>>>>>> >>> >>> > >>>
>> >>>>>>> >>> >>> > >>>
>> >>>>>>> >>> >>> > >>>
[org.ovirt.engine.core.vdsbroker.SetVdsStatusVDSCommand]
>> >>>>>>> >>> >>> > >>>
(org.ovirt.thread.pool-8-thread-45) [287d570d]
>> >>>>>>> >>> >>> > >>> START,
>> >>>>>>> >>> >>> > >>>
SetVdsStatusVDSCommand(HostName = Compute5,
HostId =
>> >>>>>>> >>> >>> > >>>
0bf6b00f-7947-4411-b55a-cc5eea2b381a,
>> >>>>>>> >>> >>> > >>>
status=Unassigned,
>> >>>>>>> >>> >>> > >>>
nonOperationalReason=NONE,
>> >>>>>>> >>> >>> > >>>
stopSpmFailureLogged=false),
>> >>>>>>> >>> >>> > >>> log
id:
>> >>>>>>> >>> >>> > >>>
1cbc7311
>> >>>>>>> >>> >>> > >>>
2014-11-24 18:05:33,820 INFO
>> >>>>>>> >>> >>> > >>>
>> >>>>>>> >>> >>> > >>>
>> >>>>>>> >>> >>> > >>>
[org.ovirt.engine.core.vdsbroker.SetVdsStatusVDSCommand]
>> >>>>>>> >>> >>> > >>>
(org.ovirt.thread.pool-8-thread-45) [287d570d]
>> >>>>>>> >>> >>> > >>>
FINISH,
>> >>>>>>> >>> >>> > >>>
SetVdsStatusVDSCommand, log id: 1cbc7311
>> >>>>>>> >>> >>> > >>>
2014-11-24 18:05:34,086 INFO
>> >>>>>>> >>> >>> > >>>
[org.ovirt.engine.core.bll.ActivateVdsCommand]
>> >>>>>>> >>> >>> > >>>
(org.ovirt.thread.pool-8-thread-45) Activate
>> >>>>>>> >>> >>> > >>>
finished.
>> >>>>>>> >>> >>> > >>> Lock
>> >>>>>>> >>> >>> > >>>
released.
>> >>>>>>> >>> >>> > >>>
Monitoring can run now for host Compute5 from
>> >>>>>>> >>> >>> > >>>
data-center
>> >>>>>>> >>> >>> > >>> SV_WTC
>> >>>>>>> >>> >>> > >>>
2014-11-24 18:05:34,088 INFO
>> >>>>>>> >>> >>> > >>>
>> >>>>>>> >>> >>> > >>>
>> >>>>>>> >>> >>> > >>>
>> >>>>>>> >>> >>> > >>>
>> >>>>>>> >>> >>> > >>>
[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
>> >>>>>>> >>> >>> > >>>
(org.ovirt.thread.pool-8-thread-45) Correlation
ID:
>> >>>>>>> >>> >>> > >>>
287d570d,
>> >>>>>>> >>> >>> > >>> Job
>> >>>>>>> >>> >>> > >>> ID:
>> >>>>>>> >>> >>> > >>>
5ef8e4d6-b2bc-469e-8e81-7ef74b2a001a, Call Stack:
>> >>>>>>> >>> >>> > >>> null,
>> >>>>>>> >>> >>> > >>> Custom
>> >>>>>>> >>> >>> > >>> Event
ID:
>> >>>>>>> >>> >>> > >>> -1,
Message: Host Compute5 was activated by admin.
>> >>>>>>> >>> >>> > >>>
2014-11-24 18:05:34,090 INFO
>> >>>>>>> >>> >>> > >>>
[org.ovirt.engine.core.bll.ActivateVdsCommand]
>> >>>>>>> >>> >>> > >>>
(org.ovirt.thread.pool-8-thread-45) Lock freed to
>> >>>>>>> >>> >>> > >>> object
>> >>>>>>> >>> >>> > >>>
EngineLock
>> >>>>>>> >>> >>> > >>>
[exclusiveLocks= key:
>> >>>>>>> >>> >>> > >>>
0bf6b00f-7947-4411-b55a-cc5eea2b381a
>> >>>>>>> >>> >>> > >>> value:
>> >>>>>>> >>> >>> > >>> VDS
>> >>>>>>> >>> >>> > >>> ,
sharedLocks= ]
>> >>>>>>> >>> >>> > >>>
2014-11-24 18:05:35,792 INFO
>> >>>>>>> >>> >>> > >>>
>> >>>>>>> >>> >>> > >>>
>> >>>>>>> >>> >>> > >>>
>> >>>>>>> >>> >>> > >>>
>> >>>>>>> >>> >>> > >>>
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand]
>> >>>>>>> >>> >>> > >>>
(DefaultQuartzScheduler_Worker-55) [3706e836]
START,
>> >>>>>>> >>> >>> > >>>
GlusterVolumesListVDSCommand(HostName = Compute4,
>> >>>>>>> >>> >>> > >>> HostId
>> >>>>>>> >>> >>> > >>> =
>> >>>>>>> >>> >>> > >>>
33648a90-200c-45ca-89d5-1ce305d79a6a), log id:
>> >>>>>>> >>> >>> > >>>
48a0c832
>> >>>>>>> >>> >>> > >>>
2014-11-24 18:05:37,064 INFO
>> >>>>>>> >>> >>> > >>>
>> >>>>>>> >>> >>> > >>>
>> >>>>>>> >>> >>> > >>>
>> >>>>>>> >>> >>> > >>>
>> >>>>>>> >>> >>> > >>>
[org.ovirt.engine.core.vdsbroker.vdsbroker.GetHardwareInfoVDSCommand]
>> >>>>>>> >>> >>> > >>>
(DefaultQuartzScheduler_Worker-69) START,
>> >>>>>>> >>> >>> > >>>
GetHardwareInfoVDSCommand(HostName = Compute5,
>> >>>>>>> >>> >>> > >>> HostId
=
>> >>>>>>> >>> >>> > >>>
0bf6b00f-7947-4411-b55a-cc5eea2b381a,
>> >>>>>>> >>> >>> > >>>
>> >>>>>>> >>> >>> > >>>
>> >>>>>>> >>> >>> > >>>
vds=Host[Compute5,0bf6b00f-7947-4411-b55a-cc5eea2b381a]), log
>> >>>>>>> >>> >>> > >>> id:
>> >>>>>>> >>> >>> > >>>
6d560cc2
>> >>>>>>> >>> >>> > >>>
2014-11-24 18:05:37,074 INFO
>> >>>>>>> >>> >>> > >>>
>> >>>>>>> >>> >>> > >>>
>> >>>>>>> >>> >>> > >>>
>> >>>>>>> >>> >>> > >>>
>> >>>>>>> >>> >>> > >>>
[org.ovirt.engine.core.vdsbroker.vdsbroker.GetHardwareInfoVDSCommand]
>> >>>>>>> >>> >>> > >>>
(DefaultQuartzScheduler_Worker-69) FINISH,
>> >>>>>>> >>> >>> > >>>
GetHardwareInfoVDSCommand, log
>> >>>>>>> >>> >>> > >>> id:
6d560cc2
>> >>>>>>> >>> >>> > >>>
2014-11-24 18:05:37,093 WARN
>> >>>>>>> >>> >>> > >>>
[org.ovirt.engine.core.vdsbroker.VdsManager]
>> >>>>>>> >>> >>> > >>>
(DefaultQuartzScheduler_Worker-69) Host Compute5
is
>> >>>>>>> >>> >>> > >>>
running
>> >>>>>>> >>> >>> > >>> with
>> >>>>>>> >>> >>> > >>>
disabled
>> >>>>>>> >>> >>> > >>>
SELinux.
>> >>>>>>> >>> >>> > >>>
2014-11-24 18:05:37,127 INFO
>> >>>>>>> >>> >>> > >>>
>> >>>>>>> >>> >>> > >>>
>> >>>>>>> >>> >>> > >>>
>> >>>>>>> >>> >>> > >>>
>> >>>>>>> >>> >>> > >>>
[org.ovirt.engine.core.bll.HandleVdsCpuFlagsOrClusterChangedCommand]
>> >>>>>>> >>> >>> > >>>
(DefaultQuartzScheduler_Worker-69) [2b4a51cf]
>> >>>>>>> >>> >>> > >>>
Running
>> >>>>>>> >>> >>> > >>>
command:
>> >>>>>>> >>> >>> > >>>
HandleVdsCpuFlagsOrClusterChangedCommand internal:
>> >>>>>>> >>> >>> > >>> true.
>> >>>>>>> >>> >>> > >>>
Entities
>> >>>>>>> >>> >>> > >>>
affected
>> >>>>>>> >>> >>> > >>> : ID:
0bf6b00f-7947-4411-b55a-cc5eea2b381a Type:
>> >>>>>>> >>> >>> > >>> VDS
>> >>>>>>> >>> >>> > >>>
2014-11-24 18:05:37,147 INFO
>> >>>>>>> >>> >>> > >>>
>> >>>>>>> >>> >>> > >>>
>> >>>>>>> >>> >>> > >>>
>> >>>>>>> >>> >>> > >>>
>> >>>>>>> >>> >>> > >>>
[org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand]
>> >>>>>>> >>> >>> > >>>
(DefaultQuartzScheduler_Worker-69) [2b4a51cf]
START,
>> >>>>>>> >>> >>> > >>>
GlusterServersListVDSCommand(HostName = Compute5,
>> >>>>>>> >>> >>> > >>> HostId
>> >>>>>>> >>> >>> > >>> =
>> >>>>>>> >>> >>> > >>>
0bf6b00f-7947-4411-b55a-cc5eea2b381a), log id:
>> >>>>>>> >>> >>> > >>>
4faed87
>> >>>>>>> >>> >>> > >>>
2014-11-24 18:05:37,164 INFO
>> >>>>>>> >>> >>> > >>>
>> >>>>>>> >>> >>> > >>>
>> >>>>>>> >>> >>> > >>>
>> >>>>>>> >>> >>> > >>>
>> >>>>>>> >>> >>> > >>>
[org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand]
>> >>>>>>> >>> >>> > >>>
(DefaultQuartzScheduler_Worker-69) [2b4a51cf]
>> >>>>>>> >>> >>> > >>>
FINISH,
>> >>>>>>> >>> >>> > >>>
GlusterServersListVDSCommand, log id: 4faed87
>> >>>>>>> >>> >>> > >>>
2014-11-24 18:05:37,189 INFO
>> >>>>>>> >>> >>> > >>>
>> >>>>>>> >>> >>> > >>>
[org.ovirt.engine.core.bll.SetNonOperationalVdsCommand]
>> >>>>>>> >>> >>> > >>>
(DefaultQuartzScheduler_Worker-69) [4a84c4e5]
>> >>>>>>> >>> >>> > >>>
Running
>> >>>>>>> >>> >>> > >>>
command:
>> >>>>>>> >>> >>> > >>>
SetNonOperationalVdsCommand internal: true.
Entities
>> >>>>>>> >>> >>> > >>>
affected :
>> >>>>>>> >>> >>> > >>> ID:
>> >>>>>>> >>> >>> > >>>
0bf6b00f-7947-4411-b55a-cc5eea2b381a Type: VDS
>> >>>>>>> >>> >>> > >>>
2014-11-24 18:05:37,206 INFO
>> >>>>>>> >>> >>> > >>>
>> >>>>>>> >>> >>> > >>>
>> >>>>>>> >>> >>> > >>>
[org.ovirt.engine.core.vdsbroker.SetVdsStatusVDSCommand]
>> >>>>>>> >>> >>> > >>>
(DefaultQuartzScheduler_Worker-69) [4a84c4e5]
START,
>> >>>>>>> >>> >>> > >>>
SetVdsStatusVDSCommand(HostName = Compute5,
HostId =
>> >>>>>>> >>> >>> > >>>
0bf6b00f-7947-4411-b55a-cc5eea2b381a,
>> >>>>>>> >>> >>> > >>>
status=NonOperational,
>> >>>>>>> >>> >>> > >>>
nonOperationalReason=GLUSTER_COMMAND_FAILED,
>> >>>>>>> >>> >>> > >>>
stopSpmFailureLogged=false),
>> >>>>>>> >>> >>> > >>> log id:
fed5617
>> >>>>>>> >>> >>> > >>>
2014-11-24 18:05:37,209 INFO
>> >>>>>>> >>> >>> > >>>
>> >>>>>>> >>> >>> > >>>
>> >>>>>>> >>> >>> > >>>
[org.ovirt.engine.core.vdsbroker.SetVdsStatusVDSCommand]
>> >>>>>>> >>> >>> > >>>
(DefaultQuartzScheduler_Worker-69) [4a84c4e5]
>> >>>>>>> >>> >>> > >>>
FINISH,
>> >>>>>>> >>> >>> > >>>
SetVdsStatusVDSCommand, log id: fed5617
>> >>>>>>> >>> >>> > >>>
2014-11-24 18:05:37,223 ERROR
>> >>>>>>> >>> >>> > >>>
>> >>>>>>> >>> >>> > >>>
>> >>>>>>> >>> >>> > >>>
>> >>>>>>> >>> >>> > >>>
>> >>>>>>> >>> >>> > >>>
[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
>> >>>>>>> >>> >>> > >>>
(DefaultQuartzScheduler_Worker-69) [4a84c4e5]
>> >>>>>>> >>> >>> > >>>
Correlation ID:
>> >>>>>>> >>> >>> > >>>
4a84c4e5,
>> >>>>>>> >>> >>> > >>> Job
>> >>>>>>> >>> >>> > >>> ID:
4bfd4a6d-c3ef-468f-a40e-a3a6ca13011b, Call
>> >>>>>>> >>> >>> > >>> Stack:
>> >>>>>>> >>> >>> > >>> null,
>> >>>>>>> >>> >>> > >>> Custom
>> >>>>>>> >>> >>> > >>> Event
>> >>>>>>> >>> >>> > >>> ID: -1,
Message: Gluster command [<UNKNOWN>]
failed
>> >>>>>>> >>> >>> > >>> on
>> >>>>>>> >>> >>> > >>> server
>> >>>>>>> >>> >>> > >>>
Compute5.
>> >>>>>>> >>> >>> > >>>
2014-11-24 18:05:37,243 INFO
>> >>>>>>> >>> >>> > >>>
>> >>>>>>> >>> >>> > >>>
>> >>>>>>> >>> >>> > >>>
>> >>>>>>> >>> >>> > >>>
>> >>>>>>> >>> >>> > >>>
[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
>> >>>>>>> >>> >>> > >>>
(DefaultQuartzScheduler_Worker-69) [4a84c4e5]
>> >>>>>>> >>> >>> > >>>
Correlation ID:
>> >>>>>>> >>> >>> > >>> null,
>> >>>>>>> >>> >>> > >>> Call
>> >>>>>>> >>> >>> > >>> Stack:
null, Custom Event ID: -1, Message: Status
of
>> >>>>>>> >>> >>> > >>> host
>> >>>>>>> >>> >>> > >>>
Compute5
>> >>>>>>> >>> >>> > >>> was
>> >>>>>>> >>> >>> > >>> set
>> >>>>>>> >>> >>> > >>> to
NonOperational.
>> >>>>>>> >>> >>> > >>>
2014-11-24 18:05:37,272 INFO
>> >>>>>>> >>> >>> > >>>
[org.ovirt.engine.core.bll.HandleVdsVersionCommand]
>> >>>>>>> >>> >>> > >>>
(DefaultQuartzScheduler_Worker-69) [a0c8a7f]
Running
>> >>>>>>> >>> >>> > >>>
command:
>> >>>>>>> >>> >>> > >>>
HandleVdsVersionCommand internal: true. Entities
>> >>>>>>> >>> >>> > >>>
affected :
>> >>>>>>> >>> >>> > >>> ID:
>> >>>>>>> >>> >>> > >>>
0bf6b00f-7947-4411-b55a-cc5eea2b381a Type: VDS
>> >>>>>>> >>> >>> > >>>
2014-11-24 18:05:37,274 INFO
>> >>>>>>> >>> >>> > >>>
>> >>>>>>> >>> >>> > >>>
[org.ovirt.engine.core.vdsbroker.VdsUpdateRunTimeInfo]
>> >>>>>>> >>> >>> > >>>
(DefaultQuartzScheduler_Worker-69) [a0c8a7f] Host
>> >>>>>>> >>> >>> > >>>
0bf6b00f-7947-4411-b55a-cc5eea2b381a : Compute5 is
>> >>>>>>> >>> >>> > >>> already
in
>> >>>>>>> >>> >>> > >>>
NonOperational status for reason
>> >>>>>>> >>> >>> > >>>
GLUSTER_COMMAND_FAILED.
>> >>>>>>> >>> >>> > >>>
SetNonOperationalVds command is skipped.
>> >>>>>>> >>> >>> > >>>
2014-11-24 18:05:38,065 INFO
>> >>>>>>> >>> >>> > >>>
>> >>>>>>> >>> >>> > >>>
>> >>>>>>> >>> >>> > >>>
>> >>>>>>> >>> >>> > >>>
>> >>>>>>> >>> >>> > >>>
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand]
>> >>>>>>> >>> >>> > >>>
(DefaultQuartzScheduler_Worker-55) [3706e836]
>> >>>>>>> >>> >>> > >>>
FINISH,
>> >>>>>>> >>> >>> > >>>
GlusterVolumesListVDSCommand, return:
>> >>>>>>> >>> >>> > >>>
>> >>>>>>> >>> >>> > >>>
>> >>>>>>> >>> >>> > >>>
>> >>>>>>> >>> >>> > >>>
>> >>>>>>> >>> >>> > >>>
{26ae1672-ee09-4a38-8fd2-72dd9974cc2b=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@4e72a1b1
},
>> >>>>>>> >>> >>> > >>> log id:
48a0c832
>> >>>>>>> >>> >>> > >>>
2014-11-24 18:05:43,243 INFO
>> >>>>>>> >>> >>> > >>>
>> >>>>>>> >>> >>> > >>>
>> >>>>>>> >>> >>> > >>>
>> >>>>>>> >>> >>> > >>>
>> >>>>>>> >>> >>> > >>>
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand]
>> >>>>>>> >>> >>> > >>>
(DefaultQuartzScheduler_Worker-35) START,
>> >>>>>>> >>> >>> > >>>
GlusterVolumesListVDSCommand(HostName = Compute4,
>> >>>>>>> >>> >>> > >>> HostId
>> >>>>>>> >>> >>> > >>> =
>> >>>>>>> >>> >>> > >>>
33648a90-200c-45ca-89d5-1ce305d79a6a), log id:
>> >>>>>>> >>> >>> > >>>
3ce13ebc
>> >>>>>>> >>> >>> > >>> ^C
>> >>>>>>> >>> >>> > >>>
[root@ccr01 ~]#
>> >>>>>>> >>> >>> > >>>
>> >>>>>>> >>> >>> > >>>
Thanks,
>> >>>>>>> >>> >>> > >>> Punit
>> >>>>>>> >>> >>> > >>>
>> >>>>>>> >>> >>> > >>>
>> >>>>>>> >>> >>> > >>>
_______________________________________________
>> >>>>>>> >>> >>> > >>> Users
mailing
>> >>>>>>> >>> >>> > >>>
>> >>>>>>> >>> >>> > >>>
>> >>>>>>> >>> >>> > >>>
>> >>>>>>> >>> >>> > >>>
listUsers@ovirt.orghttp://
lists.ovirt.org/mailman/listinfo/users
>> >>>>>>> >>> >>> > >>>
>> >>>>>>> >>> >>> > >>>
>> >>>>>>> >>> >>> > >>>
>> >>>>>>> >>> >>> > >>
>> >>>>>>> >>> >>> > >>
>> >>>>>>> >>> >>> > >
>> >>>>>>> >>> >>> > >
>> >>>>>>> >>> >>> >
>> >>>>>>> >>> >>
>> >>>>>>> >>> >>
>> >>>>>>> >>> >
>> >>>>>>> >>> >
>> >>>>>>> >>> >
_______________________________________________
>> >>>>>>> >>> > Gluster-users mailing list
>> >>>>>>> >>> > Gluster-users(a)gluster.org
>> >>>>>>> >>> >
>> >>>>>>> >>> >
http://supercolony.gluster.org/mailman/listinfo/gluster-users
>> >>>>>>> >>
>> >>>>>>> >>
>> >>>>>>> >
>> >>>>>>
>> >>>>>>
>> >>>>
>> >>
>
>
_______________________________________________
Users mailing list
Users(a)ovirt.org
http://lists.ovirt.org/mailman/listinfo/users