Dear Kaushal,
I tried various method...but still the same error...it seems it's gluster
bug..is there any body can suggest work-around here ??
Thanks,
Punit
On Sun, Dec 7, 2014 at 8:40 PM, Punit Dambiwal <hypunit(a)gmail.com> wrote:
Hi Kaushal,
Still the same error...even try your suggested workaround :-
-------------------
Can you replace 'Before=network-online.target' with
'Wants=network-online.target' and try the boot again? This should
force the network to be online before starting GlusterD.
-------------------
Thanks,
Punit
On Sat, Dec 6, 2014 at 11:44 AM, Punit Dambiwal <hypunit(a)gmail.com> wrote:
> Hi Kaushal,
>
> I already have all the hosts entry in the /etc/hosts for the easy
> resolution....i will try your method in the glusterd.services and check and
> let you know....weather problem solve or not.....
>
> On Fri, Dec 5, 2014 at 9:50 PM, Kaushal M <kshlmster(a)gmail.com> wrote:
>
>> Can you replace 'Before=network-online.target' with
>> 'Wants=network-online.target' and try the boot again? This should
>> force the network to be online before starting GlusterD.
>>
>> If even that fails, you could try adding an entry into /etc/hosts with
>> the hostname of the system. This should prevent any more failures.
>>
>> I still don't believe it's a problem with Gluster. Gluster uses apis
>> provided by the system to perform name resolution. These definitely
>> work correctly because you can start GlusterD later. Since the
>> resolution failure only happens during boot, it points to system or
>> network setup issues during boot. To me it seems like the network
>> isn't completely setup at that point of time.
>>
>> ~kaushal
>>
>> On Fri, Dec 5, 2014 at 12:47 PM, Punit Dambiwal <hypunit(a)gmail.com>
>> wrote:
>> > Hi Kaushal,
>> >
>> > It seems it's bug in glusterfs 3.6....even i manage my systemd to
>> start the
>> > network service before glusterd...but it's still fail...
>> >
>> > ---------------
>> > [Unit]
>> > Description=GlusterFS, a clustered file-system server
>> > After=network.target rpcbind.service
>> > Before=network-online.target
>> >
>> > [Service]
>> > Type=forking
>> > PIDFile=/var/run/glusterd.pid
>> > LimitNOFILE=65536
>> > ExecStartPre=/etc/rc.d/init.d/network start
>> > ExecStart=/usr/sbin/glusterd -p /var/run/glusterd.pid
>> > KillMode=process
>> >
>> > [Install]
>> > WantedBy=multi-user.target
>> > ----------------
>> >
>> > Thanks,
>> > Punit
>> >
>> > On Wed, Dec 3, 2014 at 8:56 PM, Kaushal M <kshlmster(a)gmail.com>
wrote:
>> >>
>> >> I just remembered this.
>> >>
>> >> There was another user having a similar issue of GlusterD failing to
>> >> start on the mailing list a while back. The cause of his problem was
>> >> the way his network was brought up.
>> >> IIRC, he was using a static network configuration. The problem
>> >> vanished when he began using dhcp. Or it might have been he was using
>> >> dhcp.service and it got solved after switching to NetworkManager.
>> >>
>> >> This could be one more thing you could look at.
>> >>
>> >> I'll try to find the mail thread to see if it was the same problem
as
>> you.
>> >>
>> >> ~kaushal
>> >>
>> >> On Wed, Dec 3, 2014 at 6:22 PM, Kaushal M <kshlmster(a)gmail.com>
>> wrote:
>> >> > I don't know much about how the network target is brought up
in
>> >> > CentOS7, but I'll try as much as I can.
>> >> >
>> >> > It seems to me that, after the network has been brought up and by
>> the
>> >> > time GlusterD is started,
>> >> > a. The machine hasn't yet recieved it's hostname, or
>> >> > b. It hasn't yet registered with the name server.
>> >> >
>> >> > This is causing name resolution failures.
>> >> >
>> >> > I don't know if the network target could come up without the
machine
>> >> > getting its hostname, so I'm pretty sure it's not a.
>> >> >
>> >> > So it seems to be b. But these kind of signing in happens only in
>> DDNS
>> >> > systems, which doesn't seem to be the case for you.
>> >> >
>> >> > Both of these reasons might be wrong (most likely wrong). You'd
do
>> >> > good if you could ask for help from someone with more experience
in
>> >> > systemd + networking.
>> >> >
>> >> > ~kaushal
>> >> >
>> >> > On Wed, Dec 3, 2014 at 10:54 AM, Punit Dambiwal
<hypunit(a)gmail.com>
>> >> > wrote:
>> >> >> Hi Kaushal,
>> >> >>
>> >> >> This is the host...which i rebooted...would you mind to let me
>> know how
>> >> >> i
>> >> >> can make the glusterd sevice come up after network...i am
using
>> >> >> centos7...if
>> >> >> network is the issue...
>> >> >>
>> >> >> On Wed, Dec 3, 2014 at 11:54 AM, Kaushal M
<kshlmster(a)gmail.com>
>> wrote:
>> >> >>>
>> >> >>> This peer cannot be identified.
>> >> >>>
>> >> >>> " [2014-12-03 02:29:25.998153] D
>> >> >>>
[glusterd-peer-utils.c:121:glusterd_peerinfo_find_by_hostname]
>> >> >>> 0-management:
>> >> >>> Unable to find friend:
cpu05.zne01.hkg1.ovt.36stack.com"
>> >> >>>
>> >> >>> I don't know why this address is not being resolved
during boot
>> time.
>> >> >>> If
>> >> >>> this is a valid peer, the the only reason I can think of
this
>> that the
>> >> >>> network is not up.
>> >> >>>
>> >> >>> If you had previously detached the peer forcefully, the
that could
>> >> >>> have
>> >> >>> left stale entries in some volumes. In this case as well,
GlusterD
>> >> >>> will fail
>> >> >>> to identify the peer.
>> >> >>>
>> >> >>> Do either of these reasons seem a possibility to you?
>> >> >>>
>> >> >>> On Dec 3, 2014 8:07 AM, "Punit Dambiwal"
<hypunit(a)gmail.com>
>> wrote:
>> >> >>>>
>> >> >>>> Hi Kaushal,
>> >> >>>>
>> >> >>>> Please find the logs here :-
http://ur1.ca/iyoe5 and
>> >> >>>>
http://ur1.ca/iyoed
>> >> >>>>
>> >> >>>> On Tue, Dec 2, 2014 at 10:43 PM, Kaushal M
<kshlmster(a)gmail.com>
>> >> >>>> wrote:
>> >> >>>>>
>> >> >>>>> Hey Punit,
>> >> >>>>> In the logs you've provided, GlusterD appears
to be running
>> >> >>>>> correctly.
>> >> >>>>> Could you provide the logs for the time period when
GlusterD
>> >> >>>>> attempts to
>> >> >>>>> start but fails.
>> >> >>>>>
>> >> >>>>> ~kaushal
>> >> >>>>>
>> >> >>>>> On Dec 2, 2014 8:03 PM, "Punit Dambiwal"
<hypunit(a)gmail.com>
>> wrote:
>> >> >>>>>>
>> >> >>>>>> Hi Kaushal,
>> >> >>>>>>
>> >> >>>>>> Please find the logs here :-
http://ur1.ca/iyhs5 and
>> >> >>>>>>
http://ur1.ca/iyhue
>> >> >>>>>>
>> >> >>>>>> Thanks,
>> >> >>>>>> punit
>> >> >>>>>>
>> >> >>>>>>
>> >> >>>>>> On Tue, Dec 2, 2014 at 12:00 PM, Kaushal M
<
>> kshlmster(a)gmail.com>
>> >> >>>>>> wrote:
>> >> >>>>>>>
>> >> >>>>>>> Hey Punit,
>> >> >>>>>>> Could you start Glusterd in debug mode and
provide the logs
>> here?
>> >> >>>>>>> To start it in debug mode, append
'-LDEBUG' to the ExecStart
>> line
>> >> >>>>>>> in
>> >> >>>>>>> the service file.
>> >> >>>>>>>
>> >> >>>>>>> ~kaushal
>> >> >>>>>>>
>> >> >>>>>>> On Mon, Dec 1, 2014 at 9:05 AM, Punit
Dambiwal <
>> hypunit(a)gmail.com>
>> >> >>>>>>> wrote:
>> >> >>>>>>> > Hi,
>> >> >>>>>>> >
>> >> >>>>>>> > Can Any body help me on this ??
>> >> >>>>>>> >
>> >> >>>>>>> > On Thu, Nov 27, 2014 at 9:29 AM, Punit
Dambiwal
>> >> >>>>>>> > <hypunit(a)gmail.com>
>> >> >>>>>>> > wrote:
>> >> >>>>>>> >>
>> >> >>>>>>> >> Hi Kaushal,
>> >> >>>>>>> >>
>> >> >>>>>>> >> Thanks for the detailed
reply....let me explain my setup
>> first
>> >> >>>>>>> >> :-
>> >> >>>>>>> >>
>> >> >>>>>>> >> 1. Ovirt Engine
>> >> >>>>>>> >> 2. 4* host as well as storage
machine (Host and gluster
>> >> >>>>>>> >> combined)
>> >> >>>>>>> >> 3. Every host has 24 bricks...
>> >> >>>>>>> >>
>> >> >>>>>>> >> Now whenever the host machine
reboot...it can come up but
>> can
>> >> >>>>>>> >> not
>> >> >>>>>>> >> join the
>> >> >>>>>>> >> cluster again and through the
following error "Gluster
>> command
>> >> >>>>>>> >> [<UNKNOWN>]
>> >> >>>>>>> >> failed on server.."
>> >> >>>>>>> >>
>> >> >>>>>>> >> Please check my comment in line
:-
>> >> >>>>>>> >>
>> >> >>>>>>> >> 1. Use the same string for doing
the peer probe and for the
>> >> >>>>>>> >> brick
>> >> >>>>>>> >> address
>> >> >>>>>>> >> during volume create/add-brick.
Ideally, we suggest you use
>> >> >>>>>>> >> properly
>> >> >>>>>>> >> resolvable FQDNs everywhere. If
that is not possible, then
>> use
>> >> >>>>>>> >> only
>> >> >>>>>>> >> IP
>> >> >>>>>>> >> addresses. Try to avoid short
names.
>> >> >>>>>>> >> ---------------
>> >> >>>>>>> >> [root@cpu05 ~]# gluster peer
status
>> >> >>>>>>> >> Number of Peers: 3
>> >> >>>>>>> >>
>> >> >>>>>>> >> Hostname:
cpu03.stack.com
>> >> >>>>>>> >> Uuid:
5729b8c4-e80d-4353-b456-6f467bddbdfb
>> >> >>>>>>> >> State: Peer in Cluster
(Connected)
>> >> >>>>>>> >>
>> >> >>>>>>> >> Hostname:
cpu04.stack.com
>> >> >>>>>>> >> Uuid:
d272b790-c4b2-4bed-ba68-793656e6d7b0
>> >> >>>>>>> >> State: Peer in Cluster
(Connected)
>> >> >>>>>>> >> Other names:
>> >> >>>>>>> >> 10.10.0.8
>> >> >>>>>>> >>
>> >> >>>>>>> >> Hostname:
cpu02.stack.com
>> >> >>>>>>> >> Uuid:
8d8a7041-950e-40d0-85f9-58d14340ca25
>> >> >>>>>>> >> State: Peer in Cluster
(Connected)
>> >> >>>>>>> >> [root@cpu05 ~]#
>> >> >>>>>>> >> ----------------
>> >> >>>>>>> >> 2. During boot up, make sure to
launch glusterd only after
>> the
>> >> >>>>>>> >> network is
>> >> >>>>>>> >> up. This will allow the new peer
identification mechanism
>> to do
>> >> >>>>>>> >> its
>> >> >>>>>>> >> job correctly.
>> >> >>>>>>> >> >> I think the service
itself doing the same job....
>> >> >>>>>>> >>
>> >> >>>>>>> >> [root@cpu05 ~]# cat
>> /usr/lib/systemd/system/glusterd.service
>> >> >>>>>>> >> [Unit]
>> >> >>>>>>> >> Description=GlusterFS, a clustered
file-system server
>> >> >>>>>>> >> After=network.target
rpcbind.service
>> >> >>>>>>> >> Before=network-online.target
>> >> >>>>>>> >>
>> >> >>>>>>> >> [Service]
>> >> >>>>>>> >> Type=forking
>> >> >>>>>>> >> PIDFile=/var/run/glusterd.pid
>> >> >>>>>>> >> LimitNOFILE=65536
>> >> >>>>>>> >> ExecStart=/usr/sbin/glusterd -p
/var/run/glusterd.pid
>> >> >>>>>>> >> KillMode=process
>> >> >>>>>>> >>
>> >> >>>>>>> >> [Install]
>> >> >>>>>>> >> WantedBy=multi-user.target
>> >> >>>>>>> >> [root@cpu05 ~]#
>> >> >>>>>>> >> --------------------
>> >> >>>>>>> >>
>> >> >>>>>>> >> gluster logs :-
>> >> >>>>>>> >>
>> >> >>>>>>> >> [2014-11-24 09:22:22.147471] I
[MSGID: 100030]
>> >> >>>>>>> >> [glusterfsd.c:2018:main]
>> >> >>>>>>> >> 0-/usr/sbin/glusterd: Started
running /usr/sbin/glusterd
>> >> >>>>>>> >> version
>> >> >>>>>>> >> 3.6.1
>> >> >>>>>>> >> (args: /usr/sbin/glusterd -p
/var/run/glusterd.pid)
>> >> >>>>>>> >> [2014-11-24 09:22:22.151565] I
[glusterd.c:1214:init]
>> >> >>>>>>> >> 0-management:
>> >> >>>>>>> >> Maximum allowed open file
descriptors set to 65536
>> >> >>>>>>> >> [2014-11-24 09:22:22.151599] I
[glusterd.c:1259:init]
>> >> >>>>>>> >> 0-management:
>> >> >>>>>>> >> Using
>> >> >>>>>>> >> /var/lib/glusterd as working
directory
>> >> >>>>>>> >> [2014-11-24 09:22:22.155216] W
>> >> >>>>>>> >>
[rdma.c:4195:__gf_rdma_ctx_create]
>> >> >>>>>>> >> 0-rpc-transport/rdma: rdma_cm
event channel creation
>> failed (No
>> >> >>>>>>> >> such device)
>> >> >>>>>>> >> [2014-11-24 09:22:22.155264] E
[rdma.c:4483:init]
>> >> >>>>>>> >> 0-rdma.management:
>> >> >>>>>>> >> Failed to initialize IB Device
>> >> >>>>>>> >> [2014-11-24 09:22:22.155285] E
>> >> >>>>>>> >>
[rpc-transport.c:333:rpc_transport_load]
>> >> >>>>>>> >> 0-rpc-transport: 'rdma'
initialization failed
>> >> >>>>>>> >> [2014-11-24 09:22:22.155354] W
>> >> >>>>>>> >>
[rpcsvc.c:1524:rpcsvc_transport_create]
>> >> >>>>>>> >> 0-rpc-service: cannot create
listener, initing the
>> transport
>> >> >>>>>>> >> failed
>> >> >>>>>>> >> [2014-11-24 09:22:22.156290] I
>> >> >>>>>>> >>
[glusterd.c:413:glusterd_check_gsync_present] 0-glusterd:
>> >> >>>>>>> >> geo-replication
>> >> >>>>>>> >> module not installed in the
system
>> >> >>>>>>> >> [2014-11-24 09:22:22.161318] I
>> >> >>>>>>> >>
[glusterd-store.c:2043:glusterd_restore_op_version]
>> 0-glusterd:
>> >> >>>>>>> >> retrieved
>> >> >>>>>>> >> op-version: 30600
>> >> >>>>>>> >> [2014-11-24 09:22:22.821800] I
>> >> >>>>>>> >>
[glusterd-handler.c:3146:glusterd_friend_add_from_peerinfo]
>> >> >>>>>>> >> 0-management:
>> >> >>>>>>> >> connect returned 0
>> >> >>>>>>> >> [2014-11-24 09:22:22.825810] I
>> >> >>>>>>> >>
[glusterd-handler.c:3146:glusterd_friend_add_from_peerinfo]
>> >> >>>>>>> >> 0-management:
>> >> >>>>>>> >> connect returned 0
>> >> >>>>>>> >> [2014-11-24 09:22:22.828705] I
>> >> >>>>>>> >>
[glusterd-handler.c:3146:glusterd_friend_add_from_peerinfo]
>> >> >>>>>>> >> 0-management:
>> >> >>>>>>> >> connect returned 0
>> >> >>>>>>> >> [2014-11-24 09:22:22.828771] I
>> >> >>>>>>> >>
[rpc-clnt.c:969:rpc_clnt_connection_init]
>> >> >>>>>>> >> 0-management: setting
frame-timeout to 600
>> >> >>>>>>> >> [2014-11-24 09:22:22.832670] I
>> >> >>>>>>> >>
[rpc-clnt.c:969:rpc_clnt_connection_init]
>> >> >>>>>>> >> 0-management: setting
frame-timeout to 600
>> >> >>>>>>> >> [2014-11-24 09:22:22.835919] I
>> >> >>>>>>> >>
[rpc-clnt.c:969:rpc_clnt_connection_init]
>> >> >>>>>>> >> 0-management: setting
frame-timeout to 600
>> >> >>>>>>> >> [2014-11-24 09:22:22.840209] E
>> >> >>>>>>> >>
[glusterd-store.c:4248:glusterd_resolve_all_bricks]
>> 0-glusterd:
>> >> >>>>>>> >> resolve
>> >> >>>>>>> >> brick failed in restore
>> >> >>>>>>> >> [2014-11-24 09:22:22.840233] E
[xlator.c:425:xlator_init]
>> >> >>>>>>> >> 0-management:
>> >> >>>>>>> >> Initialization of volume
'management' failed, review your
>> >> >>>>>>> >> volfile
>> >> >>>>>>> >> again
>> >> >>>>>>> >> [2014-11-24 09:22:22.840245] E
>> >> >>>>>>> >>
[graph.c:322:glusterfs_graph_init]
>> >> >>>>>>> >> 0-management: initializing
translator failed
>> >> >>>>>>> >> [2014-11-24 09:22:22.840264] E
>> >> >>>>>>> >>
[graph.c:525:glusterfs_graph_activate]
>> >> >>>>>>> >> 0-graph: init failed
>> >> >>>>>>> >> [2014-11-24 09:22:22.840754] W
>> >> >>>>>>> >>
[glusterfsd.c:1194:cleanup_and_exit]
>> >> >>>>>>> >> (-->
>> >> >>>>>>> >> 0-: received signum (0), shutting
down
>> >> >>>>>>> >>
>> >> >>>>>>> >> Thanks,
>> >> >>>>>>> >> Punit
>> >> >>>>>>> >>
>> >> >>>>>>> >>
>> >> >>>>>>> >>
>> >> >>>>>>> >>
>> >> >>>>>>> >> On Wed, Nov 26, 2014 at 7:14 PM,
Kaushal M
>> >> >>>>>>> >> <kshlmster(a)gmail.com>
>> >> >>>>>>> >> wrote:
>> >> >>>>>>> >>>
>> >> >>>>>>> >>> Based on the logs I can guess
that glusterd is being
>> started
>> >> >>>>>>> >>> before
>> >> >>>>>>> >>> the network has come up and
that the addresses given to
>> bricks
>> >> >>>>>>> >>> do
>> >> >>>>>>> >>> not
>> >> >>>>>>> >>> directly match the addresses
used in during peer probe.
>> >> >>>>>>> >>>
>> >> >>>>>>> >>> The gluster_after_reboot log
has the line "[2014-11-25
>> >> >>>>>>> >>> 06:46:09.972113] E
>> >> >>>>>>> >>>
[glusterd-store.c:2632:glusterd_resolve_all_bricks]
>> >> >>>>>>> >>> 0-glusterd: resolve brick
failed in restore".
>> >> >>>>>>> >>>
>> >> >>>>>>> >>> Brick resolution fails when
glusterd cannot match the
>> address
>> >> >>>>>>> >>> for
>> >> >>>>>>> >>> the
>> >> >>>>>>> >>> brick, with one of the peers.
Brick resolution happens in
>> two
>> >> >>>>>>> >>> phases,
>> >> >>>>>>> >>> 1. We first try to identify
the peer by performing string
>> >> >>>>>>> >>> comparisions
>> >> >>>>>>> >>> with the brick address and the
peer addresses (The peer
>> names
>> >> >>>>>>> >>> will
>> >> >>>>>>> >>> be
>> >> >>>>>>> >>> the names/addresses that were
given when the peer was
>> probed).
>> >> >>>>>>> >>> 2. If we don't find a
match from step 1, we will then
>> resolve
>> >> >>>>>>> >>> all
>> >> >>>>>>> >>> the
>> >> >>>>>>> >>> brick address and the peer
addresses into addrinfo
>> structs,
>> >> >>>>>>> >>> and
>> >> >>>>>>> >>> then
>> >> >>>>>>> >>> compare these structs to find
a match. This process should
>> >> >>>>>>> >>> generally
>> >> >>>>>>> >>> find a match if available.
This will fail only if the
>> network
>> >> >>>>>>> >>> is
>> >> >>>>>>> >>> not
>> >> >>>>>>> >>> up yet as we cannot resolve
addresses.
>> >> >>>>>>> >>>
>> >> >>>>>>> >>> The above steps are applicable
only to glusterfs versions
>> >> >>>>>>> >>> >=3.6.
>> >> >>>>>>> >>> They
>> >> >>>>>>> >>> were introduced to reduce
problems with peer
>> identification,
>> >> >>>>>>> >>> like
>> >> >>>>>>> >>> the
>> >> >>>>>>> >>> one you encountered
>> >> >>>>>>> >>>
>> >> >>>>>>> >>> Since both of the steps failed
to find a match in one
>> run, but
>> >> >>>>>>> >>> succeeded later, we can come
to the conclusion that,
>> >> >>>>>>> >>> a) the bricks don't have
the exact same string used in
>> peer
>> >> >>>>>>> >>> probe
>> >> >>>>>>> >>> for
>> >> >>>>>>> >>> their addresses as step 1
failed, and
>> >> >>>>>>> >>> b) the network was not up in
the initial run, as step 2
>> failed
>> >> >>>>>>> >>> during
>> >> >>>>>>> >>> the initial run, but passed in
the second run.
>> >> >>>>>>> >>>
>> >> >>>>>>> >>> Please let me know if my
conclusion is correct.
>> >> >>>>>>> >>>
>> >> >>>>>>> >>> If it is, you can solve your
problem in two ways.
>> >> >>>>>>> >>> 1. Use the same string for
doing the peer probe and for
>> the
>> >> >>>>>>> >>> brick
>> >> >>>>>>> >>> address during volume
create/add-brick. Ideally, we
>> suggest
>> >> >>>>>>> >>> you
>> >> >>>>>>> >>> use
>> >> >>>>>>> >>> properly resolvable FQDNs
everywhere. If that is not
>> possible,
>> >> >>>>>>> >>> then
>> >> >>>>>>> >>> use only IP addresses. Try to
avoid short names.
>> >> >>>>>>> >>> 2. During boot up, make sure
to launch glusterd only
>> after the
>> >> >>>>>>> >>> network
>> >> >>>>>>> >>> is up. This will allow the new
peer identification
>> mechanism
>> >> >>>>>>> >>> to do
>> >> >>>>>>> >>> its
>> >> >>>>>>> >>> job correctly.
>> >> >>>>>>> >>>
>> >> >>>>>>> >>>
>> >> >>>>>>> >>> If you have already followed
these steps and yet still
>> hit the
>> >> >>>>>>> >>> problem, then please provide
more information (setup,
>> logs,
>> >> >>>>>>> >>> etc.).
>> >> >>>>>>> >>> It
>> >> >>>>>>> >>> could be much different
problem that you are facing.
>> >> >>>>>>> >>>
>> >> >>>>>>> >>> ~kaushal
>> >> >>>>>>> >>>
>> >> >>>>>>> >>> On Wed, Nov 26, 2014 at 4:01
PM, Punit Dambiwal
>> >> >>>>>>> >>> <hypunit(a)gmail.com>
>> >> >>>>>>> >>> wrote:
>> >> >>>>>>> >>> > Is there any one can help
on this ??
>> >> >>>>>>> >>> >
>> >> >>>>>>> >>> > Thanks,
>> >> >>>>>>> >>> > punit
>> >> >>>>>>> >>> >
>> >> >>>>>>> >>> > On Wed, Nov 26, 2014 at
9:42 AM, Punit Dambiwal
>> >> >>>>>>> >>> >
<hypunit(a)gmail.com>
>> >> >>>>>>> >>> > wrote:
>> >> >>>>>>> >>> >>
>> >> >>>>>>> >>> >> Hi,
>> >> >>>>>>> >>> >>
>> >> >>>>>>> >>> >> My Glusterfs version
is :- glusterfs-3.6.1-1.el7
>> >> >>>>>>> >>> >>
>> >> >>>>>>> >>> >> On Wed, Nov 26, 2014
at 1:59 AM, Kanagaraj Mayilsamy
>> >> >>>>>>> >>> >>
<kmayilsa(a)redhat.com>
>> >> >>>>>>> >>> >> wrote:
>> >> >>>>>>> >>> >>>
>> >> >>>>>>> >>> >>>
[+Gluster-users(a)gluster.org]
>> >> >>>>>>> >>> >>>
>> >> >>>>>>> >>> >>>
"Initialization of volume 'management' failed, review
>> your
>> >> >>>>>>> >>> >>> volfile
>> >> >>>>>>> >>> >>> again",
glusterd throws this error when the service is
>> >> >>>>>>> >>> >>> started
>> >> >>>>>>> >>> >>> automatically
>> >> >>>>>>> >>> >>> after the reboot.
But the service is successfully
>> started
>> >> >>>>>>> >>> >>> later
>> >> >>>>>>> >>> >>> manually by
>> >> >>>>>>> >>> >>> the user.
>> >> >>>>>>> >>> >>>
>> >> >>>>>>> >>> >>> can somebody from
gluster-users please help on this?
>> >> >>>>>>> >>> >>>
>> >> >>>>>>> >>> >>> glusterfs
version: 3.5.1
>> >> >>>>>>> >>> >>>
>> >> >>>>>>> >>> >>> Thanks,
>> >> >>>>>>> >>> >>> Kanagaraj
>> >> >>>>>>> >>> >>>
>> >> >>>>>>> >>> >>> ----- Original
Message -----
>> >> >>>>>>> >>> >>> > From:
"Punit Dambiwal" <hypunit(a)gmail.com>
>> >> >>>>>>> >>> >>> > To:
"Kanagaraj" <kmayilsa(a)redhat.com>
>> >> >>>>>>> >>> >>> > Cc:
users(a)ovirt.org
>> >> >>>>>>> >>> >>> > Sent:
Tuesday, November 25, 2014 7:24:45 PM
>> >> >>>>>>> >>> >>> > Subject: Re:
[ovirt-users] Gluster command
>> [<UNKNOWN>]
>> >> >>>>>>> >>> >>> > failed on
>> >> >>>>>>> >>> >>> > server...
>> >> >>>>>>> >>> >>> >
>> >> >>>>>>> >>> >>> > Hi
Kanagraj,
>> >> >>>>>>> >>> >>> >
>> >> >>>>>>> >>> >>> > Please check
the attached log files....i didn't
>> find any
>> >> >>>>>>> >>> >>> > thing
>> >> >>>>>>> >>> >>> > special....
>> >> >>>>>>> >>> >>> >
>> >> >>>>>>> >>> >>> > On Tue, Nov
25, 2014 at 12:12 PM, Kanagaraj
>> >> >>>>>>> >>> >>> >
<kmayilsa(a)redhat.com>
>> >> >>>>>>> >>> >>> > wrote:
>> >> >>>>>>> >>> >>> >
>> >> >>>>>>> >>> >>> > > Do you
see any errors in
>> >> >>>>>>> >>> >>> > >
/var/log/glusterfs/etc-glusterfs-glusterd.vol.log
>> or
>> >> >>>>>>> >>> >>> > >
vdsm.log
>> >> >>>>>>> >>> >>> > > when
>> >> >>>>>>> >>> >>> > > the
>> >> >>>>>>> >>> >>> > > service
is trying to start automatically after the
>> >> >>>>>>> >>> >>> > >
reboot?
>> >> >>>>>>> >>> >>> > >
>> >> >>>>>>> >>> >>> > >
Thanks,
>> >> >>>>>>> >>> >>> > >
Kanagaraj
>> >> >>>>>>> >>> >>> > >
>> >> >>>>>>> >>> >>> > >
>> >> >>>>>>> >>> >>> > > On
11/24/2014 08:13 PM, Punit Dambiwal wrote:
>> >> >>>>>>> >>> >>> > >
>> >> >>>>>>> >>> >>> > > Hi
Kanagaraj,
>> >> >>>>>>> >>> >>> > >
>> >> >>>>>>> >>> >>> > >
Yes...once i will start the gluster service and
>> then
>> >> >>>>>>> >>> >>> > > vdsmd
>> >> >>>>>>> >>> >>> > > ...the
>> >> >>>>>>> >>> >>> > > host
>> >> >>>>>>> >>> >>> > > can
connect to cluster...but the question is why
>> it's
>> >> >>>>>>> >>> >>> > > not
>> >> >>>>>>> >>> >>> > >
started
>> >> >>>>>>> >>> >>> > > even
it
>> >> >>>>>>> >>> >>> > > has
chkconfig enabled...
>> >> >>>>>>> >>> >>> > >
>> >> >>>>>>> >>> >>> > > I have
tested it in two host cluster
>> >> >>>>>>> >>> >>> > >
environment...(Centos 6.6
>> >> >>>>>>> >>> >>> > > and
>> >> >>>>>>> >>> >>> > > centos
7.0) on both hypervisior cluster..it's
>> failed
>> >> >>>>>>> >>> >>> > > to
>> >> >>>>>>> >>> >>> > >
reconnect
>> >> >>>>>>> >>> >>> > > in
>> >> >>>>>>> >>> >>> > > to
>> >> >>>>>>> >>> >>> > > cluster
after reboot....
>> >> >>>>>>> >>> >>> > >
>> >> >>>>>>> >>> >>> > > In
both the environment glusterd enabled for next
>> >> >>>>>>> >>> >>> > >
boot....but
>> >> >>>>>>> >>> >>> > >
it's
>> >> >>>>>>> >>> >>> > > failed
with the same error....seems it's bug in
>> either
>> >> >>>>>>> >>> >>> > > gluster
or
>> >> >>>>>>> >>> >>> > > Ovirt
??
>> >> >>>>>>> >>> >>> > >
>> >> >>>>>>> >>> >>> > > Please
help me to find the workaround here if
>> can not
>> >> >>>>>>> >>> >>> > >
resolve
>> >> >>>>>>> >>> >>> > >
it...as
>> >> >>>>>>> >>> >>> > > without
this the Host machine can not connect
>> after
>> >> >>>>>>> >>> >>> > >
reboot....that
>> >> >>>>>>> >>> >>> > > means
>> >> >>>>>>> >>> >>> > > engine
will consider it as down and every time
>> need to
>> >> >>>>>>> >>> >>> > >
manually
>> >> >>>>>>> >>> >>> > > start
>> >> >>>>>>> >>> >>> > > the
>> >> >>>>>>> >>> >>> > > gluster
service and vdsmd... ??
>> >> >>>>>>> >>> >>> > >
>> >> >>>>>>> >>> >>> > >
Thanks,
>> >> >>>>>>> >>> >>> > > Punit
>> >> >>>>>>> >>> >>> > >
>> >> >>>>>>> >>> >>> > > On Mon,
Nov 24, 2014 at 10:20 PM, Kanagaraj
>> >> >>>>>>> >>> >>> > >
<kmayilsa(a)redhat.com>
>> >> >>>>>>> >>> >>> > > wrote:
>> >> >>>>>>> >>> >>> > >
>> >> >>>>>>> >>> >>> > >>
From vdsm.log "error: Connection failed. Please
>> >> >>>>>>> >>> >>> > >>
check if
>> >> >>>>>>> >>> >>> > >>
gluster
>> >> >>>>>>> >>> >>> > >>
daemon
>> >> >>>>>>> >>> >>> > >> is
operational."
>> >> >>>>>>> >>> >>> > >>
>> >> >>>>>>> >>> >>> > >>
Starting glusterd service should fix this issue.
>> >> >>>>>>> >>> >>> > >>
'service
>> >> >>>>>>> >>> >>> > >>
glusterd
>> >> >>>>>>> >>> >>> > >>
start'
>> >> >>>>>>> >>> >>> > >> But
i am wondering why the glusterd was not
>> started
>> >> >>>>>>> >>> >>> > >>
automatically
>> >> >>>>>>> >>> >>> > >>
after
>> >> >>>>>>> >>> >>> > >> the
reboot.
>> >> >>>>>>> >>> >>> > >>
>> >> >>>>>>> >>> >>> > >>
Thanks,
>> >> >>>>>>> >>> >>> > >>
Kanagaraj
>> >> >>>>>>> >>> >>> > >>
>> >> >>>>>>> >>> >>> > >>
>> >> >>>>>>> >>> >>> > >>
>> >> >>>>>>> >>> >>> > >> On
11/24/2014 07:18 PM, Punit Dambiwal wrote:
>> >> >>>>>>> >>> >>> > >>
>> >> >>>>>>> >>> >>> > >> Hi
Kanagaraj,
>> >> >>>>>>> >>> >>> > >>
>> >> >>>>>>> >>> >>> > >>
Please find the attached VDSM logs :-
>> >> >>>>>>> >>> >>> > >>
>> >> >>>>>>> >>> >>> > >>
----------------
>> >> >>>>>>> >>> >>> > >>
Thread-13::DEBUG::2014-11-24
>> >> >>>>>>> >>> >>> > >>
>> >> >>>>>>> >>> >>> > >>
>> >> >>>>>>> >>> >>> > >>
>> >> >>>>>>> >>> >>> > >>
>> >> >>>>>>> >>> >>> > >>
>> 21:41:17,182::resourceManager::977::Storage.ResourceManager.Owner::(cancelAll)
>> >> >>>>>>> >>> >>> > >>
Owner.cancelAll requests {}
>> >> >>>>>>> >>> >>> > >>
Thread-13::DEBUG::2014-11-24
>> >> >>>>>>> >>> >>> > >>
>> >> >>>>>>> >>> >>> > >>
>> >> >>>>>>> >>> >>> > >>
>> 21:41:17,182::task::993::Storage.TaskManager.Task::(_decref)
>> >> >>>>>>> >>> >>> > >>
Task=`1691d409-9b27-4585-8281-5ec26154367a`::ref
>> 0
>> >> >>>>>>> >>> >>> > >>
aborting
>> >> >>>>>>> >>> >>> > >>
False
>> >> >>>>>>> >>> >>> > >>
Thread-13::DEBUG::2014-11-24
>> >> >>>>>>> >>> >>> > >>
>> >> >>>>>>> >>> >>> > >>
>> >> >>>>>>> >>> >>> > >>
>> >> >>>>>>> >>> >>> > >>
>> 21:41:32,393::task::595::Storage.TaskManager.Task::(_updateState)
>> >> >>>>>>> >>> >>> > >>
>> Task=`994c7bc3-a236-4d03-a732-e068c7ed9ed4`::moving
>> >> >>>>>>> >>> >>> > >>
from
>> >> >>>>>>> >>> >>> > >>
state
>> >> >>>>>>> >>> >>> > >>
init
>> >> >>>>>>> >>> >>> > >>
->
>> >> >>>>>>> >>> >>> > >>
state preparing
>> >> >>>>>>> >>> >>> > >>
Thread-13::INFO::2014-11-24
>> >> >>>>>>> >>> >>> > >>
>> 21:41:32,393::logUtils::44::dispatcher::(wrapper) Run
>> >> >>>>>>> >>> >>> > >>
and
>> >> >>>>>>> >>> >>> > >>
protect:
>> >> >>>>>>> >>> >>> > >>
repoStats(options=None)
>> >> >>>>>>> >>> >>> > >>
Thread-13::INFO::2014-11-24
>> >> >>>>>>> >>> >>> > >>
>> 21:41:32,393::logUtils::47::dispatcher::(wrapper) Run
>> >> >>>>>>> >>> >>> > >>
and
>> >> >>>>>>> >>> >>> > >>
protect:
>> >> >>>>>>> >>> >>> > >>
repoStats, Return response: {}
>> >> >>>>>>> >>> >>> > >>
Thread-13::DEBUG::2014-11-24
>> >> >>>>>>> >>> >>> > >>
>> >> >>>>>>> >>> >>> > >>
>> >> >>>>>>> >>> >>> > >>
>> 21:41:32,393::task::1191::Storage.TaskManager.Task::(prepare)
>> >> >>>>>>> >>> >>> > >>
>> >> >>>>>>> >>> >>> > >>
>> Task=`994c7bc3-a236-4d03-a732-e068c7ed9ed4`::finished: {}
>> >> >>>>>>> >>> >>> > >>
Thread-13::DEBUG::2014-11-24
>> >> >>>>>>> >>> >>> > >>
>> >> >>>>>>> >>> >>> > >>
>> >> >>>>>>> >>> >>> > >>
>> >> >>>>>>> >>> >>> > >>
>> 21:41:32,394::task::595::Storage.TaskManager.Task::(_updateState)
>> >> >>>>>>> >>> >>> > >>
>> Task=`994c7bc3-a236-4d03-a732-e068c7ed9ed4`::moving
>> >> >>>>>>> >>> >>> > >>
from
>> >> >>>>>>> >>> >>> > >>
state
>> >> >>>>>>> >>> >>> > >>
preparing
>> >> >>>>>>> >>> >>> > >>
->
>> >> >>>>>>> >>> >>> > >>
state finished
>> >> >>>>>>> >>> >>> > >>
Thread-13::DEBUG::2014-11-24
>> >> >>>>>>> >>> >>> > >>
>> >> >>>>>>> >>> >>> > >>
>> >> >>>>>>> >>> >>> > >>
>> >> >>>>>>> >>> >>> > >>
>> >> >>>>>>> >>> >>> > >>
>> 21:41:32,394::resourceManager::940::Storage.ResourceManager.Owner::(releaseAll)
>> >> >>>>>>> >>> >>> > >>
Owner.releaseAll requests {} resources {}
>> >> >>>>>>> >>> >>> > >>
Thread-13::DEBUG::2014-11-24
>> >> >>>>>>> >>> >>> > >>
>> >> >>>>>>> >>> >>> > >>
>> >> >>>>>>> >>> >>> > >>
>> >> >>>>>>> >>> >>> > >>
>> >> >>>>>>> >>> >>> > >>
>> 21:41:32,394::resourceManager::977::Storage.ResourceManager.Owner::(cancelAll)
>> >> >>>>>>> >>> >>> > >>
Owner.cancelAll requests {}
>> >> >>>>>>> >>> >>> > >>
Thread-13::DEBUG::2014-11-24
>> >> >>>>>>> >>> >>> > >>
>> >> >>>>>>> >>> >>> > >>
>> >> >>>>>>> >>> >>> > >>
>> 21:41:32,394::task::993::Storage.TaskManager.Task::(_decref)
>> >> >>>>>>> >>> >>> > >>
Task=`994c7bc3-a236-4d03-a732-e068c7ed9ed4`::ref
>> 0
>> >> >>>>>>> >>> >>> > >>
aborting
>> >> >>>>>>> >>> >>> > >>
False
>> >> >>>>>>> >>> >>> > >>
Thread-13::DEBUG::2014-11-24
>> >> >>>>>>> >>> >>> > >>
21:41:41,550::BindingXMLRPC::1132::vds::(wrapper)
>> >> >>>>>>> >>> >>> > >>
client
>> >> >>>>>>> >>> >>> > >>
[10.10.10.2]::call
>> >> >>>>>>> >>> >>> > >>
getCapabilities with () {}
>> >> >>>>>>> >>> >>> > >>
Thread-13::DEBUG::2014-11-24
>> >> >>>>>>> >>> >>> > >>
21:41:41,553::utils::738::root::(execCmd)
>> >> >>>>>>> >>> >>> > >>
/sbin/ip route show to 0.0.0.0/0 table all (cwd
>> None)
>> >> >>>>>>> >>> >>> > >>
Thread-13::DEBUG::2014-11-24
>> >> >>>>>>> >>> >>> > >>
21:41:41,560::utils::758::root::(execCmd)
>> >> >>>>>>> >>> >>> > >>
SUCCESS: <err> = ''; <rc> = 0
>> >> >>>>>>> >>> >>> > >>
Thread-13::DEBUG::2014-11-24
>> >> >>>>>>> >>> >>> > >>
21:41:41,588::caps::728::root::(_getKeyPackages)
>> rpm
>> >> >>>>>>> >>> >>> > >>
package
>> >> >>>>>>> >>> >>> > >>
('gluster-swift',) not found
>> >> >>>>>>> >>> >>> > >>
Thread-13::DEBUG::2014-11-24
>> >> >>>>>>> >>> >>> > >>
21:41:41,592::caps::728::root::(_getKeyPackages)
>> rpm
>> >> >>>>>>> >>> >>> > >>
package
>> >> >>>>>>> >>> >>> > >>
('gluster-swift-object',) not found
>> >> >>>>>>> >>> >>> > >>
Thread-13::DEBUG::2014-11-24
>> >> >>>>>>> >>> >>> > >>
21:41:41,593::caps::728::root::(_getKeyPackages)
>> rpm
>> >> >>>>>>> >>> >>> > >>
package
>> >> >>>>>>> >>> >>> > >>
('gluster-swift-plugin',) not found
>> >> >>>>>>> >>> >>> > >>
Thread-13::DEBUG::2014-11-24
>> >> >>>>>>> >>> >>> > >>
21:41:41,598::caps::728::root::(_getKeyPackages)
>> rpm
>> >> >>>>>>> >>> >>> > >>
package
>> >> >>>>>>> >>> >>> > >>
('gluster-swift-account',) not found
>> >> >>>>>>> >>> >>> > >>
Thread-13::DEBUG::2014-11-24
>> >> >>>>>>> >>> >>> > >>
21:41:41,598::caps::728::root::(_getKeyPackages)
>> rpm
>> >> >>>>>>> >>> >>> > >>
package
>> >> >>>>>>> >>> >>> > >>
('gluster-swift-proxy',) not found
>> >> >>>>>>> >>> >>> > >>
Thread-13::DEBUG::2014-11-24
>> >> >>>>>>> >>> >>> > >>
21:41:41,598::caps::728::root::(_getKeyPackages)
>> rpm
>> >> >>>>>>> >>> >>> > >>
package
>> >> >>>>>>> >>> >>> > >>
('gluster-swift-doc',) not found
>> >> >>>>>>> >>> >>> > >>
Thread-13::DEBUG::2014-11-24
>> >> >>>>>>> >>> >>> > >>
21:41:41,599::caps::728::root::(_getKeyPackages)
>> rpm
>> >> >>>>>>> >>> >>> > >>
package
>> >> >>>>>>> >>> >>> > >>
('gluster-swift-container',) not found
>> >> >>>>>>> >>> >>> > >>
Thread-13::DEBUG::2014-11-24
>> >> >>>>>>> >>> >>> > >>
21:41:41,599::caps::728::root::(_getKeyPackages)
>> rpm
>> >> >>>>>>> >>> >>> > >>
package
>> >> >>>>>>> >>> >>> > >>
('glusterfs-geo-replication',) not found
>> >> >>>>>>> >>> >>> > >>
Thread-13::DEBUG::2014-11-24
>> >> >>>>>>> >>> >>> > >>
21:41:41,600::caps::646::root::(get)
>> >> >>>>>>> >>> >>> > >>
VirtioRNG DISABLED: libvirt version
>> 0.10.2-29.el6_5.9
>> >> >>>>>>> >>> >>> > >>
required
>> >> >>>>>>> >>> >>> > >>
>=
>> >> >>>>>>> >>> >>> > >>
0.10.2-31
>> >> >>>>>>> >>> >>> > >>
Thread-13::DEBUG::2014-11-24
>> >> >>>>>>> >>> >>> > >>
21:41:41,603::BindingXMLRPC::1139::vds::(wrapper)
>> >> >>>>>>> >>> >>> > >>
return
>> >> >>>>>>> >>> >>> > >>
getCapabilities
>> >> >>>>>>> >>> >>> > >>
with {'status': {'message': 'Done', 'code': 0},
>> >> >>>>>>> >>> >>> > >>
'info':
>> >> >>>>>>> >>> >>> > >>
{'HBAInventory':
>> >> >>>>>>> >>> >>> > >>
{'iSCSI': [{'InitiatorName':
>> >> >>>>>>> >>> >>> > >>
'iqn.1994-05.com.redhat:32151ce183c8'}],
>> >> >>>>>>> >>> >>> > >>
'FC':
>> >> >>>>>>> >>> >>> > >>
[]}, 'packages2': {'kernel': {'release':
>> >> >>>>>>> >>> >>> > >>
'431.el6.x86_64',
>> >> >>>>>>> >>> >>> > >>
'buildtime':
>> >> >>>>>>> >>> >>> > >>
1385061309.0, 'version': '2.6.32'},
>> 'glusterfs-rdma':
>> >> >>>>>>> >>> >>> > >>
{'release':
>> >> >>>>>>> >>> >>> > >>
'1.el6',
>> >> >>>>>>> >>> >>> > >>
'buildtime': 1403622628L, 'version': '3.5.1'},
>> >> >>>>>>> >>> >>> > >>
'glusterfs-fuse':
>> >> >>>>>>> >>> >>> > >>
{'release': '1.el6', 'buildtime': 1403622628L,
>> >> >>>>>>> >>> >>> > >>
'version':
>> >> >>>>>>> >>> >>> > >>
'3.5.1'},
>> >> >>>>>>> >>> >>> > >>
'spice-server': {'release': '6.el6_5.2',
>> 'buildtime':
>> >> >>>>>>> >>> >>> > >>
1402324637L,
>> >> >>>>>>> >>> >>> > >>
'version': '0.12.4'}, 'vdsm': {'release':
>> >> >>>>>>> >>> >>> > >>
'1.gitdb83943.el6',
>> >> >>>>>>> >>> >>> > >>
'buildtime':
>> >> >>>>>>> >>> >>> > >>
1412784567L, 'version': '4.16.7'}, 'qemu-kvm':
>> >> >>>>>>> >>> >>> > >>
{'release':
>> >> >>>>>>> >>> >>> > >>
'2.415.el6_5.10', 'buildtime': 1402435700L,
>> >> >>>>>>> >>> >>> > >>
'version':
>> >> >>>>>>> >>> >>> > >>
'0.12.1.2'},
>> >> >>>>>>> >>> >>> > >>
'qemu-img': {'release': '2.415.el6_5.10',
>> >> >>>>>>> >>> >>> > >>
'buildtime':
>> >> >>>>>>> >>> >>> > >>
1402435700L,
>> >> >>>>>>> >>> >>> > >>
'version': '0.12.1.2'}, 'libvirt': {'release':
>> >> >>>>>>> >>> >>> > >>
'29.el6_5.9',
>> >> >>>>>>> >>> >>> > >>
'buildtime':
>> >> >>>>>>> >>> >>> > >>
1402404612L, 'version': '0.10.2'}, 'glusterfs':
>> >> >>>>>>> >>> >>> > >>
{'release':
>> >> >>>>>>> >>> >>> > >>
'1.el6',
>> >> >>>>>>> >>> >>> > >>
'buildtime': 1403622628L, 'version': '3.5.1'},
>> 'mom':
>> >> >>>>>>> >>> >>> > >>
{'release':
>> >> >>>>>>> >>> >>> > >>
'2.el6',
>> >> >>>>>>> >>> >>> > >>
'buildtime': 1403794344L, 'version': '0.4.1'},
>> >> >>>>>>> >>> >>> > >>
'glusterfs-server':
>> >> >>>>>>> >>> >>> > >>
{'release': '1.el6', 'buildtime': 1403622628L,
>> >> >>>>>>> >>> >>> > >>
'version':
>> >> >>>>>>> >>> >>> > >>
'3.5.1'}},
>> >> >>>>>>> >>> >>> > >>
'numaNodeDistance': {'1': [20, 10], '0': [10,
>> 20]},
>> >> >>>>>>> >>> >>> > >>
'cpuModel':
>> >> >>>>>>> >>> >>> > >>
'Intel(R)
>> >> >>>>>>> >>> >>> > >>
Xeon(R) CPU X5650 @ 2.67GHz',
>> 'liveMerge':
>> >> >>>>>>> >>> >>> > >>
'false',
>> >> >>>>>>> >>> >>> > >>
'hooks':
>> >> >>>>>>> >>> >>> > >>
{},
>> >> >>>>>>> >>> >>> > >>
'cpuSockets': '2', 'vmTypes': ['kvm'], 'selinux':
>> >> >>>>>>> >>> >>> > >>
{'mode': '1'},
>> >> >>>>>>> >>> >>> > >>
'kdumpStatus': 0, 'supportedProtocols': ['2.2',
>> >> >>>>>>> >>> >>> > >>
'2.3'],
>> >> >>>>>>> >>> >>> > >>
'networks':
>> >> >>>>>>> >>> >>> > >>
{'ovirtmgmt': {'iface': u'bond0.10', 'addr':
>> >> >>>>>>> >>> >>> > >>
'43.252.176.16',
>> >> >>>>>>> >>> >>> > >>
'bridged':
>> >> >>>>>>> >>> >>> > >>
False, 'ipv6addrs':
>> ['fe80::62eb:69ff:fe20:b46c/64'],
>> >> >>>>>>> >>> >>> > >>
'mtu':
>> >> >>>>>>> >>> >>> > >>
'1500',
>> >> >>>>>>> >>> >>> > >>
'bootproto4': 'none', 'netmask': '255.255.255.0',
>> >> >>>>>>> >>> >>> > >>
'ipv4addrs':
>> >> >>>>>>> >>> >>> > >>
['
>> >> >>>>>>> >>> >>> > >>
43.252.176.16/24' <
http://43.252.176.16/24%27>],
>> >> >>>>>>> >>> >>> > >>
'interface':
>> >> >>>>>>> >>> >>> > >>
u'bond0.10', 'ipv6gateway': '::', 'gateway':
>> >> >>>>>>> >>> >>> > >>
'43.25.17.1'},
>> >> >>>>>>> >>> >>> > >>
'Internal':
>> >> >>>>>>> >>> >>> > >>
{'iface': 'Internal', 'addr': '', 'cfg':
>> {'DEFROUTE':
>> >> >>>>>>> >>> >>> > >>
'no',
>> >> >>>>>>> >>> >>> > >>
'HOTPLUG':
>> >> >>>>>>> >>> >>> > >>
'no', 'MTU': '9000', 'DELAY': '0',
>> 'NM_CONTROLLED':
>> >> >>>>>>> >>> >>> > >>
'no',
>> >> >>>>>>> >>> >>> > >>
'BOOTPROTO':
>> >> >>>>>>> >>> >>> > >>
'none', 'STP': 'off', 'DEVICE': 'Internal',
>> 'TYPE':
>> >> >>>>>>> >>> >>> > >>
'Bridge',
>> >> >>>>>>> >>> >>> > >>
'ONBOOT':
>> >> >>>>>>> >>> >>> > >>
'no'}, 'bridged': True, 'ipv6addrs':
>> >> >>>>>>> >>> >>> > >>
['fe80::210:18ff:fecd:daac/64'],
>> >> >>>>>>> >>> >>> > >>
'gateway': '', 'bootproto4': 'none', 'netmask':
>> '',
>> >> >>>>>>> >>> >>> > >>
'stp':
>> >> >>>>>>> >>> >>> > >>
'off',
>> >> >>>>>>> >>> >>> > >>
'ipv4addrs': [], 'mtu': '9000', 'ipv6gateway':
>> '::',
>> >> >>>>>>> >>> >>> > >>
'ports':
>> >> >>>>>>> >>> >>> > >>
['bond1.100']}, 'storage': {'iface': u'bond1',
>> >> >>>>>>> >>> >>> > >>
'addr':
>> >> >>>>>>> >>> >>> > >>
'10.10.10.6',
>> >> >>>>>>> >>> >>> > >>
'bridged': False, 'ipv6addrs':
>> >> >>>>>>> >>> >>> > >>
['fe80::210:18ff:fecd:daac/64'],
>> >> >>>>>>> >>> >>> > >>
'mtu':
>> >> >>>>>>> >>> >>> > >>
'9000', 'bootproto4': 'none', 'netmask':
>> >> >>>>>>> >>> >>> > >>
'255.255.255.0',
>> >> >>>>>>> >>> >>> > >>
'ipv4addrs': ['
>> >> >>>>>>> >>> >>> > >>
10.10.10.6/24' <
http://10.10.10.6/24%27>],
>> >> >>>>>>> >>> >>> > >>
'interface':
>> >> >>>>>>> >>> >>> > >>
u'bond1',
>> >> >>>>>>> >>> >>> > >>
'ipv6gateway': '::', 'gateway': ''}, 'VMNetwork':
>> >> >>>>>>> >>> >>> > >>
{'iface':
>> >> >>>>>>> >>> >>> > >>
'VMNetwork',
>> >> >>>>>>> >>> >>> > >>
'addr': '', 'cfg': {'DEFROUTE': 'no',
'HOTPLUG':
>> >> >>>>>>> >>> >>> > >>
'no',
>> >> >>>>>>> >>> >>> > >>
'MTU':
>> >> >>>>>>> >>> >>> > >>
'1500',
>> >> >>>>>>> >>> >>> > >>
'DELAY': '0', 'NM_CONTROLLED': 'no', 'BOOTPROTO':
>> >> >>>>>>> >>> >>> > >>
'none',
>> >> >>>>>>> >>> >>> > >>
'STP':
>> >> >>>>>>> >>> >>> > >>
'off',
>> >> >>>>>>> >>> >>> > >>
'DEVICE': 'VMNetwork', 'TYPE': 'Bridge',
>> 'ONBOOT':
>> >> >>>>>>> >>> >>> > >>
'no'},
>> >> >>>>>>> >>> >>> > >>
'bridged':
>> >> >>>>>>> >>> >>> > >>
True,
>> >> >>>>>>> >>> >>> > >>
'ipv6addrs': ['fe80::62eb:69ff:fe20:b46c/64'],
>> >> >>>>>>> >>> >>> > >>
'gateway':
>> >> >>>>>>> >>> >>> > >>
'',
>> >> >>>>>>> >>> >>> > >>
'bootproto4':
>> >> >>>>>>> >>> >>> > >>
'none', 'netmask': '', 'stp': 'off',
>> 'ipv4addrs': [],
>> >> >>>>>>> >>> >>> > >>
'mtu':
>> >> >>>>>>> >>> >>> > >>
'1500',
>> >> >>>>>>> >>> >>> > >>
'ipv6gateway': '::', 'ports': ['bond0.36']}},
>> >> >>>>>>> >>> >>> > >>
'bridges':
>> >> >>>>>>> >>> >>> > >>
{'Internal':
>> >> >>>>>>> >>> >>> > >>
{'addr': '', 'cfg': {'DEFROUTE': 'no',
'HOTPLUG':
>> >> >>>>>>> >>> >>> > >>
'no',
>> >> >>>>>>> >>> >>> > >>
'MTU':
>> >> >>>>>>> >>> >>> > >>
'9000',
>> >> >>>>>>> >>> >>> > >>
'DELAY': '0', 'NM_CONTROLLED': 'no', 'BOOTPROTO':
>> >> >>>>>>> >>> >>> > >>
'none',
>> >> >>>>>>> >>> >>> > >>
'STP':
>> >> >>>>>>> >>> >>> > >>
'off',
>> >> >>>>>>> >>> >>> > >>
'DEVICE': 'Internal', 'TYPE': 'Bridge', 'ONBOOT':
>> >> >>>>>>> >>> >>> > >>
'no'},
>> >> >>>>>>> >>> >>> > >>
'ipv6addrs':
>> >> >>>>>>> >>> >>> > >>
['fe80::210:18ff:fecd:daac/64'], 'mtu': '9000',
>> >> >>>>>>> >>> >>> > >>
'netmask': '',
>> >> >>>>>>> >>> >>> > >>
'stp':
>> >> >>>>>>> >>> >>> > >>
'off', 'ipv4addrs': [], 'ipv6gateway': '::',
>> >> >>>>>>> >>> >>> > >>
'gateway':
>> >> >>>>>>> >>> >>> > >>
'',
>> >> >>>>>>> >>> >>> > >>
'opts':
>> >> >>>>>>> >>> >>> > >>
{'topology_change_detected': '0',
>> >> >>>>>>> >>> >>> > >>
'multicast_last_member_count':
>> >> >>>>>>> >>> >>> > >>
'2',
>> >> >>>>>>> >>> >>> > >>
'hash_elasticity': '4',
>> >> >>>>>>> >>> >>> > >>
'multicast_query_response_interval':
>> >> >>>>>>> >>> >>> > >>
'999',
>> >> >>>>>>> >>> >>> > >>
'multicast_snooping': '1',
>> >> >>>>>>> >>> >>> > >>
'multicast_startup_query_interval':
>> >> >>>>>>> >>> >>> > >>
'3124',
>> >> >>>>>>> >>> >>> > >>
'hello_timer': '31',
>> 'multicast_querier_interval':
>> >> >>>>>>> >>> >>> > >>
'25496',
>> >> >>>>>>> >>> >>> > >>
'max_age':
>> >> >>>>>>> >>> >>> > >>
'1999', 'hash_max': '512', 'stp_state': '0',
>> >> >>>>>>> >>> >>> > >>
'root_id':
>> >> >>>>>>> >>> >>> > >>
'8000.001018cddaac', 'priority': '32768',
>> >> >>>>>>> >>> >>> > >>
'multicast_membership_interval':
>> >> >>>>>>> >>> >>> > >>
'25996', 'root_path_cost': '0', 'root_port': '0',
>> >> >>>>>>> >>> >>> > >>
'multicast_querier':
>> >> >>>>>>> >>> >>> > >>
'0',
>> >> >>>>>>> >>> >>> > >>
'multicast_startup_query_count': '2',
>> 'hello_time':
>> >> >>>>>>> >>> >>> > >>
'199',
>> >> >>>>>>> >>> >>> > >>
'topology_change': '0', 'bridge_id':
>> >> >>>>>>> >>> >>> > >>
'8000.001018cddaac',
>> >> >>>>>>> >>> >>> > >>
'topology_change_timer': '0', 'ageing_time':
>> '29995',
>> >> >>>>>>> >>> >>> > >>
'gc_timer':
>> >> >>>>>>> >>> >>> > >>
'31',
>> >> >>>>>>> >>> >>> > >>
'group_addr': '1:80:c2:0:0:0', 'tcn_timer': '0',
>> >> >>>>>>> >>> >>> > >>
'multicast_query_interval': '12498',
>> >> >>>>>>> >>> >>> > >>
'multicast_last_member_interval':
>> >> >>>>>>> >>> >>> > >>
'99', 'multicast_router': '1', 'forward_delay':
>> '0'},
>> >> >>>>>>> >>> >>> > >>
'ports':
>> >> >>>>>>> >>> >>> > >>
['bond1.100']}, 'VMNetwork': {'addr': '', 'cfg':
>> >> >>>>>>> >>> >>> > >>
{'DEFROUTE':
>> >> >>>>>>> >>> >>> > >>
'no',
>> >> >>>>>>> >>> >>> > >>
'HOTPLUG': 'no', 'MTU': '1500', 'DELAY':
'0',
>> >> >>>>>>> >>> >>> > >>
'NM_CONTROLLED':
>> >> >>>>>>> >>> >>> > >>
'no',
>> >> >>>>>>> >>> >>> > >>
'BOOTPROTO': 'none', 'STP': 'off', 'DEVICE':
>> >> >>>>>>> >>> >>> > >>
'VMNetwork',
>> >> >>>>>>> >>> >>> > >>
'TYPE':
>> >> >>>>>>> >>> >>> > >>
'Bridge',
>> >> >>>>>>> >>> >>> > >>
'ONBOOT': 'no'}, 'ipv6addrs':
>> >> >>>>>>> >>> >>> > >>
['fe80::62eb:69ff:fe20:b46c/64'],
>> >> >>>>>>> >>> >>> > >>
'mtu':
>> >> >>>>>>> >>> >>> > >>
'1500', 'netmask': '', 'stp': 'off',
>> 'ipv4addrs': [],
>> >> >>>>>>> >>> >>> > >>
'ipv6gateway':
>> >> >>>>>>> >>> >>> > >>
'::',
>> >> >>>>>>> >>> >>> > >>
'gateway': '', 'opts':
>> {'topology_change_detected':
>> >> >>>>>>> >>> >>> > >>
'0',
>> >> >>>>>>> >>> >>> > >>
'multicast_last_member_count': '2',
>> >> >>>>>>> >>> >>> > >>
'hash_elasticity':
>> >> >>>>>>> >>> >>> > >>
'4',
>> >> >>>>>>> >>> >>> > >>
'multicast_query_response_interval': '999',
>> >> >>>>>>> >>> >>> > >>
'multicast_snooping':
>> >> >>>>>>> >>> >>> > >>
'1',
>> >> >>>>>>> >>> >>> > >>
'multicast_startup_query_interval': '3124',
>> >> >>>>>>> >>> >>> > >>
'hello_timer':
>> >> >>>>>>> >>> >>> > >>
'131',
>> >> >>>>>>> >>> >>> > >>
'multicast_querier_interval': '25496', 'max_age':
>> >> >>>>>>> >>> >>> > >>
'1999',
>> >> >>>>>>> >>> >>> > >>
'hash_max':
>> >> >>>>>>> >>> >>> > >>
'512', 'stp_state': '0', 'root_id':
>> >> >>>>>>> >>> >>> > >>
'8000.60eb6920b46c',
>> >> >>>>>>> >>> >>> > >>
'priority':
>> >> >>>>>>> >>> >>> > >>
'32768', 'multicast_membership_interval':
>> '25996',
>> >> >>>>>>> >>> >>> > >>
'root_path_cost':
>> >> >>>>>>> >>> >>> > >>
'0',
>> >> >>>>>>> >>> >>> > >>
'root_port': '0', 'multicast_querier': '0',
>> >> >>>>>>> >>> >>> > >>
'multicast_startup_query_count': '2',
>> 'hello_time':
>> >> >>>>>>> >>> >>> > >>
'199',
>> >> >>>>>>> >>> >>> > >>
'topology_change': '0', 'bridge_id':
>> >> >>>>>>> >>> >>> > >>
'8000.60eb6920b46c',
>> >> >>>>>>> >>> >>> > >>
'topology_change_timer': '0', 'ageing_time':
>> '29995',
>> >> >>>>>>> >>> >>> > >>
'gc_timer':
>> >> >>>>>>> >>> >>> > >>
'31',
>> >> >>>>>>> >>> >>> > >>
'group_addr': '1:80:c2:0:0:0', 'tcn_timer': '0',
>> >> >>>>>>> >>> >>> > >>
'multicast_query_interval': '12498',
>> >> >>>>>>> >>> >>> > >>
'multicast_last_member_interval':
>> >> >>>>>>> >>> >>> > >>
'99', 'multicast_router': '1', 'forward_delay':
>> '0'},
>> >> >>>>>>> >>> >>> > >>
'ports':
>> >> >>>>>>> >>> >>> > >>
['bond0.36']}}, 'uuid':
>> >> >>>>>>> >>> >>> > >>
'44454C4C-4C00-1057-8053-B7C04F504E31',
>> >> >>>>>>> >>> >>> > >>
'lastClientIface': 'bond1', 'nics': {'eth3':
>> >> >>>>>>> >>> >>> > >>
{'permhwaddr':
>> >> >>>>>>> >>> >>> > >>
'00:10:18:cd:da:ae', 'addr': '', 'cfg': {'SLAVE':
>> >> >>>>>>> >>> >>> > >>
'yes',
>> >> >>>>>>> >>> >>> > >>
'NM_CONTROLLED':
>> >> >>>>>>> >>> >>> > >>
'no', 'MTU': '9000', 'HWADDR':
>> '00:10:18:cd:da:ae',
>> >> >>>>>>> >>> >>> > >>
'MASTER':
>> >> >>>>>>> >>> >>> > >>
'bond1',
>> >> >>>>>>> >>> >>> > >>
'DEVICE': 'eth3', 'ONBOOT': 'no'}, 'ipv6addrs':
>> [],
>> >> >>>>>>> >>> >>> > >>
'mtu':
>> >> >>>>>>> >>> >>> > >>
'9000',
>> >> >>>>>>> >>> >>> > >>
'netmask': '', 'ipv4addrs': [], 'hwaddr':
>> >> >>>>>>> >>> >>> > >>
'00:10:18:cd:da:ac',
>> >> >>>>>>> >>> >>> > >>
'speed':
>> >> >>>>>>> >>> >>> > >>
1000}, 'eth2': {'permhwaddr':
>> '00:10:18:cd:da:ac',
>> >> >>>>>>> >>> >>> > >>
'addr': '',
>> >> >>>>>>> >>> >>> > >>
'cfg':
>> >> >>>>>>> >>> >>> > >>
{'SLAVE': 'yes', 'NM_CONTROLLED': 'no', 'MTU':
>> >> >>>>>>> >>> >>> > >>
'9000',
>> >> >>>>>>> >>> >>> > >>
'HWADDR':
>> >> >>>>>>> >>> >>> > >>
'00:10:18:cd:da:ac', 'MASTER': 'bond1', 'DEVICE':
>> >> >>>>>>> >>> >>> > >>
'eth2',
>> >> >>>>>>> >>> >>> > >>
'ONBOOT':
>> >> >>>>>>> >>> >>> > >>
'no'},
>> >> >>>>>>> >>> >>> > >>
'ipv6addrs': [], 'mtu': '9000', 'netmask': '',
>> >> >>>>>>> >>> >>> > >>
'ipv4addrs': [],
>> >> >>>>>>> >>> >>> > >>
'hwaddr':
>> >> >>>>>>> >>> >>> > >>
'00:10:18:cd:da:ac', 'speed': 1000}, 'eth1':
>> >> >>>>>>> >>> >>> > >>
{'permhwaddr':
>> >> >>>>>>> >>> >>> > >>
'60:eb:69:20:b4:6d', 'addr': '', 'cfg': {'SLAVE':
>> >> >>>>>>> >>> >>> > >>
'yes',
>> >> >>>>>>> >>> >>> > >>
'NM_CONTROLLED':
>> >> >>>>>>> >>> >>> > >>
'no', 'MTU': '1500', 'HWADDR':
>> '60:eb:69:20:b4:6d',
>> >> >>>>>>> >>> >>> > >>
'MASTER':
>> >> >>>>>>> >>> >>> > >>
'bond0',
>> >> >>>>>>> >>> >>> > >>
'DEVICE': 'eth1', 'ONBOOT': 'yes'}, 'ipv6addrs':
>> [],
>> >> >>>>>>> >>> >>> > >>
'mtu':
>> >> >>>>>>> >>> >>> > >>
'1500',
>> >> >>>>>>> >>> >>> > >>
'netmask': '', 'ipv4addrs': [], 'hwaddr':
>> >> >>>>>>> >>> >>> > >>
'60:eb:69:20:b4:6c',
>> >> >>>>>>> >>> >>> > >>
'speed':
>> >> >>>>>>> >>> >>> > >>
1000}, 'eth0': {'permhwaddr':
>> '60:eb:69:20:b4:6c',
>> >> >>>>>>> >>> >>> > >>
'addr': '',
>> >> >>>>>>> >>> >>> > >>
'cfg':
>> >> >>>>>>> >>> >>> > >>
{'SLAVE': 'yes', 'NM_CONTROLLED': 'no', 'MTU':
>> >> >>>>>>> >>> >>> > >>
'1500',
>> >> >>>>>>> >>> >>> > >>
'HWADDR':
>> >> >>>>>>> >>> >>> > >>
'60:eb:69:20:b4:6c', 'MASTER': 'bond0', 'DEVICE':
>> >> >>>>>>> >>> >>> > >>
'eth0',
>> >> >>>>>>> >>> >>> > >>
'ONBOOT':
>> >> >>>>>>> >>> >>> > >>
'yes'},
>> >> >>>>>>> >>> >>> > >>
'ipv6addrs': [], 'mtu': '1500', 'netmask': '',
>> >> >>>>>>> >>> >>> > >>
'ipv4addrs': [],
>> >> >>>>>>> >>> >>> > >>
'hwaddr':
>> >> >>>>>>> >>> >>> > >>
'60:eb:69:20:b4:6c', 'speed': 1000}},
>> >> >>>>>>> >>> >>> > >>
'software_revision': '1',
>> >> >>>>>>> >>> >>> > >>
'clusterLevels': ['3.0', '3.1', '3.2', '3.3',
>> '3.4',
>> >> >>>>>>> >>> >>> > >>
'3.5'],
>> >> >>>>>>> >>> >>> > >>
'cpuFlags':
>> >> >>>>>>> >>> >>> > >>
>> >> >>>>>>> >>> >>> > >>
>> >> >>>>>>> >>> >>> > >>
>> >> >>>>>>> >>> >>> > >>
>> >> >>>>>>> >>> >>> > >>
>>
u'fpu,vme,de,pse,tsc,msr,pae,mce,cx8,apic,sep,mtrr,pge,mca,cmov,pat,pse36,clflush,dts,acpi,mmx,fxsr,sse,sse2,ss,ht,tm,pbe,syscall,nx,pdpe1gb,rdtscp,lm,constant_tsc,arch_perfmon,pebs,bts,rep_good,xtopology,nonstop_tsc,pni,pclmulqdq,dtes64,monitor,ds_cpl,vmx,smx,est,tm2,ssse3,cx16,xtpr,pdcm,pcid,dca,sse4_1,sse4_2,popcnt,aes,lahf_lm,tpr_shadow,vnmi,flexpriority,ept,vpid,model_Nehalem,model_Conroe,model_coreduo,model_core2duo,model_Penryn,model_Westmere,model_n270',
>> >> >>>>>>> >>> >>> > >>
'ISCSIInitiatorName':
>> >> >>>>>>> >>> >>> > >>
'iqn.1994-05.com.redhat:32151ce183c8',
>> >> >>>>>>> >>> >>> > >>
'netConfigDirty': 'False', 'supportedENGINEs':
>> >> >>>>>>> >>> >>> > >>
['3.0',
>> >> >>>>>>> >>> >>> > >>
'3.1',
>> >> >>>>>>> >>> >>> > >>
'3.2',
>> >> >>>>>>> >>> >>> > >>
'3.3',
>> >> >>>>>>> >>> >>> > >>
'3.4', '3.5'], 'autoNumaBalancing': 2,
>> 'reservedMem':
>> >> >>>>>>> >>> >>> > >>
'321',
>> >> >>>>>>> >>> >>> > >>
'bondings':
>> >> >>>>>>> >>> >>> > >>
{'bond4': {'addr': '', 'cfg': {}, 'mtu':
'1500',
>> >> >>>>>>> >>> >>> > >>
'netmask': '',
>> >> >>>>>>> >>> >>> > >>
'slaves':
>> >> >>>>>>> >>> >>> > >> [],
'hwaddr': '00:00:00:00:00:00'}, 'bond0':
>> {'addr':
>> >> >>>>>>> >>> >>> > >>
'',
>> >> >>>>>>> >>> >>> > >>
'cfg':
>> >> >>>>>>> >>> >>> > >>
{'HOTPLUG': 'no', 'MTU': '1500', 'NM_CONTROLLED':
>> >> >>>>>>> >>> >>> > >>
'no',
>> >> >>>>>>> >>> >>> > >>
'BONDING_OPTS':
>> >> >>>>>>> >>> >>> > >>
'mode=4 miimon=100', 'DEVICE': 'bond0', 'ONBOOT':
>> >> >>>>>>> >>> >>> > >>
'yes'},
>> >> >>>>>>> >>> >>> > >>
'ipv6addrs':
>> >> >>>>>>> >>> >>> > >>
['fe80::62eb:69ff:fe20:b46c/64'], 'mtu': '1500',
>> >> >>>>>>> >>> >>> > >>
'netmask': '',
>> >> >>>>>>> >>> >>> > >>
'ipv4addrs': [], 'hwaddr': '60:eb:69:20:b4:6c',
>> >> >>>>>>> >>> >>> > >>
'slaves':
>> >> >>>>>>> >>> >>> > >>
['eth0',
>> >> >>>>>>> >>> >>> > >>
'eth1'],
>> >> >>>>>>> >>> >>> > >>
'opts': {'miimon': '100', 'mode': '4'}},
'bond1':
>> >> >>>>>>> >>> >>> > >>
{'addr':
>> >> >>>>>>> >>> >>> > >>
'10.10.10.6',
>> >> >>>>>>> >>> >>> > >>
'cfg': {'DEFROUTE': 'no', 'IPADDR': '10.10.10.6',
>> >> >>>>>>> >>> >>> > >>
'HOTPLUG':
>> >> >>>>>>> >>> >>> > >>
'no',
>> >> >>>>>>> >>> >>> > >>
'MTU':
>> >> >>>>>>> >>> >>> > >>
'9000', 'NM_CONTROLLED': 'no', 'NETMASK':
>> >> >>>>>>> >>> >>> > >>
'255.255.255.0',
>> >> >>>>>>> >>> >>> > >>
'BOOTPROTO':
>> >> >>>>>>> >>> >>> > >>
'none', 'BONDING_OPTS': 'mode=4 miimon=100',
>> >> >>>>>>> >>> >>> > >>
'DEVICE':
>> >> >>>>>>> >>> >>> > >>
'bond1',
>> >> >>>>>>> >>> >>> > >>
'ONBOOT':
>> >> >>>>>>> >>> >>> > >>
'no'}, 'ipv6addrs':
>> ['fe80::210:18ff:fecd:daac/64'],
>> >> >>>>>>> >>> >>> > >>
'mtu':
>> >> >>>>>>> >>> >>> > >>
'9000',
>> >> >>>>>>> >>> >>> > >>
'netmask': '255.255.255.0', 'ipv4addrs':
>> >> >>>>>>> >>> >>> > >>
['10.10.10.6/24'
>> >> >>>>>>> >>> >>> > >>
<
http://10.10.10.6/24%27>], 'hwaddr':
>> >> >>>>>>> >>> >>> > >>
'00:10:18:cd:da:ac',
>> >> >>>>>>> >>> >>> > >>
'slaves':
>> >> >>>>>>> >>> >>> > >>
['eth2', 'eth3'], 'opts': {'miimon': '100',
>> 'mode':
>> >> >>>>>>> >>> >>> > >>
'4'}},
>> >> >>>>>>> >>> >>> > >>
'bond2':
>> >> >>>>>>> >>> >>> > >>
{'addr': '', 'cfg': {}, 'mtu': '1500',
>> 'netmask': '',
>> >> >>>>>>> >>> >>> > >>
'slaves':
>> >> >>>>>>> >>> >>> > >>
[],
>> >> >>>>>>> >>> >>> > >>
'hwaddr': '00:00:00:00:00:00'}, 'bond3':
>> {'addr': '',
>> >> >>>>>>> >>> >>> > >>
'cfg': {},
>> >> >>>>>>> >>> >>> > >>
'mtu':
>> >> >>>>>>> >>> >>> > >>
'1500', 'netmask': '', 'slaves': [], 'hwaddr':
>> >> >>>>>>> >>> >>> > >>
'00:00:00:00:00:00'}},
>> >> >>>>>>> >>> >>> > >>
'software_version': '4.16', 'memSize': '24019',
>> >> >>>>>>> >>> >>> > >>
'cpuSpeed':
>> >> >>>>>>> >>> >>> > >>
'2667.000',
>> >> >>>>>>> >>> >>> > >>
'numaNodes': {u'1': {'totalMemory': '12288',
>> 'cpus':
>> >> >>>>>>> >>> >>> > >>
[6,
>> >> >>>>>>> >>> >>> > >> 7,
8,
>> >> >>>>>>> >>> >>> > >> 9,
>> >> >>>>>>> >>> >>> > >> 10,
11,
>> >> >>>>>>> >>> >>> > >> 18,
19, 20, 21, 22, 23]}, u'0': {'totalMemory':
>> >> >>>>>>> >>> >>> > >>
'12278',
>> >> >>>>>>> >>> >>> > >>
'cpus':
>> >> >>>>>>> >>> >>> > >>
[0,
>> >> >>>>>>> >>> >>> > >> 1,
2,
>> >> >>>>>>> >>> >>> > >> 3,
4, 5, 12, 13, 14, 15, 16, 17]}},
>> 'version_name':
>> >> >>>>>>> >>> >>> > >>
'Snow
>> >> >>>>>>> >>> >>> > >>
Man',
>> >> >>>>>>> >>> >>> > >>
'vlans':
>> >> >>>>>>> >>> >>> > >>
{'bond0.10': {'iface': 'bond0', 'addr':
>> >> >>>>>>> >>> >>> > >>
'43.25.17.16',
>> >> >>>>>>> >>> >>> > >>
'cfg':
>> >> >>>>>>> >>> >>> > >>
{'DEFROUTE':
>> >> >>>>>>> >>> >>> > >>
'yes', 'VLAN': 'yes', 'IPADDR': '43.25.17.16',
>> >> >>>>>>> >>> >>> > >>
'HOTPLUG':
>> >> >>>>>>> >>> >>> > >>
'no',
>> >> >>>>>>> >>> >>> > >>
'GATEWAY':
>> >> >>>>>>> >>> >>> > >>
'43.25.17.1', 'NM_CONTROLLED': 'no', 'NETMASK':
>> >> >>>>>>> >>> >>> > >>
'255.255.255.0',
>> >> >>>>>>> >>> >>> > >>
'BOOTPROTO': 'none', 'DEVICE': 'bond0.10', 'MTU':
>> >> >>>>>>> >>> >>> > >>
'1500',
>> >> >>>>>>> >>> >>> > >>
'ONBOOT':
>> >> >>>>>>> >>> >>> > >>
'yes'},
>> >> >>>>>>> >>> >>> > >>
'ipv6addrs': ['fe80::62eb:69ff:fe20:b46c/64'],
>> >> >>>>>>> >>> >>> > >>
'vlanid':
>> >> >>>>>>> >>> >>> > >>
10,
>> >> >>>>>>> >>> >>> > >>
'mtu':
>> >> >>>>>>> >>> >>> > >>
'1500',
>> >> >>>>>>> >>> >>> > >>
'netmask': '255.255.255.0', 'ipv4addrs':
>> >> >>>>>>> >>> >>> > >>
['43.25.17.16/24']
>> >> >>>>>>> >>> >>> > >>
<
http://43.25.17.16/24%27%5D>}, 'bond0.36':
>> {'iface':
>> >> >>>>>>> >>> >>> > >>
'bond0',
>> >> >>>>>>> >>> >>> > >>
'addr':
>> >> >>>>>>> >>> >>> > >>
'', 'cfg': {'BRIDGE': 'VMNetwork', 'VLAN':
'yes',
>> >> >>>>>>> >>> >>> > >>
'HOTPLUG':
>> >> >>>>>>> >>> >>> > >>
'no',
>> >> >>>>>>> >>> >>> > >>
'MTU':
>> >> >>>>>>> >>> >>> > >>
'1500', 'NM_CONTROLLED': 'no', 'DEVICE':
>> 'bond0.36',
>> >> >>>>>>> >>> >>> > >>
'ONBOOT':
>> >> >>>>>>> >>> >>> > >>
'no'},
>> >> >>>>>>> >>> >>> > >>
'ipv6addrs': ['fe80::62eb:69ff:fe20:b46c/64'],
>> >> >>>>>>> >>> >>> > >>
'vlanid':
>> >> >>>>>>> >>> >>> > >>
36,
>> >> >>>>>>> >>> >>> > >>
'mtu':
>> >> >>>>>>> >>> >>> > >>
'1500',
>> >> >>>>>>> >>> >>> > >>
'netmask': '', 'ipv4addrs': []}, 'bond1.100':
>> >> >>>>>>> >>> >>> > >>
{'iface':
>> >> >>>>>>> >>> >>> > >>
'bond1',
>> >> >>>>>>> >>> >>> > >>
'addr':
>> >> >>>>>>> >>> >>> > >>
'', 'cfg': {'BRIDGE': 'Internal', 'VLAN':
'yes',
>> >> >>>>>>> >>> >>> > >>
'HOTPLUG':
>> >> >>>>>>> >>> >>> > >>
'no',
>> >> >>>>>>> >>> >>> > >>
'MTU':
>> >> >>>>>>> >>> >>> > >>
'9000', 'NM_CONTROLLED': 'no', 'DEVICE':
>> 'bond1.100',
>> >> >>>>>>> >>> >>> > >>
'ONBOOT':
>> >> >>>>>>> >>> >>> > >>
'no'},
>> >> >>>>>>> >>> >>> > >>
'ipv6addrs': ['fe80::210:18ff:fecd:daac/64'],
>> >> >>>>>>> >>> >>> > >>
'vlanid':
>> >> >>>>>>> >>> >>> > >>
100,
>> >> >>>>>>> >>> >>> > >>
'mtu':
>> >> >>>>>>> >>> >>> > >>
'9000',
>> >> >>>>>>> >>> >>> > >>
'netmask': '', 'ipv4addrs': []}}, 'cpuCores':
>> '12',
>> >> >>>>>>> >>> >>> > >>
'kvmEnabled':
>> >> >>>>>>> >>> >>> > >>
'true',
>> >> >>>>>>> >>> >>> > >>
'guestOverhead': '65', 'cpuThreads': '24',
>> >> >>>>>>> >>> >>> > >>
'emulatedMachines':
>> >> >>>>>>> >>> >>> > >>
[u'rhel6.5.0', u'pc', u'rhel6.4.0', u'rhel6.3.0',
>> >> >>>>>>> >>> >>> > >>
u'rhel6.2.0',
>> >> >>>>>>> >>> >>> > >>
u'rhel6.1.0', u'rhel6.0.0', u'rhel5.5.0',
>> >> >>>>>>> >>> >>> > >>
u'rhel5.4.4',
>> >> >>>>>>> >>> >>> > >>
u'rhel5.4.0'],
>> >> >>>>>>> >>> >>> > >>
'operatingSystem': {'release':
>> '5.el6.centos.11.1',
>> >> >>>>>>> >>> >>> > >>
'version':
>> >> >>>>>>> >>> >>> > >>
'6',
>> >> >>>>>>> >>> >>> > >>
'name':
>> >> >>>>>>> >>> >>> > >>
'RHEL'}, 'lastClient': '10.10.10.2'}}
>> >> >>>>>>> >>> >>> > >>
Thread-13::DEBUG::2014-11-24
>> >> >>>>>>> >>> >>> > >>
21:41:41,620::BindingXMLRPC::1132::vds::(wrapper)
>> >> >>>>>>> >>> >>> > >>
client
>> >> >>>>>>> >>> >>> > >>
[10.10.10.2]::call
>> >> >>>>>>> >>> >>> > >>
getHardwareInfo with () {}
>> >> >>>>>>> >>> >>> > >>
Thread-13::DEBUG::2014-11-24
>> >> >>>>>>> >>> >>> > >>
21:41:41,621::BindingXMLRPC::1139::vds::(wrapper)
>> >> >>>>>>> >>> >>> > >>
return
>> >> >>>>>>> >>> >>> > >>
getHardwareInfo
>> >> >>>>>>> >>> >>> > >>
with {'status': {'message': 'Done', 'code': 0},
>> >> >>>>>>> >>> >>> > >>
'info':
>> >> >>>>>>> >>> >>> > >>
{'systemProductName': 'CS24-TY',
>> >> >>>>>>> >>> >>> > >>
'systemSerialNumber':
>> >> >>>>>>> >>> >>> > >>
'7LWSPN1',
>> >> >>>>>>> >>> >>> > >>
'systemFamily': 'Server', 'systemVersion': 'A00',
>> >> >>>>>>> >>> >>> > >>
'systemUUID':
>> >> >>>>>>> >>> >>> > >>
'44454c4c-4c00-1057-8053-b7c04f504e31',
>> >> >>>>>>> >>> >>> > >>
'systemManufacturer':
>> >> >>>>>>> >>> >>> > >>
'Dell'}}
>> >> >>>>>>> >>> >>> > >>
Thread-13::DEBUG::2014-11-24
>> >> >>>>>>> >>> >>> > >>
21:41:41,733::BindingXMLRPC::1132::vds::(wrapper)
>> >> >>>>>>> >>> >>> > >>
client
>> >> >>>>>>> >>> >>> > >>
[10.10.10.2]::call
>> >> >>>>>>> >>> >>> > >>
hostsList with () {} flowID [222e8036]
>> >> >>>>>>> >>> >>> > >>
Thread-13::ERROR::2014-11-24
>> >> >>>>>>> >>> >>> > >>
21:41:44,753::BindingXMLRPC::1148::vds::(wrapper)
>> >> >>>>>>> >>> >>> > >>
vdsm
>> >> >>>>>>> >>> >>> > >>
exception
>> >> >>>>>>> >>> >>> > >>
occured
>> >> >>>>>>> >>> >>> > >>
Traceback (most recent call last):
>> >> >>>>>>> >>> >>> > >>
File "/usr/share/vdsm/rpc/BindingXMLRPC.py",
>> line
>> >> >>>>>>> >>> >>> > >>
1135,
>> >> >>>>>>> >>> >>> > >> in
>> >> >>>>>>> >>> >>> > >>
wrapper
>> >> >>>>>>> >>> >>> > >>
res = f(*args, **kwargs)
>> >> >>>>>>> >>> >>> > >>
File "/usr/share/vdsm/gluster/api.py", line
>> 54, in
>> >> >>>>>>> >>> >>> > >>
wrapper
>> >> >>>>>>> >>> >>> > >>
rv = func(*args, **kwargs)
>> >> >>>>>>> >>> >>> > >>
File "/usr/share/vdsm/gluster/api.py", line
>> 251, in
>> >> >>>>>>> >>> >>> > >>
hostsList
>> >> >>>>>>> >>> >>> > >>
return {'hosts':
>> >> >>>>>>> >>> >>> > >>
self.svdsmProxy.glusterPeerStatus()}
>> >> >>>>>>> >>> >>> > >>
File "/usr/share/vdsm/supervdsm.py", line 50,
>> in
>> >> >>>>>>> >>> >>> > >>
__call__
>> >> >>>>>>> >>> >>> > >>
return callMethod()
>> >> >>>>>>> >>> >>> > >>
File "/usr/share/vdsm/supervdsm.py", line 48,
>> in
>> >> >>>>>>> >>> >>> > >>
<lambda>
>> >> >>>>>>> >>> >>> > >>
**kwargs)
>> >> >>>>>>> >>> >>> > >>
File "<string>", line 2, in glusterPeerStatus
>> >> >>>>>>> >>> >>> > >>
File
>> >> >>>>>>> >>> >>> > >>
>> "/usr/lib64/python2.6/multiprocessing/managers.py",
>> >> >>>>>>> >>> >>> > >>
line
>> >> >>>>>>> >>> >>> > >>
740,
>> >> >>>>>>> >>> >>> > >> in
>> >> >>>>>>> >>> >>> > >>
_callmethod
>> >> >>>>>>> >>> >>> > >>
raise convert_to_error(kind, result)
>> >> >>>>>>> >>> >>> > >>
GlusterCmdExecFailedException: Command execution
>> >> >>>>>>> >>> >>> > >>
failed
>> >> >>>>>>> >>> >>> > >>
error: Connection failed. Please check if gluster
>> >> >>>>>>> >>> >>> > >>
daemon
>> >> >>>>>>> >>> >>> > >> is
>> >> >>>>>>> >>> >>> > >>
operational.
>> >> >>>>>>> >>> >>> > >>
return code: 1
>> >> >>>>>>> >>> >>> > >>
Thread-13::DEBUG::2014-11-24
>> >> >>>>>>> >>> >>> > >>
>> >> >>>>>>> >>> >>> > >>
>> >> >>>>>>> >>> >>> > >>
>> >> >>>>>>> >>> >>> > >>
>> 21:41:50,949::task::595::Storage.TaskManager.Task::(_updateState)
>> >> >>>>>>> >>> >>> > >>
>> Task=`c9042986-c978-4b08-adb2-616f5299e115`::moving
>> >> >>>>>>> >>> >>> > >>
from
>> >> >>>>>>> >>> >>> > >>
state
>> >> >>>>>>> >>> >>> > >>
init
>> >> >>>>>>> >>> >>> > >>
->
>> >> >>>>>>> >>> >>> > >>
state preparing
>> >> >>>>>>> >>> >>> > >>
Thread-13::INFO::2014-11-24
>> >> >>>>>>> >>> >>> > >>
>> 21:41:50,950::logUtils::44::dispatcher::(wrapper) Run
>> >> >>>>>>> >>> >>> > >>
and
>> >> >>>>>>> >>> >>> > >>
protect:
>> >> >>>>>>> >>> >>> > >>
repoStats(options=None)
>> >> >>>>>>> >>> >>> > >>
Thread-13::INFO::2014-11-24
>> >> >>>>>>> >>> >>> > >>
>> 21:41:50,950::logUtils::47::dispatcher::(wrapper) Run
>> >> >>>>>>> >>> >>> > >>
and
>> >> >>>>>>> >>> >>> > >>
protect:
>> >> >>>>>>> >>> >>> > >>
repoStats, Return response: {}
>> >> >>>>>>> >>> >>> > >>
Thread-13::DEBUG::2014-11-24
>> >> >>>>>>> >>> >>> > >>
>> >> >>>>>>> >>> >>> > >>
>> >> >>>>>>> >>> >>> > >>
>> 21:41:50,950::task::1191::Storage.TaskManager.Task::(prepare)
>> >> >>>>>>> >>> >>> > >>
>> >> >>>>>>> >>> >>> > >>
>> Task=`c9042986-c978-4b08-adb2-616f5299e115`::finished: {}
>> >> >>>>>>> >>> >>> > >>
Thread-13::DEBUG::2014-11-24
>> >> >>>>>>> >>> >>> > >>
>> >> >>>>>>> >>> >>> > >>
>> >> >>>>>>> >>> >>> > >>
>> >> >>>>>>> >>> >>> > >>
>> 21:41:50,950::task::595::Storage.TaskManager.Task::(_updateState)
>> >> >>>>>>> >>> >>> > >>
>> Task=`c9042986-c978-4b08-adb2-616f5299e115`::moving
>> >> >>>>>>> >>> >>> > >>
from
>> >> >>>>>>> >>> >>> > >>
state
>> >> >>>>>>> >>> >>> > >>
preparing
>> >> >>>>>>> >>> >>> > >>
->
>> >> >>>>>>> >>> >>> > >>
state finished
>> >> >>>>>>> >>> >>> > >>
Thread-13::DEBUG::2014-11-24
>> >> >>>>>>> >>> >>> > >>
>> >> >>>>>>> >>> >>> > >>
>> >> >>>>>>> >>> >>> > >>
>> >> >>>>>>> >>> >>> > >>
>> >> >>>>>>> >>> >>> > >>
>> 21:41:50,951::resourceManager::940::Storage.ResourceManager.Owner::(releaseAll)
>> >> >>>>>>> >>> >>> > >>
Owner.releaseAll requests {} resources {}
>> >> >>>>>>> >>> >>> > >>
Thread-13::DEBUG::2014-11-24
>> >> >>>>>>> >>> >>> > >>
>> >> >>>>>>> >>> >>> > >>
>> >> >>>>>>> >>> >>> > >>
>> >> >>>>>>> >>> >>> > >>
>> >> >>>>>>> >>> >>> > >>
>> 21:41:50,951::resourceManager::977::Storage.ResourceManager.Owner::(cancelAll)
>> >> >>>>>>> >>> >>> > >>
Owner.cancelAll requests {}
>> >> >>>>>>> >>> >>> > >>
Thread-13::DEBUG::2014-11-24
>> >> >>>>>>> >>> >>> > >>
>> >> >>>>>>> >>> >>> > >>
>> >> >>>>>>> >>> >>> > >>
>> 21:41:50,951::task::993::Storage.TaskManager.Task::(_decref)
>> >> >>>>>>> >>> >>> > >>
Task=`c9042986-c978-4b08-adb2-616f5299e115`::ref
>> 0
>> >> >>>>>>> >>> >>> > >>
aborting
>> >> >>>>>>> >>> >>> > >>
False
>> >> >>>>>>> >>> >>> > >>
-------------------------------
>> >> >>>>>>> >>> >>> > >>
>> >> >>>>>>> >>> >>> > >>
[root@compute4 ~]# service glusterd status
>> >> >>>>>>> >>> >>> > >>
glusterd is stopped
>> >> >>>>>>> >>> >>> > >>
[root@compute4 ~]# chkconfig --list | grep
>> glusterd
>> >> >>>>>>> >>> >>> > >>
glusterd 0:off 1:off 2:on 3:on
>> 4:on
>> >> >>>>>>> >>> >>> > >>
5:on
>> >> >>>>>>> >>> >>> > >>
6:off
>> >> >>>>>>> >>> >>> > >>
[root@compute4 ~]#
>> >> >>>>>>> >>> >>> > >>
>> >> >>>>>>> >>> >>> > >>
Thanks,
>> >> >>>>>>> >>> >>> > >>
Punit
>> >> >>>>>>> >>> >>> > >>
>> >> >>>>>>> >>> >>> > >> On
Mon, Nov 24, 2014 at 6:36 PM, Kanagaraj
>> >> >>>>>>> >>> >>> > >>
<kmayilsa(a)redhat.com>
>> >> >>>>>>> >>> >>> > >>
wrote:
>> >> >>>>>>> >>> >>> > >>
>> >> >>>>>>> >>> >>> > >>>
Can you send the corresponding error in
>> vdsm.log
>> >> >>>>>>> >>> >>> > >>>
from
>> >> >>>>>>> >>> >>> > >>>
the
>> >> >>>>>>> >>> >>> > >>>
host?
>> >> >>>>>>> >>> >>> >
>>>
>> >> >>>>>>> >>> >>> > >>>
Also check if glusterd service is running.
>> >> >>>>>>> >>> >>> >
>>>
>> >> >>>>>>> >>> >>> > >>>
Thanks,
>> >> >>>>>>> >>> >>> > >>>
Kanagaraj
>> >> >>>>>>> >>> >>> >
>>>
>> >> >>>>>>> >>> >>> >
>>>
>> >> >>>>>>> >>> >>> > >>>
On 11/24/2014 03:39 PM, Punit Dambiwal wrote:
>> >> >>>>>>> >>> >>> >
>>>
>> >> >>>>>>> >>> >>> > >>>
Hi,
>> >> >>>>>>> >>> >>> >
>>>
>> >> >>>>>>> >>> >>> > >>>
After reboot my Hypervisior host can not
>> activate
>> >> >>>>>>> >>> >>> > >>>
again
>> >> >>>>>>> >>> >>> > >>>
in the
>> >> >>>>>>> >>> >>> > >>>
cluster
>> >> >>>>>>> >>> >>> > >>>
and failed with the following error :-
>> >> >>>>>>> >>> >>> >
>>>
>> >> >>>>>>> >>> >>> > >>>
Gluster command [<UNKNOWN>] failed on server...
>> >> >>>>>>> >>> >>> >
>>>
>> >> >>>>>>> >>> >>> > >>>
Engine logs :-
>> >> >>>>>>> >>> >>> >
>>>
>> >> >>>>>>> >>> >>> > >>>
2014-11-24 18:05:28,397 INFO
>> >> >>>>>>> >>> >>> >
>>>
>> >> >>>>>>> >>> >>> >
>>>
>> >> >>>>>>> >>> >>> >
>>>
>> >> >>>>>>> >>> >>> >
>>>
>> >> >>>>>>> >>> >>> >
>>>
>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand]
>> >> >>>>>>> >>> >>> > >>>
(DefaultQuartzScheduler_Worker-64) START,
>> >> >>>>>>> >>> >>> > >>>
GlusterVolumesListVDSCommand(HostName =
>> Compute4,
>> >> >>>>>>> >>> >>> > >>>
HostId
>> >> >>>>>>> >>> >>> > >>>
=
>> >> >>>>>>> >>> >>> > >>>
33648a90-200c-45ca-89d5-1ce305d79a6a), log id:
>> >> >>>>>>> >>> >>> > >>>
5f251c90
>> >> >>>>>>> >>> >>> > >>>
2014-11-24 18:05:30,609 INFO
>> >> >>>>>>> >>> >>> >
>>>
>> >> >>>>>>> >>> >>> >
>>>
>> >> >>>>>>> >>> >>> >
>>>
>> >> >>>>>>> >>> >>> >
>>>
>> >> >>>>>>> >>> >>> >
>>>
>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand]
>> >> >>>>>>> >>> >>> > >>>
(DefaultQuartzScheduler_Worker-64) FINISH,
>> >> >>>>>>> >>> >>> > >>>
GlusterVolumesListVDSCommand,
>> >> >>>>>>> >>> >>> > >>>
return:
>> >> >>>>>>> >>> >>> >
>>>
>> >> >>>>>>> >>> >>> >
>>>
>> >> >>>>>>> >>> >>> >
>>>
>> >> >>>>>>> >>> >>> >
>>>
>> >> >>>>>>> >>> >>> >
>>>
>>
{26ae1672-ee09-4a38-8fd2-72dd9974cc2b=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@d95203e0
>> },
>> >> >>>>>>> >>> >>> > >>>
log id: 5f251c90
>> >> >>>>>>> >>> >>> > >>>
2014-11-24 18:05:33,768 INFO
>> >> >>>>>>> >>> >>> > >>>
[org.ovirt.engine.core.bll.ActivateVdsCommand]
>> >> >>>>>>> >>> >>> > >>>
(ajp--127.0.0.1-8702-8)
>> >> >>>>>>> >>> >>> > >>>
[287d570d] Lock Acquired to object EngineLock
>> >> >>>>>>> >>> >>> > >>>
[exclusiveLocks=
>> >> >>>>>>> >>> >>> > >>>
key:
>> >> >>>>>>> >>> >>> > >>>
0bf6b00f-7947-4411-b55a-cc5eea2b381a value: VDS
>> >> >>>>>>> >>> >>> > >>>
, sharedLocks= ]
>> >> >>>>>>> >>> >>> > >>>
2014-11-24 18:05:33,795 INFO
>> >> >>>>>>> >>> >>> > >>>
[org.ovirt.engine.core.bll.ActivateVdsCommand]
>> >> >>>>>>> >>> >>> > >>>
(org.ovirt.thread.pool-8-thread-45) [287d570d]
>> >> >>>>>>> >>> >>> > >>>
Running
>> >> >>>>>>> >>> >>> > >>>
command:
>> >> >>>>>>> >>> >>> > >>>
ActivateVdsCommand internal: false. Entities
>> >> >>>>>>> >>> >>> > >>>
affected :
>> >> >>>>>>> >>> >>> > >>>
ID:
>> >> >>>>>>> >>> >>> > >>>
0bf6b00f-7947-4411-b55a-cc5eea2b381a Type:
>> VDSAction
>> >> >>>>>>> >>> >>> > >>>
group
>> >> >>>>>>> >>> >>> > >>>
MANIPULATE_HOST
>> >> >>>>>>> >>> >>> > >>>
with role type ADMIN
>> >> >>>>>>> >>> >>> > >>>
2014-11-24 18:05:33,796 INFO
>> >> >>>>>>> >>> >>> > >>>
[org.ovirt.engine.core.bll.ActivateVdsCommand]
>> >> >>>>>>> >>> >>> > >>>
(org.ovirt.thread.pool-8-thread-45) [287d570d]
>> >> >>>>>>> >>> >>> > >>>
Before
>> >> >>>>>>> >>> >>> > >>>
acquiring
>> >> >>>>>>> >>> >>> > >>>
lock in
>> >> >>>>>>> >>> >>> > >>>
order to prevent monitoring for host Compute5
>> from
>> >> >>>>>>> >>> >>> > >>>
data-center
>> >> >>>>>>> >>> >>> > >>>
SV_WTC
>> >> >>>>>>> >>> >>> > >>>
2014-11-24 18:05:33,797 INFO
>> >> >>>>>>> >>> >>> > >>>
[org.ovirt.engine.core.bll.ActivateVdsCommand]
>> >> >>>>>>> >>> >>> > >>>
(org.ovirt.thread.pool-8-thread-45) [287d570d]
>> Lock
>> >> >>>>>>> >>> >>> > >>>
acquired,
>> >> >>>>>>> >>> >>> > >>>
from
>> >> >>>>>>> >>> >>> > >>>
now a
>> >> >>>>>>> >>> >>> > >>>
monitoring of host will be skipped for host
>> Compute5
>> >> >>>>>>> >>> >>> > >>>
from
>> >> >>>>>>> >>> >>> > >>>
data-center
>> >> >>>>>>> >>> >>> > >>>
SV_WTC
>> >> >>>>>>> >>> >>> > >>>
2014-11-24 18:05:33,817 INFO
>> >> >>>>>>> >>> >>> >
>>>
>> >> >>>>>>> >>> >>> >
>>>
>> >> >>>>>>> >>> >>> >
>>>
>> [org.ovirt.engine.core.vdsbroker.SetVdsStatusVDSCommand]
>> >> >>>>>>> >>> >>> > >>>
(org.ovirt.thread.pool-8-thread-45) [287d570d]
>> >> >>>>>>> >>> >>> > >>>
START,
>> >> >>>>>>> >>> >>> > >>>
SetVdsStatusVDSCommand(HostName = Compute5,
>> HostId =
>> >> >>>>>>> >>> >>> > >>>
0bf6b00f-7947-4411-b55a-cc5eea2b381a,
>> >> >>>>>>> >>> >>> > >>>
status=Unassigned,
>> >> >>>>>>> >>> >>> > >>>
nonOperationalReason=NONE,
>> >> >>>>>>> >>> >>> > >>>
stopSpmFailureLogged=false),
>> >> >>>>>>> >>> >>> > >>>
log id:
>> >> >>>>>>> >>> >>> > >>>
1cbc7311
>> >> >>>>>>> >>> >>> > >>>
2014-11-24 18:05:33,820 INFO
>> >> >>>>>>> >>> >>> >
>>>
>> >> >>>>>>> >>> >>> >
>>>
>> >> >>>>>>> >>> >>> >
>>>
>> [org.ovirt.engine.core.vdsbroker.SetVdsStatusVDSCommand]
>> >> >>>>>>> >>> >>> > >>>
(org.ovirt.thread.pool-8-thread-45) [287d570d]
>> >> >>>>>>> >>> >>> > >>>
FINISH,
>> >> >>>>>>> >>> >>> > >>>
SetVdsStatusVDSCommand, log id: 1cbc7311
>> >> >>>>>>> >>> >>> > >>>
2014-11-24 18:05:34,086 INFO
>> >> >>>>>>> >>> >>> > >>>
[org.ovirt.engine.core.bll.ActivateVdsCommand]
>> >> >>>>>>> >>> >>> > >>>
(org.ovirt.thread.pool-8-thread-45) Activate
>> >> >>>>>>> >>> >>> > >>>
finished.
>> >> >>>>>>> >>> >>> > >>>
Lock
>> >> >>>>>>> >>> >>> > >>>
released.
>> >> >>>>>>> >>> >>> > >>>
Monitoring can run now for host Compute5 from
>> >> >>>>>>> >>> >>> > >>>
data-center
>> >> >>>>>>> >>> >>> > >>>
SV_WTC
>> >> >>>>>>> >>> >>> > >>>
2014-11-24 18:05:34,088 INFO
>> >> >>>>>>> >>> >>> >
>>>
>> >> >>>>>>> >>> >>> >
>>>
>> >> >>>>>>> >>> >>> >
>>>
>> >> >>>>>>> >>> >>> >
>>>
>> >> >>>>>>> >>> >>> >
>>>
>> [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
>> >> >>>>>>> >>> >>> > >>>
(org.ovirt.thread.pool-8-thread-45) Correlation
>> ID:
>> >> >>>>>>> >>> >>> > >>>
287d570d,
>> >> >>>>>>> >>> >>> > >>>
Job
>> >> >>>>>>> >>> >>> > >>>
ID:
>> >> >>>>>>> >>> >>> > >>>
5ef8e4d6-b2bc-469e-8e81-7ef74b2a001a, Call
>> Stack:
>> >> >>>>>>> >>> >>> > >>>
null,
>> >> >>>>>>> >>> >>> > >>>
Custom
>> >> >>>>>>> >>> >>> > >>>
Event ID:
>> >> >>>>>>> >>> >>> > >>>
-1, Message: Host Compute5 was activated by
>> admin.
>> >> >>>>>>> >>> >>> > >>>
2014-11-24 18:05:34,090 INFO
>> >> >>>>>>> >>> >>> > >>>
[org.ovirt.engine.core.bll.ActivateVdsCommand]
>> >> >>>>>>> >>> >>> > >>>
(org.ovirt.thread.pool-8-thread-45) Lock freed
>> to
>> >> >>>>>>> >>> >>> > >>>
object
>> >> >>>>>>> >>> >>> > >>>
EngineLock
>> >> >>>>>>> >>> >>> > >>>
[exclusiveLocks= key:
>> >> >>>>>>> >>> >>> > >>>
0bf6b00f-7947-4411-b55a-cc5eea2b381a
>> >> >>>>>>> >>> >>> > >>>
value:
>> >> >>>>>>> >>> >>> > >>>
VDS
>> >> >>>>>>> >>> >>> > >>>
, sharedLocks= ]
>> >> >>>>>>> >>> >>> > >>>
2014-11-24 18:05:35,792 INFO
>> >> >>>>>>> >>> >>> >
>>>
>> >> >>>>>>> >>> >>> >
>>>
>> >> >>>>>>> >>> >>> >
>>>
>> >> >>>>>>> >>> >>> >
>>>
>> >> >>>>>>> >>> >>> >
>>>
>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand]
>> >> >>>>>>> >>> >>> > >>>
(DefaultQuartzScheduler_Worker-55) [3706e836]
>> START,
>> >> >>>>>>> >>> >>> > >>>
GlusterVolumesListVDSCommand(HostName =
>> Compute4,
>> >> >>>>>>> >>> >>> > >>>
HostId
>> >> >>>>>>> >>> >>> > >>>
=
>> >> >>>>>>> >>> >>> > >>>
33648a90-200c-45ca-89d5-1ce305d79a6a), log id:
>> >> >>>>>>> >>> >>> > >>>
48a0c832
>> >> >>>>>>> >>> >>> > >>>
2014-11-24 18:05:37,064 INFO
>> >> >>>>>>> >>> >>> >
>>>
>> >> >>>>>>> >>> >>> >
>>>
>> >> >>>>>>> >>> >>> >
>>>
>> >> >>>>>>> >>> >>> >
>>>
>> >> >>>>>>> >>> >>> >
>>>
>> [org.ovirt.engine.core.vdsbroker.vdsbroker.GetHardwareInfoVDSCommand]
>> >> >>>>>>> >>> >>> > >>>
(DefaultQuartzScheduler_Worker-69) START,
>> >> >>>>>>> >>> >>> > >>>
GetHardwareInfoVDSCommand(HostName = Compute5,
>> >> >>>>>>> >>> >>> > >>>
HostId =
>> >> >>>>>>> >>> >>> > >>>
0bf6b00f-7947-4411-b55a-cc5eea2b381a,
>> >> >>>>>>> >>> >>> >
>>>
>> >> >>>>>>> >>> >>> >
>>>
>> >> >>>>>>> >>> >>> >
>>>
>> vds=Host[Compute5,0bf6b00f-7947-4411-b55a-cc5eea2b381a]), log
>> >> >>>>>>> >>> >>> > >>>
id:
>> >> >>>>>>> >>> >>> > >>>
6d560cc2
>> >> >>>>>>> >>> >>> > >>>
2014-11-24 18:05:37,074 INFO
>> >> >>>>>>> >>> >>> >
>>>
>> >> >>>>>>> >>> >>> >
>>>
>> >> >>>>>>> >>> >>> >
>>>
>> >> >>>>>>> >>> >>> >
>>>
>> >> >>>>>>> >>> >>> >
>>>
>> [org.ovirt.engine.core.vdsbroker.vdsbroker.GetHardwareInfoVDSCommand]
>> >> >>>>>>> >>> >>> > >>>
(DefaultQuartzScheduler_Worker-69) FINISH,
>> >> >>>>>>> >>> >>> > >>>
GetHardwareInfoVDSCommand, log
>> >> >>>>>>> >>> >>> > >>>
id: 6d560cc2
>> >> >>>>>>> >>> >>> > >>>
2014-11-24 18:05:37,093 WARN
>> >> >>>>>>> >>> >>> > >>>
[org.ovirt.engine.core.vdsbroker.VdsManager]
>> >> >>>>>>> >>> >>> > >>>
(DefaultQuartzScheduler_Worker-69) Host
>> Compute5 is
>> >> >>>>>>> >>> >>> > >>>
running
>> >> >>>>>>> >>> >>> > >>>
with
>> >> >>>>>>> >>> >>> > >>>
disabled
>> >> >>>>>>> >>> >>> > >>>
SELinux.
>> >> >>>>>>> >>> >>> > >>>
2014-11-24 18:05:37,127 INFO
>> >> >>>>>>> >>> >>> >
>>>
>> >> >>>>>>> >>> >>> >
>>>
>> >> >>>>>>> >>> >>> >
>>>
>> >> >>>>>>> >>> >>> >
>>>
>> >> >>>>>>> >>> >>> >
>>>
>> [org.ovirt.engine.core.bll.HandleVdsCpuFlagsOrClusterChangedCommand]
>> >> >>>>>>> >>> >>> > >>>
(DefaultQuartzScheduler_Worker-69) [2b4a51cf]
>> >> >>>>>>> >>> >>> > >>>
Running
>> >> >>>>>>> >>> >>> > >>>
command:
>> >> >>>>>>> >>> >>> > >>>
HandleVdsCpuFlagsOrClusterChangedCommand
>> internal:
>> >> >>>>>>> >>> >>> > >>>
true.
>> >> >>>>>>> >>> >>> > >>>
Entities
>> >> >>>>>>> >>> >>> > >>>
affected
>> >> >>>>>>> >>> >>> > >>>
: ID: 0bf6b00f-7947-4411-b55a-cc5eea2b381a
>> Type:
>> >> >>>>>>> >>> >>> > >>>
VDS
>> >> >>>>>>> >>> >>> > >>>
2014-11-24 18:05:37,147 INFO
>> >> >>>>>>> >>> >>> >
>>>
>> >> >>>>>>> >>> >>> >
>>>
>> >> >>>>>>> >>> >>> >
>>>
>> >> >>>>>>> >>> >>> >
>>>
>> >> >>>>>>> >>> >>> >
>>>
>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand]
>> >> >>>>>>> >>> >>> > >>>
(DefaultQuartzScheduler_Worker-69) [2b4a51cf]
>> START,
>> >> >>>>>>> >>> >>> > >>>
GlusterServersListVDSCommand(HostName =
>> Compute5,
>> >> >>>>>>> >>> >>> > >>>
HostId
>> >> >>>>>>> >>> >>> > >>>
=
>> >> >>>>>>> >>> >>> > >>>
0bf6b00f-7947-4411-b55a-cc5eea2b381a), log id:
>> >> >>>>>>> >>> >>> > >>>
4faed87
>> >> >>>>>>> >>> >>> > >>>
2014-11-24 18:05:37,164 INFO
>> >> >>>>>>> >>> >>> >
>>>
>> >> >>>>>>> >>> >>> >
>>>
>> >> >>>>>>> >>> >>> >
>>>
>> >> >>>>>>> >>> >>> >
>>>
>> >> >>>>>>> >>> >>> >
>>>
>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand]
>> >> >>>>>>> >>> >>> > >>>
(DefaultQuartzScheduler_Worker-69) [2b4a51cf]
>> >> >>>>>>> >>> >>> > >>>
FINISH,
>> >> >>>>>>> >>> >>> > >>>
GlusterServersListVDSCommand, log id: 4faed87
>> >> >>>>>>> >>> >>> > >>>
2014-11-24 18:05:37,189 INFO
>> >> >>>>>>> >>> >>> >
>>>
>> >> >>>>>>> >>> >>> >
>>>
>> [org.ovirt.engine.core.bll.SetNonOperationalVdsCommand]
>> >> >>>>>>> >>> >>> > >>>
(DefaultQuartzScheduler_Worker-69) [4a84c4e5]
>> >> >>>>>>> >>> >>> > >>>
Running
>> >> >>>>>>> >>> >>> > >>>
command:
>> >> >>>>>>> >>> >>> > >>>
SetNonOperationalVdsCommand internal: true.
>> Entities
>> >> >>>>>>> >>> >>> > >>>
affected :
>> >> >>>>>>> >>> >>> > >>>
ID:
>> >> >>>>>>> >>> >>> > >>>
0bf6b00f-7947-4411-b55a-cc5eea2b381a Type: VDS
>> >> >>>>>>> >>> >>> > >>>
2014-11-24 18:05:37,206 INFO
>> >> >>>>>>> >>> >>> >
>>>
>> >> >>>>>>> >>> >>> >
>>>
>> >> >>>>>>> >>> >>> >
>>>
>> [org.ovirt.engine.core.vdsbroker.SetVdsStatusVDSCommand]
>> >> >>>>>>> >>> >>> > >>>
(DefaultQuartzScheduler_Worker-69) [4a84c4e5]
>> START,
>> >> >>>>>>> >>> >>> > >>>
SetVdsStatusVDSCommand(HostName = Compute5,
>> HostId =
>> >> >>>>>>> >>> >>> > >>>
0bf6b00f-7947-4411-b55a-cc5eea2b381a,
>> >> >>>>>>> >>> >>> > >>>
status=NonOperational,
>> >> >>>>>>> >>> >>> > >>>
nonOperationalReason=GLUSTER_COMMAND_FAILED,
>> >> >>>>>>> >>> >>> > >>>
stopSpmFailureLogged=false),
>> >> >>>>>>> >>> >>> > >>>
log id: fed5617
>> >> >>>>>>> >>> >>> > >>>
2014-11-24 18:05:37,209 INFO
>> >> >>>>>>> >>> >>> >
>>>
>> >> >>>>>>> >>> >>> >
>>>
>> >> >>>>>>> >>> >>> >
>>>
>> [org.ovirt.engine.core.vdsbroker.SetVdsStatusVDSCommand]
>> >> >>>>>>> >>> >>> > >>>
(DefaultQuartzScheduler_Worker-69) [4a84c4e5]
>> >> >>>>>>> >>> >>> > >>>
FINISH,
>> >> >>>>>>> >>> >>> > >>>
SetVdsStatusVDSCommand, log id: fed5617
>> >> >>>>>>> >>> >>> > >>>
2014-11-24 18:05:37,223 ERROR
>> >> >>>>>>> >>> >>> >
>>>
>> >> >>>>>>> >>> >>> >
>>>
>> >> >>>>>>> >>> >>> >
>>>
>> >> >>>>>>> >>> >>> >
>>>
>> >> >>>>>>> >>> >>> >
>>>
>> [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
>> >> >>>>>>> >>> >>> > >>>
(DefaultQuartzScheduler_Worker-69) [4a84c4e5]
>> >> >>>>>>> >>> >>> > >>>
Correlation ID:
>> >> >>>>>>> >>> >>> > >>>
4a84c4e5,
>> >> >>>>>>> >>> >>> > >>>
Job
>> >> >>>>>>> >>> >>> > >>>
ID: 4bfd4a6d-c3ef-468f-a40e-a3a6ca13011b, Call
>> >> >>>>>>> >>> >>> > >>>
Stack:
>> >> >>>>>>> >>> >>> > >>>
null,
>> >> >>>>>>> >>> >>> > >>>
Custom
>> >> >>>>>>> >>> >>> > >>>
Event
>> >> >>>>>>> >>> >>> > >>>
ID: -1, Message: Gluster command [<UNKNOWN>]
>> failed
>> >> >>>>>>> >>> >>> > >>>
on
>> >> >>>>>>> >>> >>> > >>>
server
>> >> >>>>>>> >>> >>> > >>>
Compute5.
>> >> >>>>>>> >>> >>> > >>>
2014-11-24 18:05:37,243 INFO
>> >> >>>>>>> >>> >>> >
>>>
>> >> >>>>>>> >>> >>> >
>>>
>> >> >>>>>>> >>> >>> >
>>>
>> >> >>>>>>> >>> >>> >
>>>
>> >> >>>>>>> >>> >>> >
>>>
>> [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
>> >> >>>>>>> >>> >>> > >>>
(DefaultQuartzScheduler_Worker-69) [4a84c4e5]
>> >> >>>>>>> >>> >>> > >>>
Correlation ID:
>> >> >>>>>>> >>> >>> > >>>
null,
>> >> >>>>>>> >>> >>> > >>>
Call
>> >> >>>>>>> >>> >>> > >>>
Stack: null, Custom Event ID: -1, Message:
>> Status of
>> >> >>>>>>> >>> >>> > >>>
host
>> >> >>>>>>> >>> >>> > >>>
Compute5
>> >> >>>>>>> >>> >>> > >>>
was
>> >> >>>>>>> >>> >>> > >>>
set
>> >> >>>>>>> >>> >>> > >>>
to NonOperational.
>> >> >>>>>>> >>> >>> > >>>
2014-11-24 18:05:37,272 INFO
>> >> >>>>>>> >>> >>> >
>>>
>> [org.ovirt.engine.core.bll.HandleVdsVersionCommand]
>> >> >>>>>>> >>> >>> > >>>
(DefaultQuartzScheduler_Worker-69) [a0c8a7f]
>> Running
>> >> >>>>>>> >>> >>> > >>>
command:
>> >> >>>>>>> >>> >>> > >>>
HandleVdsVersionCommand internal: true. Entities
>> >> >>>>>>> >>> >>> > >>>
affected :
>> >> >>>>>>> >>> >>> > >>>
ID:
>> >> >>>>>>> >>> >>> > >>>
0bf6b00f-7947-4411-b55a-cc5eea2b381a Type: VDS
>> >> >>>>>>> >>> >>> > >>>
2014-11-24 18:05:37,274 INFO
>> >> >>>>>>> >>> >>> >
>>>
>> >> >>>>>>> >>> >>> >
>>>
>> [org.ovirt.engine.core.vdsbroker.VdsUpdateRunTimeInfo]
>> >> >>>>>>> >>> >>> > >>>
(DefaultQuartzScheduler_Worker-69) [a0c8a7f]
>> Host
>> >> >>>>>>> >>> >>> > >>>
0bf6b00f-7947-4411-b55a-cc5eea2b381a : Compute5
>> is
>> >> >>>>>>> >>> >>> > >>>
already in
>> >> >>>>>>> >>> >>> > >>>
NonOperational status for reason
>> >> >>>>>>> >>> >>> > >>>
GLUSTER_COMMAND_FAILED.
>> >> >>>>>>> >>> >>> > >>>
SetNonOperationalVds command is skipped.
>> >> >>>>>>> >>> >>> > >>>
2014-11-24 18:05:38,065 INFO
>> >> >>>>>>> >>> >>> >
>>>
>> >> >>>>>>> >>> >>> >
>>>
>> >> >>>>>>> >>> >>> >
>>>
>> >> >>>>>>> >>> >>> >
>>>
>> >> >>>>>>> >>> >>> >
>>>
>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand]
>> >> >>>>>>> >>> >>> > >>>
(DefaultQuartzScheduler_Worker-55) [3706e836]
>> >> >>>>>>> >>> >>> > >>>
FINISH,
>> >> >>>>>>> >>> >>> > >>>
GlusterVolumesListVDSCommand, return:
>> >> >>>>>>> >>> >>> >
>>>
>> >> >>>>>>> >>> >>> >
>>>
>> >> >>>>>>> >>> >>> >
>>>
>> >> >>>>>>> >>> >>> >
>>>
>> >> >>>>>>> >>> >>> >
>>>
>>
{26ae1672-ee09-4a38-8fd2-72dd9974cc2b=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@4e72a1b1
>> },
>> >> >>>>>>> >>> >>> > >>>
log id: 48a0c832
>> >> >>>>>>> >>> >>> > >>>
2014-11-24 18:05:43,243 INFO
>> >> >>>>>>> >>> >>> >
>>>
>> >> >>>>>>> >>> >>> >
>>>
>> >> >>>>>>> >>> >>> >
>>>
>> >> >>>>>>> >>> >>> >
>>>
>> >> >>>>>>> >>> >>> >
>>>
>> [org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand]
>> >> >>>>>>> >>> >>> > >>>
(DefaultQuartzScheduler_Worker-35) START,
>> >> >>>>>>> >>> >>> > >>>
GlusterVolumesListVDSCommand(HostName =
>> Compute4,
>> >> >>>>>>> >>> >>> > >>>
HostId
>> >> >>>>>>> >>> >>> > >>>
=
>> >> >>>>>>> >>> >>> > >>>
33648a90-200c-45ca-89d5-1ce305d79a6a), log id:
>> >> >>>>>>> >>> >>> > >>>
3ce13ebc
>> >> >>>>>>> >>> >>> > >>>
^C
>> >> >>>>>>> >>> >>> > >>>
[root@ccr01 ~]#
>> >> >>>>>>> >>> >>> >
>>>
>> >> >>>>>>> >>> >>> > >>>
Thanks,
>> >> >>>>>>> >>> >>> > >>>
Punit
>> >> >>>>>>> >>> >>> >
>>>
>> >> >>>>>>> >>> >>> >
>>>
>> >> >>>>>>> >>> >>> > >>>
_______________________________________________
>> >> >>>>>>> >>> >>> > >>>
Users mailing
>> >> >>>>>>> >>> >>> >
>>>
>> >> >>>>>>> >>> >>> >
>>>
>> >> >>>>>>> >>> >>> >
>>>
>> >> >>>>>>> >>> >>> > >>>
listUsers@ovirt.orghttp://
>>
lists.ovirt.org/mailman/listinfo/users
>> >> >>>>>>> >>> >>> >
>>>
>> >> >>>>>>> >>> >>> >
>>>
>> >> >>>>>>> >>> >>> >
>>>
>> >> >>>>>>> >>> >>> > >>
>> >> >>>>>>> >>> >>> > >>
>> >> >>>>>>> >>> >>> > >
>> >> >>>>>>> >>> >>> > >
>> >> >>>>>>> >>> >>> >
>> >> >>>>>>> >>> >>
>> >> >>>>>>> >>> >>
>> >> >>>>>>> >>> >
>> >> >>>>>>> >>> >
>> >> >>>>>>> >>> >
_______________________________________________
>> >> >>>>>>> >>> > Gluster-users mailing
list
>> >> >>>>>>> >>> >
Gluster-users(a)gluster.org
>> >> >>>>>>> >>> >
>> >> >>>>>>> >>> >
>>
http://supercolony.gluster.org/mailman/listinfo/gluster-users
>> >> >>>>>>> >>
>> >> >>>>>>> >>
>> >> >>>>>>> >
>> >> >>>>>>
>> >> >>>>>>
>> >> >>>>
>> >> >>
>> >
>> >
>>
>
>