I don't know much about how the network target is brought up in
CentOS7, but I'll try as much as I can.
It seems to me that, after the network has been brought up and by the
time GlusterD is started,
a. The machine hasn't yet recieved it's hostname, or
b. It hasn't yet registered with the name server.
This is causing name resolution failures.
I don't know if the network target could come up without the machine
getting its hostname, so I'm pretty sure it's not a.
So it seems to be b. But these kind of signing in happens only in DDNS
systems, which doesn't seem to be the case for you.
Both of these reasons might be wrong (most likely wrong). You'd do
good if you could ask for help from someone with more experience in
systemd + networking.
~kaushal
On Wed, Dec 3, 2014 at 10:54 AM, Punit Dambiwal <hypunit(a)gmail.com> wrote:
Hi Kaushal,
This is the host...which i rebooted...would you mind to let me know how i
can make the glusterd sevice come up after network...i am using centos7...if
network is the issue...
On Wed, Dec 3, 2014 at 11:54 AM, Kaushal M <kshlmster(a)gmail.com> wrote:
>
> This peer cannot be identified.
>
> " [2014-12-03 02:29:25.998153] D
> [glusterd-peer-utils.c:121:glusterd_peerinfo_find_by_hostname] 0-management:
> Unable to find friend: cpu05.zne01.hkg1.ovt.36stack.com"
>
> I don't know why this address is not being resolved during boot time. If
> this is a valid peer, the the only reason I can think of this that the
> network is not up.
>
> If you had previously detached the peer forcefully, the that could have
> left stale entries in some volumes. In this case as well, GlusterD will fail
> to identify the peer.
>
> Do either of these reasons seem a possibility to you?
>
> On Dec 3, 2014 8:07 AM, "Punit Dambiwal" <hypunit(a)gmail.com> wrote:
>>
>> Hi Kaushal,
>>
>> Please find the logs here :-
http://ur1.ca/iyoe5 and
http://ur1.ca/iyoed
>>
>> On Tue, Dec 2, 2014 at 10:43 PM, Kaushal M <kshlmster(a)gmail.com> wrote:
>>>
>>> Hey Punit,
>>> In the logs you've provided, GlusterD appears to be running correctly.
>>> Could you provide the logs for the time period when GlusterD attempts to
>>> start but fails.
>>>
>>> ~kaushal
>>>
>>> On Dec 2, 2014 8:03 PM, "Punit Dambiwal" <hypunit(a)gmail.com>
wrote:
>>>>
>>>> Hi Kaushal,
>>>>
>>>> Please find the logs here :-
http://ur1.ca/iyhs5 and
>>>>
http://ur1.ca/iyhue
>>>>
>>>> Thanks,
>>>> punit
>>>>
>>>>
>>>> On Tue, Dec 2, 2014 at 12:00 PM, Kaushal M <kshlmster(a)gmail.com>
wrote:
>>>>>
>>>>> Hey Punit,
>>>>> Could you start Glusterd in debug mode and provide the logs here?
>>>>> To start it in debug mode, append '-LDEBUG' to the ExecStart
line in
>>>>> the service file.
>>>>>
>>>>> ~kaushal
>>>>>
>>>>> On Mon, Dec 1, 2014 at 9:05 AM, Punit Dambiwal
<hypunit(a)gmail.com>
>>>>> wrote:
>>>>> > Hi,
>>>>> >
>>>>> > Can Any body help me on this ??
>>>>> >
>>>>> > On Thu, Nov 27, 2014 at 9:29 AM, Punit Dambiwal
<hypunit(a)gmail.com>
>>>>> > wrote:
>>>>> >>
>>>>> >> Hi Kaushal,
>>>>> >>
>>>>> >> Thanks for the detailed reply....let me explain my setup
first :-
>>>>> >>
>>>>> >> 1. Ovirt Engine
>>>>> >> 2. 4* host as well as storage machine (Host and gluster
combined)
>>>>> >> 3. Every host has 24 bricks...
>>>>> >>
>>>>> >> Now whenever the host machine reboot...it can come up but
can not
>>>>> >> join the
>>>>> >> cluster again and through the following error "Gluster
command
>>>>> >> [<UNKNOWN>]
>>>>> >> failed on server.."
>>>>> >>
>>>>> >> Please check my comment in line :-
>>>>> >>
>>>>> >> 1. Use the same string for doing the peer probe and for the
brick
>>>>> >> address
>>>>> >> during volume create/add-brick. Ideally, we suggest you use
>>>>> >> properly
>>>>> >> resolvable FQDNs everywhere. If that is not possible, then
use only
>>>>> >> IP
>>>>> >> addresses. Try to avoid short names.
>>>>> >> ---------------
>>>>> >> [root@cpu05 ~]# gluster peer status
>>>>> >> Number of Peers: 3
>>>>> >>
>>>>> >> Hostname:
cpu03.stack.com
>>>>> >> Uuid: 5729b8c4-e80d-4353-b456-6f467bddbdfb
>>>>> >> State: Peer in Cluster (Connected)
>>>>> >>
>>>>> >> Hostname:
cpu04.stack.com
>>>>> >> Uuid: d272b790-c4b2-4bed-ba68-793656e6d7b0
>>>>> >> State: Peer in Cluster (Connected)
>>>>> >> Other names:
>>>>> >> 10.10.0.8
>>>>> >>
>>>>> >> Hostname:
cpu02.stack.com
>>>>> >> Uuid: 8d8a7041-950e-40d0-85f9-58d14340ca25
>>>>> >> State: Peer in Cluster (Connected)
>>>>> >> [root@cpu05 ~]#
>>>>> >> ----------------
>>>>> >> 2. During boot up, make sure to launch glusterd only after
the
>>>>> >> network is
>>>>> >> up. This will allow the new peer identification mechanism to
do its
>>>>> >> job correctly.
>>>>> >> >> I think the service itself doing the same job....
>>>>> >>
>>>>> >> [root@cpu05 ~]# cat
/usr/lib/systemd/system/glusterd.service
>>>>> >> [Unit]
>>>>> >> Description=GlusterFS, a clustered file-system server
>>>>> >> After=network.target rpcbind.service
>>>>> >> Before=network-online.target
>>>>> >>
>>>>> >> [Service]
>>>>> >> Type=forking
>>>>> >> PIDFile=/var/run/glusterd.pid
>>>>> >> LimitNOFILE=65536
>>>>> >> ExecStart=/usr/sbin/glusterd -p /var/run/glusterd.pid
>>>>> >> KillMode=process
>>>>> >>
>>>>> >> [Install]
>>>>> >> WantedBy=multi-user.target
>>>>> >> [root@cpu05 ~]#
>>>>> >> --------------------
>>>>> >>
>>>>> >> gluster logs :-
>>>>> >>
>>>>> >> [2014-11-24 09:22:22.147471] I [MSGID: 100030]
>>>>> >> [glusterfsd.c:2018:main]
>>>>> >> 0-/usr/sbin/glusterd: Started running /usr/sbin/glusterd
version
>>>>> >> 3.6.1
>>>>> >> (args: /usr/sbin/glusterd -p /var/run/glusterd.pid)
>>>>> >> [2014-11-24 09:22:22.151565] I [glusterd.c:1214:init]
0-management:
>>>>> >> Maximum allowed open file descriptors set to 65536
>>>>> >> [2014-11-24 09:22:22.151599] I [glusterd.c:1259:init]
0-management:
>>>>> >> Using
>>>>> >> /var/lib/glusterd as working directory
>>>>> >> [2014-11-24 09:22:22.155216] W
[rdma.c:4195:__gf_rdma_ctx_create]
>>>>> >> 0-rpc-transport/rdma: rdma_cm event channel creation failed
(No
>>>>> >> such device)
>>>>> >> [2014-11-24 09:22:22.155264] E [rdma.c:4483:init]
>>>>> >> 0-rdma.management:
>>>>> >> Failed to initialize IB Device
>>>>> >> [2014-11-24 09:22:22.155285] E
>>>>> >> [rpc-transport.c:333:rpc_transport_load]
>>>>> >> 0-rpc-transport: 'rdma' initialization failed
>>>>> >> [2014-11-24 09:22:22.155354] W
>>>>> >> [rpcsvc.c:1524:rpcsvc_transport_create]
>>>>> >> 0-rpc-service: cannot create listener, initing the transport
failed
>>>>> >> [2014-11-24 09:22:22.156290] I
>>>>> >> [glusterd.c:413:glusterd_check_gsync_present] 0-glusterd:
>>>>> >> geo-replication
>>>>> >> module not installed in the system
>>>>> >> [2014-11-24 09:22:22.161318] I
>>>>> >> [glusterd-store.c:2043:glusterd_restore_op_version]
0-glusterd:
>>>>> >> retrieved
>>>>> >> op-version: 30600
>>>>> >> [2014-11-24 09:22:22.821800] I
>>>>> >> [glusterd-handler.c:3146:glusterd_friend_add_from_peerinfo]
>>>>> >> 0-management:
>>>>> >> connect returned 0
>>>>> >> [2014-11-24 09:22:22.825810] I
>>>>> >> [glusterd-handler.c:3146:glusterd_friend_add_from_peerinfo]
>>>>> >> 0-management:
>>>>> >> connect returned 0
>>>>> >> [2014-11-24 09:22:22.828705] I
>>>>> >> [glusterd-handler.c:3146:glusterd_friend_add_from_peerinfo]
>>>>> >> 0-management:
>>>>> >> connect returned 0
>>>>> >> [2014-11-24 09:22:22.828771] I
>>>>> >> [rpc-clnt.c:969:rpc_clnt_connection_init]
>>>>> >> 0-management: setting frame-timeout to 600
>>>>> >> [2014-11-24 09:22:22.832670] I
>>>>> >> [rpc-clnt.c:969:rpc_clnt_connection_init]
>>>>> >> 0-management: setting frame-timeout to 600
>>>>> >> [2014-11-24 09:22:22.835919] I
>>>>> >> [rpc-clnt.c:969:rpc_clnt_connection_init]
>>>>> >> 0-management: setting frame-timeout to 600
>>>>> >> [2014-11-24 09:22:22.840209] E
>>>>> >> [glusterd-store.c:4248:glusterd_resolve_all_bricks]
0-glusterd:
>>>>> >> resolve
>>>>> >> brick failed in restore
>>>>> >> [2014-11-24 09:22:22.840233] E [xlator.c:425:xlator_init]
>>>>> >> 0-management:
>>>>> >> Initialization of volume 'management' failed, review
your volfile
>>>>> >> again
>>>>> >> [2014-11-24 09:22:22.840245] E
[graph.c:322:glusterfs_graph_init]
>>>>> >> 0-management: initializing translator failed
>>>>> >> [2014-11-24 09:22:22.840264] E
>>>>> >> [graph.c:525:glusterfs_graph_activate]
>>>>> >> 0-graph: init failed
>>>>> >> [2014-11-24 09:22:22.840754] W
[glusterfsd.c:1194:cleanup_and_exit]
>>>>> >> (-->
>>>>> >> 0-: received signum (0), shutting down
>>>>> >>
>>>>> >> Thanks,
>>>>> >> Punit
>>>>> >>
>>>>> >>
>>>>> >>
>>>>> >>
>>>>> >> On Wed, Nov 26, 2014 at 7:14 PM, Kaushal M
<kshlmster(a)gmail.com>
>>>>> >> wrote:
>>>>> >>>
>>>>> >>> Based on the logs I can guess that glusterd is being
started
>>>>> >>> before
>>>>> >>> the network has come up and that the addresses given to
bricks do
>>>>> >>> not
>>>>> >>> directly match the addresses used in during peer probe.
>>>>> >>>
>>>>> >>> The gluster_after_reboot log has the line
"[2014-11-25
>>>>> >>> 06:46:09.972113] E
>>>>> >>> [glusterd-store.c:2632:glusterd_resolve_all_bricks]
>>>>> >>> 0-glusterd: resolve brick failed in restore".
>>>>> >>>
>>>>> >>> Brick resolution fails when glusterd cannot match the
address for
>>>>> >>> the
>>>>> >>> brick, with one of the peers. Brick resolution happens
in two
>>>>> >>> phases,
>>>>> >>> 1. We first try to identify the peer by performing
string
>>>>> >>> comparisions
>>>>> >>> with the brick address and the peer addresses (The peer
names will
>>>>> >>> be
>>>>> >>> the names/addresses that were given when the peer was
probed).
>>>>> >>> 2. If we don't find a match from step 1, we will
then resolve all
>>>>> >>> the
>>>>> >>> brick address and the peer addresses into addrinfo
structs, and
>>>>> >>> then
>>>>> >>> compare these structs to find a match. This process
should
>>>>> >>> generally
>>>>> >>> find a match if available. This will fail only if the
network is
>>>>> >>> not
>>>>> >>> up yet as we cannot resolve addresses.
>>>>> >>>
>>>>> >>> The above steps are applicable only to glusterfs
versions >=3.6.
>>>>> >>> They
>>>>> >>> were introduced to reduce problems with peer
identification, like
>>>>> >>> the
>>>>> >>> one you encountered
>>>>> >>>
>>>>> >>> Since both of the steps failed to find a match in one
run, but
>>>>> >>> succeeded later, we can come to the conclusion that,
>>>>> >>> a) the bricks don't have the exact same string used
in peer probe
>>>>> >>> for
>>>>> >>> their addresses as step 1 failed, and
>>>>> >>> b) the network was not up in the initial run, as step 2
failed
>>>>> >>> during
>>>>> >>> the initial run, but passed in the second run.
>>>>> >>>
>>>>> >>> Please let me know if my conclusion is correct.
>>>>> >>>
>>>>> >>> If it is, you can solve your problem in two ways.
>>>>> >>> 1. Use the same string for doing the peer probe and for
the brick
>>>>> >>> address during volume create/add-brick. Ideally, we
suggest you
>>>>> >>> use
>>>>> >>> properly resolvable FQDNs everywhere. If that is not
possible,
>>>>> >>> then
>>>>> >>> use only IP addresses. Try to avoid short names.
>>>>> >>> 2. During boot up, make sure to launch glusterd only
after the
>>>>> >>> network
>>>>> >>> is up. This will allow the new peer identification
mechanism to do
>>>>> >>> its
>>>>> >>> job correctly.
>>>>> >>>
>>>>> >>>
>>>>> >>> If you have already followed these steps and yet still
hit the
>>>>> >>> problem, then please provide more information (setup,
logs, etc.).
>>>>> >>> It
>>>>> >>> could be much different problem that you are facing.
>>>>> >>>
>>>>> >>> ~kaushal
>>>>> >>>
>>>>> >>> On Wed, Nov 26, 2014 at 4:01 PM, Punit Dambiwal
>>>>> >>> <hypunit(a)gmail.com>
>>>>> >>> wrote:
>>>>> >>> > Is there any one can help on this ??
>>>>> >>> >
>>>>> >>> > Thanks,
>>>>> >>> > punit
>>>>> >>> >
>>>>> >>> > On Wed, Nov 26, 2014 at 9:42 AM, Punit Dambiwal
>>>>> >>> > <hypunit(a)gmail.com>
>>>>> >>> > wrote:
>>>>> >>> >>
>>>>> >>> >> Hi,
>>>>> >>> >>
>>>>> >>> >> My Glusterfs version is :-
glusterfs-3.6.1-1.el7
>>>>> >>> >>
>>>>> >>> >> On Wed, Nov 26, 2014 at 1:59 AM, Kanagaraj
Mayilsamy
>>>>> >>> >> <kmayilsa(a)redhat.com>
>>>>> >>> >> wrote:
>>>>> >>> >>>
>>>>> >>> >>> [+Gluster-users(a)gluster.org]
>>>>> >>> >>>
>>>>> >>> >>> "Initialization of volume
'management' failed, review your
>>>>> >>> >>> volfile
>>>>> >>> >>> again", glusterd throws this error
when the service is started
>>>>> >>> >>> automatically
>>>>> >>> >>> after the reboot. But the service is
successfully started
>>>>> >>> >>> later
>>>>> >>> >>> manually by
>>>>> >>> >>> the user.
>>>>> >>> >>>
>>>>> >>> >>> can somebody from gluster-users please help
on this?
>>>>> >>> >>>
>>>>> >>> >>> glusterfs version: 3.5.1
>>>>> >>> >>>
>>>>> >>> >>> Thanks,
>>>>> >>> >>> Kanagaraj
>>>>> >>> >>>
>>>>> >>> >>> ----- Original Message -----
>>>>> >>> >>> > From: "Punit Dambiwal"
<hypunit(a)gmail.com>
>>>>> >>> >>> > To: "Kanagaraj"
<kmayilsa(a)redhat.com>
>>>>> >>> >>> > Cc: users(a)ovirt.org
>>>>> >>> >>> > Sent: Tuesday, November 25, 2014
7:24:45 PM
>>>>> >>> >>> > Subject: Re: [ovirt-users] Gluster
command [<UNKNOWN>]
>>>>> >>> >>> > failed on
>>>>> >>> >>> > server...
>>>>> >>> >>> >
>>>>> >>> >>> > Hi Kanagraj,
>>>>> >>> >>> >
>>>>> >>> >>> > Please check the attached log
files....i didn't find any
>>>>> >>> >>> > thing
>>>>> >>> >>> > special....
>>>>> >>> >>> >
>>>>> >>> >>> > On Tue, Nov 25, 2014 at 12:12 PM,
Kanagaraj
>>>>> >>> >>> > <kmayilsa(a)redhat.com>
>>>>> >>> >>> > wrote:
>>>>> >>> >>> >
>>>>> >>> >>> > > Do you see any errors in
>>>>> >>> >>> > >
/var/log/glusterfs/etc-glusterfs-glusterd.vol.log or
>>>>> >>> >>> > > vdsm.log
>>>>> >>> >>> > > when
>>>>> >>> >>> > > the
>>>>> >>> >>> > > service is trying to start
automatically after the reboot?
>>>>> >>> >>> > >
>>>>> >>> >>> > > Thanks,
>>>>> >>> >>> > > Kanagaraj
>>>>> >>> >>> > >
>>>>> >>> >>> > >
>>>>> >>> >>> > > On 11/24/2014 08:13 PM, Punit
Dambiwal wrote:
>>>>> >>> >>> > >
>>>>> >>> >>> > > Hi Kanagaraj,
>>>>> >>> >>> > >
>>>>> >>> >>> > > Yes...once i will start the
gluster service and then
>>>>> >>> >>> > > vdsmd
>>>>> >>> >>> > > ...the
>>>>> >>> >>> > > host
>>>>> >>> >>> > > can connect to cluster...but the
question is why it's not
>>>>> >>> >>> > > started
>>>>> >>> >>> > > even it
>>>>> >>> >>> > > has chkconfig enabled...
>>>>> >>> >>> > >
>>>>> >>> >>> > > I have tested it in two host
cluster
>>>>> >>> >>> > > environment...(Centos 6.6
>>>>> >>> >>> > > and
>>>>> >>> >>> > > centos 7.0) on both hypervisior
cluster..it's failed to
>>>>> >>> >>> > > reconnect
>>>>> >>> >>> > > in
>>>>> >>> >>> > > to
>>>>> >>> >>> > > cluster after reboot....
>>>>> >>> >>> > >
>>>>> >>> >>> > > In both the environment glusterd
enabled for next
>>>>> >>> >>> > > boot....but
>>>>> >>> >>> > > it's
>>>>> >>> >>> > > failed with the same
error....seems it's bug in either
>>>>> >>> >>> > > gluster or
>>>>> >>> >>> > > Ovirt ??
>>>>> >>> >>> > >
>>>>> >>> >>> > > Please help me to find the
workaround here if can not
>>>>> >>> >>> > > resolve
>>>>> >>> >>> > > it...as
>>>>> >>> >>> > > without this the Host machine can
not connect after
>>>>> >>> >>> > > reboot....that
>>>>> >>> >>> > > means
>>>>> >>> >>> > > engine will consider it as down
and every time need to
>>>>> >>> >>> > > manually
>>>>> >>> >>> > > start
>>>>> >>> >>> > > the
>>>>> >>> >>> > > gluster service and vdsmd... ??
>>>>> >>> >>> > >
>>>>> >>> >>> > > Thanks,
>>>>> >>> >>> > > Punit
>>>>> >>> >>> > >
>>>>> >>> >>> > > On Mon, Nov 24, 2014 at 10:20 PM,
Kanagaraj
>>>>> >>> >>> > > <kmayilsa(a)redhat.com>
>>>>> >>> >>> > > wrote:
>>>>> >>> >>> > >
>>>>> >>> >>> > >> From vdsm.log "error:
Connection failed. Please check if
>>>>> >>> >>> > >> gluster
>>>>> >>> >>> > >> daemon
>>>>> >>> >>> > >> is operational."
>>>>> >>> >>> > >>
>>>>> >>> >>> > >> Starting glusterd service
should fix this issue. 'service
>>>>> >>> >>> > >> glusterd
>>>>> >>> >>> > >> start'
>>>>> >>> >>> > >> But i am wondering why the
glusterd was not started
>>>>> >>> >>> > >> automatically
>>>>> >>> >>> > >> after
>>>>> >>> >>> > >> the reboot.
>>>>> >>> >>> > >>
>>>>> >>> >>> > >> Thanks,
>>>>> >>> >>> > >> Kanagaraj
>>>>> >>> >>> > >>
>>>>> >>> >>> > >>
>>>>> >>> >>> > >>
>>>>> >>> >>> > >> On 11/24/2014 07:18 PM, Punit
Dambiwal wrote:
>>>>> >>> >>> > >>
>>>>> >>> >>> > >> Hi Kanagaraj,
>>>>> >>> >>> > >>
>>>>> >>> >>> > >> Please find the attached
VDSM logs :-
>>>>> >>> >>> > >>
>>>>> >>> >>> > >> ----------------
>>>>> >>> >>> > >>
Thread-13::DEBUG::2014-11-24
>>>>> >>> >>> > >>
>>>>> >>> >>> > >>
>>>>> >>> >>> > >>
>>>>> >>> >>> > >>
21:41:17,182::resourceManager::977::Storage.ResourceManager.Owner::(cancelAll)
>>>>> >>> >>> > >> Owner.cancelAll requests {}
>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24
>>>>> >>> >>> > >>
>>>>> >>> >>> > >>
21:41:17,182::task::993::Storage.TaskManager.Task::(_decref)
>>>>> >>> >>> > >>
Task=`1691d409-9b27-4585-8281-5ec26154367a`::ref 0
>>>>> >>> >>> > >> aborting
>>>>> >>> >>> > >> False
>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24
>>>>> >>> >>> > >>
>>>>> >>> >>> > >>
>>>>> >>> >>> > >>
21:41:32,393::task::595::Storage.TaskManager.Task::(_updateState)
>>>>> >>> >>> > >>
Task=`994c7bc3-a236-4d03-a732-e068c7ed9ed4`::moving from
>>>>> >>> >>> > >> state
>>>>> >>> >>> > >> init
>>>>> >>> >>> > >> ->
>>>>> >>> >>> > >> state preparing
>>>>> >>> >>> > >> Thread-13::INFO::2014-11-24
>>>>> >>> >>> > >>
21:41:32,393::logUtils::44::dispatcher::(wrapper) Run and
>>>>> >>> >>> > >> protect:
>>>>> >>> >>> > >> repoStats(options=None)
>>>>> >>> >>> > >> Thread-13::INFO::2014-11-24
>>>>> >>> >>> > >>
21:41:32,393::logUtils::47::dispatcher::(wrapper) Run and
>>>>> >>> >>> > >> protect:
>>>>> >>> >>> > >> repoStats, Return response:
{}
>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24
>>>>> >>> >>> > >>
>>>>> >>> >>> > >>
21:41:32,393::task::1191::Storage.TaskManager.Task::(prepare)
>>>>> >>> >>> > >>
Task=`994c7bc3-a236-4d03-a732-e068c7ed9ed4`::finished: {}
>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24
>>>>> >>> >>> > >>
>>>>> >>> >>> > >>
>>>>> >>> >>> > >>
21:41:32,394::task::595::Storage.TaskManager.Task::(_updateState)
>>>>> >>> >>> > >>
Task=`994c7bc3-a236-4d03-a732-e068c7ed9ed4`::moving from
>>>>> >>> >>> > >> state
>>>>> >>> >>> > >> preparing
>>>>> >>> >>> > >> ->
>>>>> >>> >>> > >> state finished
>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24
>>>>> >>> >>> > >>
>>>>> >>> >>> > >>
>>>>> >>> >>> > >>
>>>>> >>> >>> > >>
21:41:32,394::resourceManager::940::Storage.ResourceManager.Owner::(releaseAll)
>>>>> >>> >>> > >> Owner.releaseAll requests {}
resources {}
>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24
>>>>> >>> >>> > >>
>>>>> >>> >>> > >>
>>>>> >>> >>> > >>
>>>>> >>> >>> > >>
21:41:32,394::resourceManager::977::Storage.ResourceManager.Owner::(cancelAll)
>>>>> >>> >>> > >> Owner.cancelAll requests {}
>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24
>>>>> >>> >>> > >>
>>>>> >>> >>> > >>
21:41:32,394::task::993::Storage.TaskManager.Task::(_decref)
>>>>> >>> >>> > >>
Task=`994c7bc3-a236-4d03-a732-e068c7ed9ed4`::ref 0
>>>>> >>> >>> > >> aborting
>>>>> >>> >>> > >> False
>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24
>>>>> >>> >>> > >>
21:41:41,550::BindingXMLRPC::1132::vds::(wrapper) client
>>>>> >>> >>> > >> [10.10.10.2]::call
>>>>> >>> >>> > >> getCapabilities with () {}
>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24
>>>>> >>> >>> > >>
21:41:41,553::utils::738::root::(execCmd)
>>>>> >>> >>> > >> /sbin/ip route show to
0.0.0.0/0 table all (cwd None)
>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24
>>>>> >>> >>> > >>
21:41:41,560::utils::758::root::(execCmd)
>>>>> >>> >>> > >> SUCCESS: <err> =
''; <rc> = 0
>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24
>>>>> >>> >>> > >>
21:41:41,588::caps::728::root::(_getKeyPackages) rpm
>>>>> >>> >>> > >> package
>>>>> >>> >>> > >> ('gluster-swift',)
not found
>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24
>>>>> >>> >>> > >>
21:41:41,592::caps::728::root::(_getKeyPackages) rpm
>>>>> >>> >>> > >> package
>>>>> >>> >>> > >>
('gluster-swift-object',) not found
>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24
>>>>> >>> >>> > >>
21:41:41,593::caps::728::root::(_getKeyPackages) rpm
>>>>> >>> >>> > >> package
>>>>> >>> >>> > >>
('gluster-swift-plugin',) not found
>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24
>>>>> >>> >>> > >>
21:41:41,598::caps::728::root::(_getKeyPackages) rpm
>>>>> >>> >>> > >> package
>>>>> >>> >>> > >>
('gluster-swift-account',) not found
>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24
>>>>> >>> >>> > >>
21:41:41,598::caps::728::root::(_getKeyPackages) rpm
>>>>> >>> >>> > >> package
>>>>> >>> >>> > >>
('gluster-swift-proxy',) not found
>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24
>>>>> >>> >>> > >>
21:41:41,598::caps::728::root::(_getKeyPackages) rpm
>>>>> >>> >>> > >> package
>>>>> >>> >>> > >>
('gluster-swift-doc',) not found
>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24
>>>>> >>> >>> > >>
21:41:41,599::caps::728::root::(_getKeyPackages) rpm
>>>>> >>> >>> > >> package
>>>>> >>> >>> > >>
('gluster-swift-container',) not found
>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24
>>>>> >>> >>> > >>
21:41:41,599::caps::728::root::(_getKeyPackages) rpm
>>>>> >>> >>> > >> package
>>>>> >>> >>> > >>
('glusterfs-geo-replication',) not found
>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24
>>>>> >>> >>> > >>
21:41:41,600::caps::646::root::(get)
>>>>> >>> >>> > >> VirtioRNG DISABLED: libvirt
version 0.10.2-29.el6_5.9
>>>>> >>> >>> > >> required
>>>>> >>> >>> > >> >=
>>>>> >>> >>> > >> 0.10.2-31
>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24
>>>>> >>> >>> > >>
21:41:41,603::BindingXMLRPC::1139::vds::(wrapper) return
>>>>> >>> >>> > >> getCapabilities
>>>>> >>> >>> > >> with {'status':
{'message': 'Done', 'code': 0}, 'info':
>>>>> >>> >>> > >> {'HBAInventory':
>>>>> >>> >>> > >> {'iSCSI':
[{'InitiatorName':
>>>>> >>> >>> > >>
'iqn.1994-05.com.redhat:32151ce183c8'}],
>>>>> >>> >>> > >> 'FC':
>>>>> >>> >>> > >> []}, 'packages2':
{'kernel': {'release':
>>>>> >>> >>> > >> '431.el6.x86_64',
>>>>> >>> >>> > >> 'buildtime':
>>>>> >>> >>> > >> 1385061309.0,
'version': '2.6.32'}, 'glusterfs-rdma':
>>>>> >>> >>> > >> {'release':
>>>>> >>> >>> > >> '1.el6',
>>>>> >>> >>> > >> 'buildtime':
1403622628L, 'version': '3.5.1'},
>>>>> >>> >>> > >> 'glusterfs-fuse':
>>>>> >>> >>> > >> {'release':
'1.el6', 'buildtime': 1403622628L, 'version':
>>>>> >>> >>> > >> '3.5.1'},
>>>>> >>> >>> > >> 'spice-server':
{'release': '6.el6_5.2', 'buildtime':
>>>>> >>> >>> > >> 1402324637L,
>>>>> >>> >>> > >> 'version':
'0.12.4'}, 'vdsm': {'release':
>>>>> >>> >>> > >> '1.gitdb83943.el6',
>>>>> >>> >>> > >> 'buildtime':
>>>>> >>> >>> > >> 1412784567L,
'version': '4.16.7'}, 'qemu-kvm':
>>>>> >>> >>> > >> {'release':
>>>>> >>> >>> > >> '2.415.el6_5.10',
'buildtime': 1402435700L, 'version':
>>>>> >>> >>> > >> '0.12.1.2'},
>>>>> >>> >>> > >> 'qemu-img':
{'release': '2.415.el6_5.10', 'buildtime':
>>>>> >>> >>> > >> 1402435700L,
>>>>> >>> >>> > >> 'version':
'0.12.1.2'}, 'libvirt': {'release':
>>>>> >>> >>> > >> '29.el6_5.9',
>>>>> >>> >>> > >> 'buildtime':
>>>>> >>> >>> > >> 1402404612L,
'version': '0.10.2'}, 'glusterfs':
>>>>> >>> >>> > >> {'release':
>>>>> >>> >>> > >> '1.el6',
>>>>> >>> >>> > >> 'buildtime':
1403622628L, 'version': '3.5.1'}, 'mom':
>>>>> >>> >>> > >> {'release':
>>>>> >>> >>> > >> '2.el6',
>>>>> >>> >>> > >> 'buildtime':
1403794344L, 'version': '0.4.1'},
>>>>> >>> >>> > >> 'glusterfs-server':
>>>>> >>> >>> > >> {'release':
'1.el6', 'buildtime': 1403622628L, 'version':
>>>>> >>> >>> > >> '3.5.1'}},
>>>>> >>> >>> > >> 'numaNodeDistance':
{'1': [20, 10], '0': [10, 20]},
>>>>> >>> >>> > >> 'cpuModel':
>>>>> >>> >>> > >> 'Intel(R)
>>>>> >>> >>> > >> Xeon(R) CPU X5650
@ 2.67GHz', 'liveMerge':
>>>>> >>> >>> > >> 'false',
>>>>> >>> >>> > >> 'hooks':
>>>>> >>> >>> > >> {},
>>>>> >>> >>> > >> 'cpuSockets':
'2', 'vmTypes': ['kvm'], 'selinux':
>>>>> >>> >>> > >> {'mode':
'1'},
>>>>> >>> >>> > >> 'kdumpStatus': 0,
'supportedProtocols': ['2.2', '2.3'],
>>>>> >>> >>> > >> 'networks':
>>>>> >>> >>> > >> {'ovirtmgmt':
{'iface': u'bond0.10', 'addr':
>>>>> >>> >>> > >> '43.252.176.16',
>>>>> >>> >>> > >> 'bridged':
>>>>> >>> >>> > >> False, 'ipv6addrs':
['fe80::62eb:69ff:fe20:b46c/64'],
>>>>> >>> >>> > >> 'mtu':
>>>>> >>> >>> > >> '1500',
>>>>> >>> >>> > >> 'bootproto4':
'none', 'netmask': '255.255.255.0',
>>>>> >>> >>> > >> 'ipv4addrs':
>>>>> >>> >>> > >> ['
>>>>> >>> >>> > >> 43.252.176.16/24'
<
http://43.252.176.16/24%27>],
>>>>> >>> >>> > >> 'interface':
>>>>> >>> >>> > >> u'bond0.10',
'ipv6gateway': '::', 'gateway':
>>>>> >>> >>> > >> '43.25.17.1'},
>>>>> >>> >>> > >> 'Internal':
>>>>> >>> >>> > >> {'iface':
'Internal', 'addr': '', 'cfg': {'DEFROUTE':
>>>>> >>> >>> > >> 'no',
>>>>> >>> >>> > >> 'HOTPLUG':
>>>>> >>> >>> > >> 'no', 'MTU':
'9000', 'DELAY': '0', 'NM_CONTROLLED': 'no',
>>>>> >>> >>> > >> 'BOOTPROTO':
>>>>> >>> >>> > >> 'none',
'STP': 'off', 'DEVICE': 'Internal', 'TYPE':
>>>>> >>> >>> > >> 'Bridge',
>>>>> >>> >>> > >> 'ONBOOT':
>>>>> >>> >>> > >> 'no'},
'bridged': True, 'ipv6addrs':
>>>>> >>> >>> > >>
['fe80::210:18ff:fecd:daac/64'],
>>>>> >>> >>> > >> 'gateway':
'', 'bootproto4': 'none', 'netmask': '',
>>>>> >>> >>> > >> 'stp':
>>>>> >>> >>> > >> 'off',
>>>>> >>> >>> > >> 'ipv4addrs': [],
'mtu': '9000', 'ipv6gateway': '::',
>>>>> >>> >>> > >> 'ports':
>>>>> >>> >>> > >> ['bond1.100']},
'storage': {'iface': u'bond1', 'addr':
>>>>> >>> >>> > >> '10.10.10.6',
>>>>> >>> >>> > >> 'bridged': False,
'ipv6addrs':
>>>>> >>> >>> > >>
['fe80::210:18ff:fecd:daac/64'],
>>>>> >>> >>> > >> 'mtu':
>>>>> >>> >>> > >> '9000',
'bootproto4': 'none', 'netmask': '255.255.255.0',
>>>>> >>> >>> > >> 'ipv4addrs': ['
>>>>> >>> >>> > >> 10.10.10.6/24'
<
http://10.10.10.6/24%27>], 'interface':
>>>>> >>> >>> > >> u'bond1',
>>>>> >>> >>> > >> 'ipv6gateway':
'::', 'gateway': ''}, 'VMNetwork':
>>>>> >>> >>> > >> {'iface':
>>>>> >>> >>> > >> 'VMNetwork',
>>>>> >>> >>> > >> 'addr': '',
'cfg': {'DEFROUTE': 'no', 'HOTPLUG': 'no',
>>>>> >>> >>> > >> 'MTU':
>>>>> >>> >>> > >> '1500',
>>>>> >>> >>> > >> 'DELAY': '0',
'NM_CONTROLLED': 'no', 'BOOTPROTO': 'none',
>>>>> >>> >>> > >> 'STP':
>>>>> >>> >>> > >> 'off',
>>>>> >>> >>> > >> 'DEVICE':
'VMNetwork', 'TYPE': 'Bridge', 'ONBOOT': 'no'},
>>>>> >>> >>> > >> 'bridged':
>>>>> >>> >>> > >> True,
>>>>> >>> >>> > >> 'ipv6addrs':
['fe80::62eb:69ff:fe20:b46c/64'], 'gateway':
>>>>> >>> >>> > >> '',
>>>>> >>> >>> > >> 'bootproto4':
>>>>> >>> >>> > >> 'none',
'netmask': '', 'stp': 'off', 'ipv4addrs': [],
>>>>> >>> >>> > >> 'mtu':
>>>>> >>> >>> > >> '1500',
>>>>> >>> >>> > >> 'ipv6gateway':
'::', 'ports': ['bond0.36']}}, 'bridges':
>>>>> >>> >>> > >> {'Internal':
>>>>> >>> >>> > >> {'addr': '',
'cfg': {'DEFROUTE': 'no', 'HOTPLUG': 'no',
>>>>> >>> >>> > >> 'MTU':
>>>>> >>> >>> > >> '9000',
>>>>> >>> >>> > >> 'DELAY': '0',
'NM_CONTROLLED': 'no', 'BOOTPROTO': 'none',
>>>>> >>> >>> > >> 'STP':
>>>>> >>> >>> > >> 'off',
>>>>> >>> >>> > >> 'DEVICE':
'Internal', 'TYPE': 'Bridge', 'ONBOOT': 'no'},
>>>>> >>> >>> > >> 'ipv6addrs':
>>>>> >>> >>> > >>
['fe80::210:18ff:fecd:daac/64'], 'mtu': '9000',
>>>>> >>> >>> > >> 'netmask':
'',
>>>>> >>> >>> > >> 'stp':
>>>>> >>> >>> > >> 'off',
'ipv4addrs': [], 'ipv6gateway': '::', 'gateway':
>>>>> >>> >>> > >> '',
>>>>> >>> >>> > >> 'opts':
>>>>> >>> >>> > >>
{'topology_change_detected': '0',
>>>>> >>> >>> > >>
'multicast_last_member_count':
>>>>> >>> >>> > >> '2',
>>>>> >>> >>> > >> 'hash_elasticity':
'4',
>>>>> >>> >>> > >>
'multicast_query_response_interval':
>>>>> >>> >>> > >> '999',
>>>>> >>> >>> > >> 'multicast_snooping':
'1',
>>>>> >>> >>> > >>
'multicast_startup_query_interval':
>>>>> >>> >>> > >> '3124',
>>>>> >>> >>> > >> 'hello_timer':
'31', 'multicast_querier_interval':
>>>>> >>> >>> > >> '25496',
>>>>> >>> >>> > >> 'max_age':
>>>>> >>> >>> > >> '1999',
'hash_max': '512', 'stp_state': '0', 'root_id':
>>>>> >>> >>> > >> '8000.001018cddaac',
'priority': '32768',
>>>>> >>> >>> > >>
'multicast_membership_interval':
>>>>> >>> >>> > >> '25996',
'root_path_cost': '0', 'root_port': '0',
>>>>> >>> >>> > >> 'multicast_querier':
>>>>> >>> >>> > >> '0',
>>>>> >>> >>> > >>
'multicast_startup_query_count': '2', 'hello_time':
>>>>> >>> >>> > >> '199',
>>>>> >>> >>> > >> 'topology_change':
'0', 'bridge_id': '8000.001018cddaac',
>>>>> >>> >>> > >>
'topology_change_timer': '0', 'ageing_time': '29995',
>>>>> >>> >>> > >> 'gc_timer':
>>>>> >>> >>> > >> '31',
>>>>> >>> >>> > >> 'group_addr':
'1:80:c2:0:0:0', 'tcn_timer': '0',
>>>>> >>> >>> > >>
'multicast_query_interval': '12498',
>>>>> >>> >>> > >>
'multicast_last_member_interval':
>>>>> >>> >>> > >> '99',
'multicast_router': '1', 'forward_delay': '0'},
>>>>> >>> >>> > >> 'ports':
>>>>> >>> >>> > >> ['bond1.100']},
'VMNetwork': {'addr': '', 'cfg':
>>>>> >>> >>> > >> {'DEFROUTE':
>>>>> >>> >>> > >> 'no',
>>>>> >>> >>> > >> 'HOTPLUG':
'no', 'MTU': '1500', 'DELAY': '0',
>>>>> >>> >>> > >> 'NM_CONTROLLED':
>>>>> >>> >>> > >> 'no',
>>>>> >>> >>> > >> 'BOOTPROTO':
'none', 'STP': 'off', 'DEVICE': 'VMNetwork',
>>>>> >>> >>> > >> 'TYPE':
>>>>> >>> >>> > >> 'Bridge',
>>>>> >>> >>> > >> 'ONBOOT':
'no'}, 'ipv6addrs':
>>>>> >>> >>> > >>
['fe80::62eb:69ff:fe20:b46c/64'],
>>>>> >>> >>> > >> 'mtu':
>>>>> >>> >>> > >> '1500',
'netmask': '', 'stp': 'off', 'ipv4addrs': [],
>>>>> >>> >>> > >> 'ipv6gateway':
>>>>> >>> >>> > >> '::',
>>>>> >>> >>> > >> 'gateway':
'', 'opts': {'topology_change_detected': '0',
>>>>> >>> >>> > >>
'multicast_last_member_count': '2', 'hash_elasticity':
>>>>> >>> >>> > >> '4',
>>>>> >>> >>> > >>
'multicast_query_response_interval': '999',
>>>>> >>> >>> > >>
'multicast_snooping':
>>>>> >>> >>> > >> '1',
>>>>> >>> >>> > >>
'multicast_startup_query_interval': '3124',
>>>>> >>> >>> > >> 'hello_timer':
>>>>> >>> >>> > >> '131',
>>>>> >>> >>> > >>
'multicast_querier_interval': '25496', 'max_age': '1999',
>>>>> >>> >>> > >> 'hash_max':
>>>>> >>> >>> > >> '512',
'stp_state': '0', 'root_id': '8000.60eb6920b46c',
>>>>> >>> >>> > >> 'priority':
>>>>> >>> >>> > >> '32768',
'multicast_membership_interval': '25996',
>>>>> >>> >>> > >> 'root_path_cost':
>>>>> >>> >>> > >> '0',
>>>>> >>> >>> > >> 'root_port':
'0', 'multicast_querier': '0',
>>>>> >>> >>> > >>
'multicast_startup_query_count': '2', 'hello_time':
>>>>> >>> >>> > >> '199',
>>>>> >>> >>> > >> 'topology_change':
'0', 'bridge_id': '8000.60eb6920b46c',
>>>>> >>> >>> > >>
'topology_change_timer': '0', 'ageing_time': '29995',
>>>>> >>> >>> > >> 'gc_timer':
>>>>> >>> >>> > >> '31',
>>>>> >>> >>> > >> 'group_addr':
'1:80:c2:0:0:0', 'tcn_timer': '0',
>>>>> >>> >>> > >>
'multicast_query_interval': '12498',
>>>>> >>> >>> > >>
'multicast_last_member_interval':
>>>>> >>> >>> > >> '99',
'multicast_router': '1', 'forward_delay': '0'},
>>>>> >>> >>> > >> 'ports':
>>>>> >>> >>> > >> ['bond0.36']}},
'uuid':
>>>>> >>> >>> > >>
'44454C4C-4C00-1057-8053-B7C04F504E31',
>>>>> >>> >>> > >> 'lastClientIface':
'bond1', 'nics': {'eth3':
>>>>> >>> >>> > >> {'permhwaddr':
>>>>> >>> >>> > >> '00:10:18:cd:da:ae',
'addr': '', 'cfg': {'SLAVE': 'yes',
>>>>> >>> >>> > >> 'NM_CONTROLLED':
>>>>> >>> >>> > >> 'no', 'MTU':
'9000', 'HWADDR': '00:10:18:cd:da:ae',
>>>>> >>> >>> > >> 'MASTER':
>>>>> >>> >>> > >> 'bond1',
>>>>> >>> >>> > >> 'DEVICE':
'eth3', 'ONBOOT': 'no'}, 'ipv6addrs': [],
>>>>> >>> >>> > >> 'mtu':
>>>>> >>> >>> > >> '9000',
>>>>> >>> >>> > >> 'netmask':
'', 'ipv4addrs': [], 'hwaddr':
>>>>> >>> >>> > >> '00:10:18:cd:da:ac',
>>>>> >>> >>> > >> 'speed':
>>>>> >>> >>> > >> 1000}, 'eth2':
{'permhwaddr': '00:10:18:cd:da:ac',
>>>>> >>> >>> > >> 'addr': '',
>>>>> >>> >>> > >> 'cfg':
>>>>> >>> >>> > >> {'SLAVE':
'yes', 'NM_CONTROLLED': 'no', 'MTU': '9000',
>>>>> >>> >>> > >> 'HWADDR':
>>>>> >>> >>> > >> '00:10:18:cd:da:ac',
'MASTER': 'bond1', 'DEVICE': 'eth2',
>>>>> >>> >>> > >> 'ONBOOT':
>>>>> >>> >>> > >> 'no'},
>>>>> >>> >>> > >> 'ipv6addrs': [],
'mtu': '9000', 'netmask': '',
>>>>> >>> >>> > >> 'ipv4addrs': [],
>>>>> >>> >>> > >> 'hwaddr':
>>>>> >>> >>> > >> '00:10:18:cd:da:ac',
'speed': 1000}, 'eth1':
>>>>> >>> >>> > >> {'permhwaddr':
>>>>> >>> >>> > >> '60:eb:69:20:b4:6d',
'addr': '', 'cfg': {'SLAVE': 'yes',
>>>>> >>> >>> > >> 'NM_CONTROLLED':
>>>>> >>> >>> > >> 'no', 'MTU':
'1500', 'HWADDR': '60:eb:69:20:b4:6d',
>>>>> >>> >>> > >> 'MASTER':
>>>>> >>> >>> > >> 'bond0',
>>>>> >>> >>> > >> 'DEVICE':
'eth1', 'ONBOOT': 'yes'}, 'ipv6addrs': [],
>>>>> >>> >>> > >> 'mtu':
>>>>> >>> >>> > >> '1500',
>>>>> >>> >>> > >> 'netmask':
'', 'ipv4addrs': [], 'hwaddr':
>>>>> >>> >>> > >> '60:eb:69:20:b4:6c',
>>>>> >>> >>> > >> 'speed':
>>>>> >>> >>> > >> 1000}, 'eth0':
{'permhwaddr': '60:eb:69:20:b4:6c',
>>>>> >>> >>> > >> 'addr': '',
>>>>> >>> >>> > >> 'cfg':
>>>>> >>> >>> > >> {'SLAVE':
'yes', 'NM_CONTROLLED': 'no', 'MTU': '1500',
>>>>> >>> >>> > >> 'HWADDR':
>>>>> >>> >>> > >> '60:eb:69:20:b4:6c',
'MASTER': 'bond0', 'DEVICE': 'eth0',
>>>>> >>> >>> > >> 'ONBOOT':
>>>>> >>> >>> > >> 'yes'},
>>>>> >>> >>> > >> 'ipv6addrs': [],
'mtu': '1500', 'netmask': '',
>>>>> >>> >>> > >> 'ipv4addrs': [],
>>>>> >>> >>> > >> 'hwaddr':
>>>>> >>> >>> > >> '60:eb:69:20:b4:6c',
'speed': 1000}},
>>>>> >>> >>> > >> 'software_revision':
'1',
>>>>> >>> >>> > >> 'clusterLevels':
['3.0', '3.1', '3.2', '3.3', '3.4',
>>>>> >>> >>> > >> '3.5'],
>>>>> >>> >>> > >> 'cpuFlags':
>>>>> >>> >>> > >>
>>>>> >>> >>> > >>
>>>>> >>> >>> > >>
>>>>> >>> >>> > >>
u'fpu,vme,de,pse,tsc,msr,pae,mce,cx8,apic,sep,mtrr,pge,mca,cmov,pat,pse36,clflush,dts,acpi,mmx,fxsr,sse,sse2,ss,ht,tm,pbe,syscall,nx,pdpe1gb,rdtscp,lm,constant_tsc,arch_perfmon,pebs,bts,rep_good,xtopology,nonstop_tsc,pni,pclmulqdq,dtes64,monitor,ds_cpl,vmx,smx,est,tm2,ssse3,cx16,xtpr,pdcm,pcid,dca,sse4_1,sse4_2,popcnt,aes,lahf_lm,tpr_shadow,vnmi,flexpriority,ept,vpid,model_Nehalem,model_Conroe,model_coreduo,model_core2duo,model_Penryn,model_Westmere,model_n270',
>>>>> >>> >>> > >>
'ISCSIInitiatorName':
>>>>> >>> >>> > >>
'iqn.1994-05.com.redhat:32151ce183c8',
>>>>> >>> >>> > >> 'netConfigDirty':
'False', 'supportedENGINEs': ['3.0',
>>>>> >>> >>> > >> '3.1',
>>>>> >>> >>> > >> '3.2',
>>>>> >>> >>> > >> '3.3',
>>>>> >>> >>> > >> '3.4',
'3.5'], 'autoNumaBalancing': 2, 'reservedMem':
>>>>> >>> >>> > >> '321',
>>>>> >>> >>> > >> 'bondings':
>>>>> >>> >>> > >> {'bond4':
{'addr': '', 'cfg': {}, 'mtu': '1500',
>>>>> >>> >>> > >> 'netmask':
'',
>>>>> >>> >>> > >> 'slaves':
>>>>> >>> >>> > >> [], 'hwaddr':
'00:00:00:00:00:00'}, 'bond0': {'addr': '',
>>>>> >>> >>> > >> 'cfg':
>>>>> >>> >>> > >> {'HOTPLUG':
'no', 'MTU': '1500', 'NM_CONTROLLED': 'no',
>>>>> >>> >>> > >> 'BONDING_OPTS':
>>>>> >>> >>> > >> 'mode=4 miimon=100',
'DEVICE': 'bond0', 'ONBOOT': 'yes'},
>>>>> >>> >>> > >> 'ipv6addrs':
>>>>> >>> >>> > >>
['fe80::62eb:69ff:fe20:b46c/64'], 'mtu': '1500',
>>>>> >>> >>> > >> 'netmask':
'',
>>>>> >>> >>> > >> 'ipv4addrs': [],
'hwaddr': '60:eb:69:20:b4:6c', 'slaves':
>>>>> >>> >>> > >> ['eth0',
>>>>> >>> >>> > >> 'eth1'],
>>>>> >>> >>> > >> 'opts':
{'miimon': '100', 'mode': '4'}}, 'bond1':
>>>>> >>> >>> > >> {'addr':
>>>>> >>> >>> > >> '10.10.10.6',
>>>>> >>> >>> > >> 'cfg':
{'DEFROUTE': 'no', 'IPADDR': '10.10.10.6',
>>>>> >>> >>> > >> 'HOTPLUG':
>>>>> >>> >>> > >> 'no',
>>>>> >>> >>> > >> 'MTU':
>>>>> >>> >>> > >> '9000',
'NM_CONTROLLED': 'no', 'NETMASK':
>>>>> >>> >>> > >> '255.255.255.0',
>>>>> >>> >>> > >> 'BOOTPROTO':
>>>>> >>> >>> > >> 'none',
'BONDING_OPTS': 'mode=4 miimon=100', 'DEVICE':
>>>>> >>> >>> > >> 'bond1',
>>>>> >>> >>> > >> 'ONBOOT':
>>>>> >>> >>> > >> 'no'},
'ipv6addrs': ['fe80::210:18ff:fecd:daac/64'],
>>>>> >>> >>> > >> 'mtu':
>>>>> >>> >>> > >> '9000',
>>>>> >>> >>> > >> 'netmask':
'255.255.255.0', 'ipv4addrs': ['10.10.10.6/24'
>>>>> >>> >>> > >>
<
http://10.10.10.6/24%27>], 'hwaddr':
>>>>> >>> >>> > >> '00:10:18:cd:da:ac',
>>>>> >>> >>> > >> 'slaves':
>>>>> >>> >>> > >> ['eth2',
'eth3'], 'opts': {'miimon': '100', 'mode':
>>>>> >>> >>> > >> '4'}},
>>>>> >>> >>> > >> 'bond2':
>>>>> >>> >>> > >> {'addr': '',
'cfg': {}, 'mtu': '1500', 'netmask': '',
>>>>> >>> >>> > >> 'slaves':
>>>>> >>> >>> > >> [],
>>>>> >>> >>> > >> 'hwaddr':
'00:00:00:00:00:00'}, 'bond3': {'addr': '',
>>>>> >>> >>> > >> 'cfg': {},
>>>>> >>> >>> > >> 'mtu':
>>>>> >>> >>> > >> '1500',
'netmask': '', 'slaves': [], 'hwaddr':
>>>>> >>> >>> > >>
'00:00:00:00:00:00'}},
>>>>> >>> >>> > >> 'software_version':
'4.16', 'memSize': '24019',
>>>>> >>> >>> > >> 'cpuSpeed':
>>>>> >>> >>> > >> '2667.000',
>>>>> >>> >>> > >> 'numaNodes':
{u'1': {'totalMemory': '12288', 'cpus': [6,
>>>>> >>> >>> > >> 7, 8,
>>>>> >>> >>> > >> 9,
>>>>> >>> >>> > >> 10, 11,
>>>>> >>> >>> > >> 18, 19, 20, 21, 22, 23]},
u'0': {'totalMemory': '12278',
>>>>> >>> >>> > >> 'cpus':
>>>>> >>> >>> > >> [0,
>>>>> >>> >>> > >> 1, 2,
>>>>> >>> >>> > >> 3, 4, 5, 12, 13, 14, 15, 16,
17]}}, 'version_name': 'Snow
>>>>> >>> >>> > >> Man',
>>>>> >>> >>> > >> 'vlans':
>>>>> >>> >>> > >> {'bond0.10':
{'iface': 'bond0', 'addr': '43.25.17.16',
>>>>> >>> >>> > >> 'cfg':
>>>>> >>> >>> > >> {'DEFROUTE':
>>>>> >>> >>> > >> 'yes',
'VLAN': 'yes', 'IPADDR': '43.25.17.16',
'HOTPLUG':
>>>>> >>> >>> > >> 'no',
>>>>> >>> >>> > >> 'GATEWAY':
>>>>> >>> >>> > >> '43.25.17.1',
'NM_CONTROLLED': 'no', 'NETMASK':
>>>>> >>> >>> > >> '255.255.255.0',
>>>>> >>> >>> > >> 'BOOTPROTO':
'none', 'DEVICE': 'bond0.10', 'MTU': '1500',
>>>>> >>> >>> > >> 'ONBOOT':
>>>>> >>> >>> > >> 'yes'},
>>>>> >>> >>> > >> 'ipv6addrs':
['fe80::62eb:69ff:fe20:b46c/64'], 'vlanid':
>>>>> >>> >>> > >> 10,
>>>>> >>> >>> > >> 'mtu':
>>>>> >>> >>> > >> '1500',
>>>>> >>> >>> > >> 'netmask':
'255.255.255.0', 'ipv4addrs':
>>>>> >>> >>> > >> ['43.25.17.16/24']
>>>>> >>> >>> > >>
<
http://43.25.17.16/24%27%5D>}, 'bond0.36': {'iface':
>>>>> >>> >>> > >> 'bond0',
>>>>> >>> >>> > >> 'addr':
>>>>> >>> >>> > >> '', 'cfg':
{'BRIDGE': 'VMNetwork', 'VLAN': 'yes',
>>>>> >>> >>> > >> 'HOTPLUG':
>>>>> >>> >>> > >> 'no',
>>>>> >>> >>> > >> 'MTU':
>>>>> >>> >>> > >> '1500',
'NM_CONTROLLED': 'no', 'DEVICE': 'bond0.36',
>>>>> >>> >>> > >> 'ONBOOT':
>>>>> >>> >>> > >> 'no'},
>>>>> >>> >>> > >> 'ipv6addrs':
['fe80::62eb:69ff:fe20:b46c/64'], 'vlanid':
>>>>> >>> >>> > >> 36,
>>>>> >>> >>> > >> 'mtu':
>>>>> >>> >>> > >> '1500',
>>>>> >>> >>> > >> 'netmask':
'', 'ipv4addrs': []}, 'bond1.100': {'iface':
>>>>> >>> >>> > >> 'bond1',
>>>>> >>> >>> > >> 'addr':
>>>>> >>> >>> > >> '', 'cfg':
{'BRIDGE': 'Internal', 'VLAN': 'yes',
>>>>> >>> >>> > >> 'HOTPLUG':
>>>>> >>> >>> > >> 'no',
>>>>> >>> >>> > >> 'MTU':
>>>>> >>> >>> > >> '9000',
'NM_CONTROLLED': 'no', 'DEVICE': 'bond1.100',
>>>>> >>> >>> > >> 'ONBOOT':
>>>>> >>> >>> > >> 'no'},
>>>>> >>> >>> > >> 'ipv6addrs':
['fe80::210:18ff:fecd:daac/64'], 'vlanid':
>>>>> >>> >>> > >> 100,
>>>>> >>> >>> > >> 'mtu':
>>>>> >>> >>> > >> '9000',
>>>>> >>> >>> > >> 'netmask':
'', 'ipv4addrs': []}}, 'cpuCores': '12',
>>>>> >>> >>> > >> 'kvmEnabled':
>>>>> >>> >>> > >> 'true',
>>>>> >>> >>> > >> 'guestOverhead':
'65', 'cpuThreads': '24',
>>>>> >>> >>> > >> 'emulatedMachines':
>>>>> >>> >>> > >> [u'rhel6.5.0',
u'pc', u'rhel6.4.0', u'rhel6.3.0',
>>>>> >>> >>> > >> u'rhel6.2.0',
>>>>> >>> >>> > >> u'rhel6.1.0',
u'rhel6.0.0', u'rhel5.5.0', u'rhel5.4.4',
>>>>> >>> >>> > >> u'rhel5.4.0'],
>>>>> >>> >>> > >> 'operatingSystem':
{'release': '5.el6.centos.11.1',
>>>>> >>> >>> > >> 'version':
>>>>> >>> >>> > >> '6',
>>>>> >>> >>> > >> 'name':
>>>>> >>> >>> > >> 'RHEL'},
'lastClient': '10.10.10.2'}}
>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24
>>>>> >>> >>> > >>
21:41:41,620::BindingXMLRPC::1132::vds::(wrapper) client
>>>>> >>> >>> > >> [10.10.10.2]::call
>>>>> >>> >>> > >> getHardwareInfo with () {}
>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24
>>>>> >>> >>> > >>
21:41:41,621::BindingXMLRPC::1139::vds::(wrapper) return
>>>>> >>> >>> > >> getHardwareInfo
>>>>> >>> >>> > >> with {'status':
{'message': 'Done', 'code': 0}, 'info':
>>>>> >>> >>> > >> {'systemProductName':
'CS24-TY', 'systemSerialNumber':
>>>>> >>> >>> > >> '7LWSPN1',
>>>>> >>> >>> > >> 'systemFamily':
'Server', 'systemVersion': 'A00',
>>>>> >>> >>> > >> 'systemUUID':
>>>>> >>> >>> > >>
'44454c4c-4c00-1057-8053-b7c04f504e31',
>>>>> >>> >>> > >>
'systemManufacturer':
>>>>> >>> >>> > >> 'Dell'}}
>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24
>>>>> >>> >>> > >>
21:41:41,733::BindingXMLRPC::1132::vds::(wrapper) client
>>>>> >>> >>> > >> [10.10.10.2]::call
>>>>> >>> >>> > >> hostsList with () {} flowID
[222e8036]
>>>>> >>> >>> > >> Thread-13::ERROR::2014-11-24
>>>>> >>> >>> > >>
21:41:44,753::BindingXMLRPC::1148::vds::(wrapper) vdsm
>>>>> >>> >>> > >> exception
>>>>> >>> >>> > >> occured
>>>>> >>> >>> > >> Traceback (most recent call
last):
>>>>> >>> >>> > >> File
"/usr/share/vdsm/rpc/BindingXMLRPC.py", line 1135,
>>>>> >>> >>> > >> in
>>>>> >>> >>> > >> wrapper
>>>>> >>> >>> > >> res = f(*args, **kwargs)
>>>>> >>> >>> > >> File
"/usr/share/vdsm/gluster/api.py", line 54, in
>>>>> >>> >>> > >> wrapper
>>>>> >>> >>> > >> rv = func(*args,
**kwargs)
>>>>> >>> >>> > >> File
"/usr/share/vdsm/gluster/api.py", line 251, in
>>>>> >>> >>> > >> hostsList
>>>>> >>> >>> > >> return {'hosts':
self.svdsmProxy.glusterPeerStatus()}
>>>>> >>> >>> > >> File
"/usr/share/vdsm/supervdsm.py", line 50, in
>>>>> >>> >>> > >> __call__
>>>>> >>> >>> > >> return callMethod()
>>>>> >>> >>> > >> File
"/usr/share/vdsm/supervdsm.py", line 48, in
>>>>> >>> >>> > >> <lambda>
>>>>> >>> >>> > >> **kwargs)
>>>>> >>> >>> > >> File
"<string>", line 2, in glusterPeerStatus
>>>>> >>> >>> > >> File
>>>>> >>> >>> > >>
"/usr/lib64/python2.6/multiprocessing/managers.py", line
>>>>> >>> >>> > >> 740,
>>>>> >>> >>> > >> in
>>>>> >>> >>> > >> _callmethod
>>>>> >>> >>> > >> raise
convert_to_error(kind, result)
>>>>> >>> >>> > >>
GlusterCmdExecFailedException: Command execution failed
>>>>> >>> >>> > >> error: Connection failed.
Please check if gluster daemon
>>>>> >>> >>> > >> is
>>>>> >>> >>> > >> operational.
>>>>> >>> >>> > >> return code: 1
>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24
>>>>> >>> >>> > >>
>>>>> >>> >>> > >>
>>>>> >>> >>> > >>
21:41:50,949::task::595::Storage.TaskManager.Task::(_updateState)
>>>>> >>> >>> > >>
Task=`c9042986-c978-4b08-adb2-616f5299e115`::moving from
>>>>> >>> >>> > >> state
>>>>> >>> >>> > >> init
>>>>> >>> >>> > >> ->
>>>>> >>> >>> > >> state preparing
>>>>> >>> >>> > >> Thread-13::INFO::2014-11-24
>>>>> >>> >>> > >>
21:41:50,950::logUtils::44::dispatcher::(wrapper) Run and
>>>>> >>> >>> > >> protect:
>>>>> >>> >>> > >> repoStats(options=None)
>>>>> >>> >>> > >> Thread-13::INFO::2014-11-24
>>>>> >>> >>> > >>
21:41:50,950::logUtils::47::dispatcher::(wrapper) Run and
>>>>> >>> >>> > >> protect:
>>>>> >>> >>> > >> repoStats, Return response:
{}
>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24
>>>>> >>> >>> > >>
>>>>> >>> >>> > >>
21:41:50,950::task::1191::Storage.TaskManager.Task::(prepare)
>>>>> >>> >>> > >>
Task=`c9042986-c978-4b08-adb2-616f5299e115`::finished: {}
>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24
>>>>> >>> >>> > >>
>>>>> >>> >>> > >>
>>>>> >>> >>> > >>
21:41:50,950::task::595::Storage.TaskManager.Task::(_updateState)
>>>>> >>> >>> > >>
Task=`c9042986-c978-4b08-adb2-616f5299e115`::moving from
>>>>> >>> >>> > >> state
>>>>> >>> >>> > >> preparing
>>>>> >>> >>> > >> ->
>>>>> >>> >>> > >> state finished
>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24
>>>>> >>> >>> > >>
>>>>> >>> >>> > >>
>>>>> >>> >>> > >>
>>>>> >>> >>> > >>
21:41:50,951::resourceManager::940::Storage.ResourceManager.Owner::(releaseAll)
>>>>> >>> >>> > >> Owner.releaseAll requests {}
resources {}
>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24
>>>>> >>> >>> > >>
>>>>> >>> >>> > >>
>>>>> >>> >>> > >>
>>>>> >>> >>> > >>
21:41:50,951::resourceManager::977::Storage.ResourceManager.Owner::(cancelAll)
>>>>> >>> >>> > >> Owner.cancelAll requests {}
>>>>> >>> >>> > >> Thread-13::DEBUG::2014-11-24
>>>>> >>> >>> > >>
>>>>> >>> >>> > >>
21:41:50,951::task::993::Storage.TaskManager.Task::(_decref)
>>>>> >>> >>> > >>
Task=`c9042986-c978-4b08-adb2-616f5299e115`::ref 0
>>>>> >>> >>> > >> aborting
>>>>> >>> >>> > >> False
>>>>> >>> >>> > >>
-------------------------------
>>>>> >>> >>> > >>
>>>>> >>> >>> > >> [root@compute4 ~]# service
glusterd status
>>>>> >>> >>> > >> glusterd is stopped
>>>>> >>> >>> > >> [root@compute4 ~]# chkconfig
--list | grep glusterd
>>>>> >>> >>> > >> glusterd 0:off 1:off
2:on 3:on 4:on
>>>>> >>> >>> > >> 5:on
>>>>> >>> >>> > >> 6:off
>>>>> >>> >>> > >> [root@compute4 ~]#
>>>>> >>> >>> > >>
>>>>> >>> >>> > >> Thanks,
>>>>> >>> >>> > >> Punit
>>>>> >>> >>> > >>
>>>>> >>> >>> > >> On Mon, Nov 24, 2014 at 6:36
PM, Kanagaraj
>>>>> >>> >>> > >> <kmayilsa(a)redhat.com>
>>>>> >>> >>> > >> wrote:
>>>>> >>> >>> > >>
>>>>> >>> >>> > >>> Can you send the
corresponding error in vdsm.log from
>>>>> >>> >>> > >>> the
>>>>> >>> >>> > >>> host?
>>>>> >>> >>> > >>>
>>>>> >>> >>> > >>> Also check if glusterd
service is running.
>>>>> >>> >>> > >>>
>>>>> >>> >>> > >>> Thanks,
>>>>> >>> >>> > >>> Kanagaraj
>>>>> >>> >>> > >>>
>>>>> >>> >>> > >>>
>>>>> >>> >>> > >>> On 11/24/2014 03:39 PM,
Punit Dambiwal wrote:
>>>>> >>> >>> > >>>
>>>>> >>> >>> > >>> Hi,
>>>>> >>> >>> > >>>
>>>>> >>> >>> > >>> After reboot my
Hypervisior host can not activate again
>>>>> >>> >>> > >>> in the
>>>>> >>> >>> > >>> cluster
>>>>> >>> >>> > >>> and failed with the
following error :-
>>>>> >>> >>> > >>>
>>>>> >>> >>> > >>> Gluster command
[<UNKNOWN>] failed on server...
>>>>> >>> >>> > >>>
>>>>> >>> >>> > >>> Engine logs :-
>>>>> >>> >>> > >>>
>>>>> >>> >>> > >>> 2014-11-24 18:05:28,397
INFO
>>>>> >>> >>> > >>>
>>>>> >>> >>> > >>>
>>>>> >>> >>> > >>>
>>>>> >>> >>> > >>>
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand]
>>>>> >>> >>> > >>>
(DefaultQuartzScheduler_Worker-64) START,
>>>>> >>> >>> > >>>
GlusterVolumesListVDSCommand(HostName = Compute4, HostId
>>>>> >>> >>> > >>> =
>>>>> >>> >>> > >>>
33648a90-200c-45ca-89d5-1ce305d79a6a), log id: 5f251c90
>>>>> >>> >>> > >>> 2014-11-24 18:05:30,609
INFO
>>>>> >>> >>> > >>>
>>>>> >>> >>> > >>>
>>>>> >>> >>> > >>>
>>>>> >>> >>> > >>>
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand]
>>>>> >>> >>> > >>>
(DefaultQuartzScheduler_Worker-64) FINISH,
>>>>> >>> >>> > >>>
GlusterVolumesListVDSCommand,
>>>>> >>> >>> > >>> return:
>>>>> >>> >>> > >>>
>>>>> >>> >>> > >>>
>>>>> >>> >>> > >>>
>>>>> >>> >>> > >>>
{26ae1672-ee09-4a38-8fd2-72dd9974cc2b=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@d95203e0},
>>>>> >>> >>> > >>> log id: 5f251c90
>>>>> >>> >>> > >>> 2014-11-24 18:05:33,768
INFO
>>>>> >>> >>> > >>>
[org.ovirt.engine.core.bll.ActivateVdsCommand]
>>>>> >>> >>> > >>> (ajp--127.0.0.1-8702-8)
>>>>> >>> >>> > >>> [287d570d] Lock Acquired
to object EngineLock
>>>>> >>> >>> > >>> [exclusiveLocks=
>>>>> >>> >>> > >>> key:
>>>>> >>> >>> > >>>
0bf6b00f-7947-4411-b55a-cc5eea2b381a value: VDS
>>>>> >>> >>> > >>> , sharedLocks= ]
>>>>> >>> >>> > >>> 2014-11-24 18:05:33,795
INFO
>>>>> >>> >>> > >>>
[org.ovirt.engine.core.bll.ActivateVdsCommand]
>>>>> >>> >>> > >>>
(org.ovirt.thread.pool-8-thread-45) [287d570d] Running
>>>>> >>> >>> > >>> command:
>>>>> >>> >>> > >>> ActivateVdsCommand
internal: false. Entities affected :
>>>>> >>> >>> > >>> ID:
>>>>> >>> >>> > >>>
0bf6b00f-7947-4411-b55a-cc5eea2b381a Type: VDSAction
>>>>> >>> >>> > >>> group
>>>>> >>> >>> > >>> MANIPULATE_HOST
>>>>> >>> >>> > >>> with role type ADMIN
>>>>> >>> >>> > >>> 2014-11-24 18:05:33,796
INFO
>>>>> >>> >>> > >>>
[org.ovirt.engine.core.bll.ActivateVdsCommand]
>>>>> >>> >>> > >>>
(org.ovirt.thread.pool-8-thread-45) [287d570d] Before
>>>>> >>> >>> > >>> acquiring
>>>>> >>> >>> > >>> lock in
>>>>> >>> >>> > >>> order to prevent
monitoring for host Compute5 from
>>>>> >>> >>> > >>> data-center
>>>>> >>> >>> > >>> SV_WTC
>>>>> >>> >>> > >>> 2014-11-24 18:05:33,797
INFO
>>>>> >>> >>> > >>>
[org.ovirt.engine.core.bll.ActivateVdsCommand]
>>>>> >>> >>> > >>>
(org.ovirt.thread.pool-8-thread-45) [287d570d] Lock
>>>>> >>> >>> > >>> acquired,
>>>>> >>> >>> > >>> from
>>>>> >>> >>> > >>> now a
>>>>> >>> >>> > >>> monitoring of host will
be skipped for host Compute5
>>>>> >>> >>> > >>> from
>>>>> >>> >>> > >>> data-center
>>>>> >>> >>> > >>> SV_WTC
>>>>> >>> >>> > >>> 2014-11-24 18:05:33,817
INFO
>>>>> >>> >>> > >>>
>>>>> >>> >>> > >>>
[org.ovirt.engine.core.vdsbroker.SetVdsStatusVDSCommand]
>>>>> >>> >>> > >>>
(org.ovirt.thread.pool-8-thread-45) [287d570d] START,
>>>>> >>> >>> > >>>
SetVdsStatusVDSCommand(HostName = Compute5, HostId =
>>>>> >>> >>> > >>>
0bf6b00f-7947-4411-b55a-cc5eea2b381a, status=Unassigned,
>>>>> >>> >>> > >>>
nonOperationalReason=NONE, stopSpmFailureLogged=false),
>>>>> >>> >>> > >>> log id:
>>>>> >>> >>> > >>> 1cbc7311
>>>>> >>> >>> > >>> 2014-11-24 18:05:33,820
INFO
>>>>> >>> >>> > >>>
>>>>> >>> >>> > >>>
[org.ovirt.engine.core.vdsbroker.SetVdsStatusVDSCommand]
>>>>> >>> >>> > >>>
(org.ovirt.thread.pool-8-thread-45) [287d570d] FINISH,
>>>>> >>> >>> > >>> SetVdsStatusVDSCommand,
log id: 1cbc7311
>>>>> >>> >>> > >>> 2014-11-24 18:05:34,086
INFO
>>>>> >>> >>> > >>>
[org.ovirt.engine.core.bll.ActivateVdsCommand]
>>>>> >>> >>> > >>>
(org.ovirt.thread.pool-8-thread-45) Activate finished.
>>>>> >>> >>> > >>> Lock
>>>>> >>> >>> > >>> released.
>>>>> >>> >>> > >>> Monitoring can run now
for host Compute5 from
>>>>> >>> >>> > >>> data-center
>>>>> >>> >>> > >>> SV_WTC
>>>>> >>> >>> > >>> 2014-11-24 18:05:34,088
INFO
>>>>> >>> >>> > >>>
>>>>> >>> >>> > >>>
>>>>> >>> >>> > >>>
>>>>> >>> >>> > >>>
[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
>>>>> >>> >>> > >>>
(org.ovirt.thread.pool-8-thread-45) Correlation ID:
>>>>> >>> >>> > >>> 287d570d,
>>>>> >>> >>> > >>> Job
>>>>> >>> >>> > >>> ID:
>>>>> >>> >>> > >>>
5ef8e4d6-b2bc-469e-8e81-7ef74b2a001a, Call Stack: null,
>>>>> >>> >>> > >>> Custom
>>>>> >>> >>> > >>> Event ID:
>>>>> >>> >>> > >>> -1, Message: Host
Compute5 was activated by admin.
>>>>> >>> >>> > >>> 2014-11-24 18:05:34,090
INFO
>>>>> >>> >>> > >>>
[org.ovirt.engine.core.bll.ActivateVdsCommand]
>>>>> >>> >>> > >>>
(org.ovirt.thread.pool-8-thread-45) Lock freed to object
>>>>> >>> >>> > >>> EngineLock
>>>>> >>> >>> > >>> [exclusiveLocks= key:
>>>>> >>> >>> > >>>
0bf6b00f-7947-4411-b55a-cc5eea2b381a
>>>>> >>> >>> > >>> value:
>>>>> >>> >>> > >>> VDS
>>>>> >>> >>> > >>> , sharedLocks= ]
>>>>> >>> >>> > >>> 2014-11-24 18:05:35,792
INFO
>>>>> >>> >>> > >>>
>>>>> >>> >>> > >>>
>>>>> >>> >>> > >>>
>>>>> >>> >>> > >>>
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand]
>>>>> >>> >>> > >>>
(DefaultQuartzScheduler_Worker-55) [3706e836] START,
>>>>> >>> >>> > >>>
GlusterVolumesListVDSCommand(HostName = Compute4, HostId
>>>>> >>> >>> > >>> =
>>>>> >>> >>> > >>>
33648a90-200c-45ca-89d5-1ce305d79a6a), log id: 48a0c832
>>>>> >>> >>> > >>> 2014-11-24 18:05:37,064
INFO
>>>>> >>> >>> > >>>
>>>>> >>> >>> > >>>
>>>>> >>> >>> > >>>
>>>>> >>> >>> > >>>
[org.ovirt.engine.core.vdsbroker.vdsbroker.GetHardwareInfoVDSCommand]
>>>>> >>> >>> > >>>
(DefaultQuartzScheduler_Worker-69) START,
>>>>> >>> >>> > >>>
GetHardwareInfoVDSCommand(HostName = Compute5, HostId =
>>>>> >>> >>> > >>>
0bf6b00f-7947-4411-b55a-cc5eea2b381a,
>>>>> >>> >>> > >>>
>>>>> >>> >>> > >>>
vds=Host[Compute5,0bf6b00f-7947-4411-b55a-cc5eea2b381a]), log
>>>>> >>> >>> > >>> id:
>>>>> >>> >>> > >>> 6d560cc2
>>>>> >>> >>> > >>> 2014-11-24 18:05:37,074
INFO
>>>>> >>> >>> > >>>
>>>>> >>> >>> > >>>
>>>>> >>> >>> > >>>
>>>>> >>> >>> > >>>
[org.ovirt.engine.core.vdsbroker.vdsbroker.GetHardwareInfoVDSCommand]
>>>>> >>> >>> > >>>
(DefaultQuartzScheduler_Worker-69) FINISH,
>>>>> >>> >>> > >>>
GetHardwareInfoVDSCommand, log
>>>>> >>> >>> > >>> id: 6d560cc2
>>>>> >>> >>> > >>> 2014-11-24 18:05:37,093
WARN
>>>>> >>> >>> > >>>
[org.ovirt.engine.core.vdsbroker.VdsManager]
>>>>> >>> >>> > >>>
(DefaultQuartzScheduler_Worker-69) Host Compute5 is
>>>>> >>> >>> > >>> running
>>>>> >>> >>> > >>> with
>>>>> >>> >>> > >>> disabled
>>>>> >>> >>> > >>> SELinux.
>>>>> >>> >>> > >>> 2014-11-24 18:05:37,127
INFO
>>>>> >>> >>> > >>>
>>>>> >>> >>> > >>>
>>>>> >>> >>> > >>>
>>>>> >>> >>> > >>>
[org.ovirt.engine.core.bll.HandleVdsCpuFlagsOrClusterChangedCommand]
>>>>> >>> >>> > >>>
(DefaultQuartzScheduler_Worker-69) [2b4a51cf] Running
>>>>> >>> >>> > >>> command:
>>>>> >>> >>> > >>>
HandleVdsCpuFlagsOrClusterChangedCommand internal: true.
>>>>> >>> >>> > >>> Entities
>>>>> >>> >>> > >>> affected
>>>>> >>> >>> > >>> : ID:
0bf6b00f-7947-4411-b55a-cc5eea2b381a Type: VDS
>>>>> >>> >>> > >>> 2014-11-24 18:05:37,147
INFO
>>>>> >>> >>> > >>>
>>>>> >>> >>> > >>>
>>>>> >>> >>> > >>>
>>>>> >>> >>> > >>>
[org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand]
>>>>> >>> >>> > >>>
(DefaultQuartzScheduler_Worker-69) [2b4a51cf] START,
>>>>> >>> >>> > >>>
GlusterServersListVDSCommand(HostName = Compute5, HostId
>>>>> >>> >>> > >>> =
>>>>> >>> >>> > >>>
0bf6b00f-7947-4411-b55a-cc5eea2b381a), log id: 4faed87
>>>>> >>> >>> > >>> 2014-11-24 18:05:37,164
INFO
>>>>> >>> >>> > >>>
>>>>> >>> >>> > >>>
>>>>> >>> >>> > >>>
>>>>> >>> >>> > >>>
[org.ovirt.engine.core.vdsbroker.gluster.GlusterServersListVDSCommand]
>>>>> >>> >>> > >>>
(DefaultQuartzScheduler_Worker-69) [2b4a51cf] FINISH,
>>>>> >>> >>> > >>>
GlusterServersListVDSCommand, log id: 4faed87
>>>>> >>> >>> > >>> 2014-11-24 18:05:37,189
INFO
>>>>> >>> >>> > >>>
[org.ovirt.engine.core.bll.SetNonOperationalVdsCommand]
>>>>> >>> >>> > >>>
(DefaultQuartzScheduler_Worker-69) [4a84c4e5] Running
>>>>> >>> >>> > >>> command:
>>>>> >>> >>> > >>>
SetNonOperationalVdsCommand internal: true. Entities
>>>>> >>> >>> > >>> affected :
>>>>> >>> >>> > >>> ID:
>>>>> >>> >>> > >>>
0bf6b00f-7947-4411-b55a-cc5eea2b381a Type: VDS
>>>>> >>> >>> > >>> 2014-11-24 18:05:37,206
INFO
>>>>> >>> >>> > >>>
>>>>> >>> >>> > >>>
[org.ovirt.engine.core.vdsbroker.SetVdsStatusVDSCommand]
>>>>> >>> >>> > >>>
(DefaultQuartzScheduler_Worker-69) [4a84c4e5] START,
>>>>> >>> >>> > >>>
SetVdsStatusVDSCommand(HostName = Compute5, HostId =
>>>>> >>> >>> > >>>
0bf6b00f-7947-4411-b55a-cc5eea2b381a,
>>>>> >>> >>> > >>> status=NonOperational,
>>>>> >>> >>> > >>>
nonOperationalReason=GLUSTER_COMMAND_FAILED,
>>>>> >>> >>> > >>>
stopSpmFailureLogged=false),
>>>>> >>> >>> > >>> log id: fed5617
>>>>> >>> >>> > >>> 2014-11-24 18:05:37,209
INFO
>>>>> >>> >>> > >>>
>>>>> >>> >>> > >>>
[org.ovirt.engine.core.vdsbroker.SetVdsStatusVDSCommand]
>>>>> >>> >>> > >>>
(DefaultQuartzScheduler_Worker-69) [4a84c4e5] FINISH,
>>>>> >>> >>> > >>> SetVdsStatusVDSCommand,
log id: fed5617
>>>>> >>> >>> > >>> 2014-11-24 18:05:37,223
ERROR
>>>>> >>> >>> > >>>
>>>>> >>> >>> > >>>
>>>>> >>> >>> > >>>
>>>>> >>> >>> > >>>
[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
>>>>> >>> >>> > >>>
(DefaultQuartzScheduler_Worker-69) [4a84c4e5]
>>>>> >>> >>> > >>> Correlation ID:
>>>>> >>> >>> > >>> 4a84c4e5,
>>>>> >>> >>> > >>> Job
>>>>> >>> >>> > >>> ID:
4bfd4a6d-c3ef-468f-a40e-a3a6ca13011b, Call Stack:
>>>>> >>> >>> > >>> null,
>>>>> >>> >>> > >>> Custom
>>>>> >>> >>> > >>> Event
>>>>> >>> >>> > >>> ID: -1, Message: Gluster
command [<UNKNOWN>] failed on
>>>>> >>> >>> > >>> server
>>>>> >>> >>> > >>> Compute5.
>>>>> >>> >>> > >>> 2014-11-24 18:05:37,243
INFO
>>>>> >>> >>> > >>>
>>>>> >>> >>> > >>>
>>>>> >>> >>> > >>>
>>>>> >>> >>> > >>>
[org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector]
>>>>> >>> >>> > >>>
(DefaultQuartzScheduler_Worker-69) [4a84c4e5]
>>>>> >>> >>> > >>> Correlation ID:
>>>>> >>> >>> > >>> null,
>>>>> >>> >>> > >>> Call
>>>>> >>> >>> > >>> Stack: null, Custom Event
ID: -1, Message: Status of
>>>>> >>> >>> > >>> host
>>>>> >>> >>> > >>> Compute5
>>>>> >>> >>> > >>> was
>>>>> >>> >>> > >>> set
>>>>> >>> >>> > >>> to NonOperational.
>>>>> >>> >>> > >>> 2014-11-24 18:05:37,272
INFO
>>>>> >>> >>> > >>>
[org.ovirt.engine.core.bll.HandleVdsVersionCommand]
>>>>> >>> >>> > >>>
(DefaultQuartzScheduler_Worker-69) [a0c8a7f] Running
>>>>> >>> >>> > >>> command:
>>>>> >>> >>> > >>> HandleVdsVersionCommand
internal: true. Entities
>>>>> >>> >>> > >>> affected :
>>>>> >>> >>> > >>> ID:
>>>>> >>> >>> > >>>
0bf6b00f-7947-4411-b55a-cc5eea2b381a Type: VDS
>>>>> >>> >>> > >>> 2014-11-24 18:05:37,274
INFO
>>>>> >>> >>> > >>>
[org.ovirt.engine.core.vdsbroker.VdsUpdateRunTimeInfo]
>>>>> >>> >>> > >>>
(DefaultQuartzScheduler_Worker-69) [a0c8a7f] Host
>>>>> >>> >>> > >>>
0bf6b00f-7947-4411-b55a-cc5eea2b381a : Compute5 is
>>>>> >>> >>> > >>> already in
>>>>> >>> >>> > >>> NonOperational status for
reason GLUSTER_COMMAND_FAILED.
>>>>> >>> >>> > >>> SetNonOperationalVds
command is skipped.
>>>>> >>> >>> > >>> 2014-11-24 18:05:38,065
INFO
>>>>> >>> >>> > >>>
>>>>> >>> >>> > >>>
>>>>> >>> >>> > >>>
>>>>> >>> >>> > >>>
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand]
>>>>> >>> >>> > >>>
(DefaultQuartzScheduler_Worker-55) [3706e836] FINISH,
>>>>> >>> >>> > >>>
GlusterVolumesListVDSCommand, return:
>>>>> >>> >>> > >>>
>>>>> >>> >>> > >>>
>>>>> >>> >>> > >>>
>>>>> >>> >>> > >>>
{26ae1672-ee09-4a38-8fd2-72dd9974cc2b=org.ovirt.engine.core.common.businessentities.gluster.GlusterVolumeEntity@4e72a1b1},
>>>>> >>> >>> > >>> log id: 48a0c832
>>>>> >>> >>> > >>> 2014-11-24 18:05:43,243
INFO
>>>>> >>> >>> > >>>
>>>>> >>> >>> > >>>
>>>>> >>> >>> > >>>
>>>>> >>> >>> > >>>
[org.ovirt.engine.core.vdsbroker.gluster.GlusterVolumesListVDSCommand]
>>>>> >>> >>> > >>>
(DefaultQuartzScheduler_Worker-35) START,
>>>>> >>> >>> > >>>
GlusterVolumesListVDSCommand(HostName = Compute4, HostId
>>>>> >>> >>> > >>> =
>>>>> >>> >>> > >>>
33648a90-200c-45ca-89d5-1ce305d79a6a), log id: 3ce13ebc
>>>>> >>> >>> > >>> ^C
>>>>> >>> >>> > >>> [root@ccr01 ~]#
>>>>> >>> >>> > >>>
>>>>> >>> >>> > >>> Thanks,
>>>>> >>> >>> > >>> Punit
>>>>> >>> >>> > >>>
>>>>> >>> >>> > >>>
>>>>> >>> >>> > >>>
_______________________________________________
>>>>> >>> >>> > >>> Users mailing
>>>>> >>> >>> > >>>
>>>>> >>> >>> > >>>
>>>>> >>> >>> > >>>
listUsers@ovirt.orghttp://lists.ovirt.org/mailman/listinfo/users
>>>>> >>> >>> > >>>
>>>>> >>> >>> > >>>
>>>>> >>> >>> > >>>
>>>>> >>> >>> > >>
>>>>> >>> >>> > >>
>>>>> >>> >>> > >
>>>>> >>> >>> > >
>>>>> >>> >>> >
>>>>> >>> >>
>>>>> >>> >>
>>>>> >>> >
>>>>> >>> >
>>>>> >>> > _______________________________________________
>>>>> >>> > Gluster-users mailing list
>>>>> >>> > Gluster-users(a)gluster.org
>>>>> >>> >
http://supercolony.gluster.org/mailman/listinfo/gluster-users
>>>>> >>
>>>>> >>
>>>>> >
>>>>
>>>>
>>