I found this on vdsm log
Thread-432::DEBUG::2015-09-02
19:37:30,854::bindingxmlrpc::1256::vds::(wrapper) client [127.0.0.1]::call
vmGetStats with ('ab
1dc1a9-b6e9-4890-8485-1019da2f328f',) {}
Thread-432::DEBUG::2015-09-02
19:37:30,854::bindingxmlrpc::1263::vds::(wrapper) return vmGetStats with
{'status': {'message':
'*Virtual machine does not exist*', 'code': 1}}
I really don't understand anything
2015-09-02 17:01 GMT+01:00 wodel youchi <wodel.youchi(a)gmail.com>:
Thanks,
but before that I stuck again with the storage of the VM engine not
detected after reboot.
the /rhev is populated, but ovirt-ha-agent crashes with
MainThread::INFO::2015-09-02
16:12:20,261::brokerlink::129::ovirt_hosted_engine_ha.lib.brokerlink.BrokerLink::(start_monitor)
Starting monitor engine-health, options {'use_ssl': 'true',
'vm_uuid':
'ab1dc1a9-b6e9-4890-8485-1019da2f328f', 'address': '0
'}
MainThread::INFO::2015-09-02
16:12:20,283::brokerlink::140::ovirt_hosted_engine_ha.lib.brokerlink.BrokerLink::(start_monitor)
Success, id 139994237094736
MainThread::INFO::2015-09-02
16:12:20,702::brokerlink::178::ovirt_hosted_engine_ha.lib.brokerlink.BrokerLink::(set_storage_do
main) Success, id 139994236985168
MainThread::INFO::2015-09-02
16:12:20,702::hosted_engine::574::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_ini
tialize_broker) Broker initialized, all submonitors started
MainThread::INFO::2015-09-02
16:12:20,799::hosted_engine::678::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_ini
tialize_sanlock) Ensuring lease for lockspace hosted-engine, host id 1 is
acquired (file: /var/run/vdsm/storage/8b25f3be-7574
-4f7a-8851-363129704e52/a44d1302-3165-4632-9d99-3e035dfc3ac7/0f260ab0-3631-4c71-b332-c6c7f67f7342)
MainThread::INFO::2015-09-02
16:12:20,800::hosted_engine::401::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(star
t_monitoring) Reloading vm.conf from the shared storage domain
MainThread::ERROR::2015-09-02
16:12:20,927::agent::201::ovirt_hosted_engine_ha.agent.agent.Agent::(_run_agent)
*Error: ''Configuration value not found:
file=/var/run/ovirt-hosted-engine-ha/vm.conf, key=memSize'' - trying to
restart agent *
MainThread::WARNING::2015-09-02
16:12:25,932::agent::204::ovirt_hosted_engine_ha.agent.agent.Agent::(_run_agent)
Restarting a
gent, attempt '9'
MainThread::ERROR::2015-09-02
16:12:25,933::agent::206::ovirt_hosted_engine_ha.agent.agent.Agent::(_run_agent)
Too many error
s occurred, giving up. Please review the log and consider filing a bug.
MainThread::INFO::2015-09-02
16:12:25,933::agent::143::ovirt_hosted_engine_ha.agent.agent.Agent::(run)
Agent shutting down
I restared vdsm ha-agent and broker-agent wihtout success
When executed
[root@noveria ~]# hosted-engine --vm-status
You must run deploy first
I got this
[root@noveria ~]# tree /var/run/vdsm/storage/
/var/run/vdsm/storage/
└── 8b25f3be-7574-4f7a-8851-363129704e52
├── 8e49032f-680b-40c2-b422-80d86dc7beda
│ └── f05762e5-e8cd-45e7-ac19-303c1ade79d1 ->
/dev/8b25f3be-7574-4f7a-8851-363129704e52/f05762e5-e8cd-45e7-ac19-303c1
ade79d1
├── a44d1302-3165-4632-9d99-3e035dfc3ac7
│ └── 0f260ab0-3631-4c71-b332-c6c7f67f7342 ->
/dev/8b25f3be-7574-4f7a-8851-363129704e52/0f260ab0-3631-4c71-b332-c6c7f
67f7342
├── a5475e57-c6f5-4dc5-a3f2-7fb782d613a7
│ └── ae352fab-7477-4376-aa27-04c321b4fbd1 ->
/dev/8b25f3be-7574-4f7a-8851-363129704e52/ae352fab-7477-4376-aa27-04c32
1b4fbd1
└── bf3bdae1-7318-4443-a19b-7371de30b982
└── cbb10cf0-9600-465e-aed9-412f7157706b ->
/dev/8b25f3be-7574-4f7a-8851-363129704e52/cbb10cf0-9600-465e-aed9-412f7
157706b
and this
[root@noveria rhev]# tree
.
└── data-center
├── 00000001-0001-0001-0001-000000000221
└── mnt
├── blockSD
│ └── 8b25f3be-7574-4f7a-8851-363129704e52
│ ├── dom_md
│ │ ├── ids ->
/dev/8b25f3be-7574-4f7a-8851-363129704e52/ids
│ │ ├── inbox ->
/dev/8b25f3be-7574-4f7a-8851-363129704e52/inbox
│ │ ├── leases ->
/dev/8b25f3be-7574-4f7a-8851-363129704e52/leases
│ │ ├── master ->
/dev/8b25f3be-7574-4f7a-8851-363129704e52/master
│ │ ├── metadata ->
/dev/8b25f3be-7574-4f7a-8851-363129704e52/metadata
│ │ └── outbox ->
/dev/8b25f3be-7574-4f7a-8851-363129704e52/outbox
│ ├── ha_agent
│ │ ├── hosted-engine.lockspace ->
/var/run/vdsm/storage/8b25f3be-7574-4f7a-8851-363129704e52/a44d1302-3165-4
632-9d99-3e035dfc3ac7/0f260ab0-3631-4c71-b332-c6c7f67f7342
│ │ └── hosted-engine.metadata ->
/var/run/vdsm/storage/8b25f3be-7574-4f7a-8851-363129704e52/8e49032f-680b-40
c2-b422-80d86dc7beda/f05762e5-e8cd-45e7-ac19-303c1ade79d1
│ └── images
│ ├── 8e49032f-680b-40c2-b422-80d86dc7beda
│ │ └── f05762e5-e8cd-45e7-ac19-303c1ade79d1 ->
/dev/8b25f3be-7574-4f7a-8851-363129704e52/f05762e5-e8cd-4
5e7-ac19-303c1ade79d1
│ ├── a44d1302-3165-4632-9d99-3e035dfc3ac7
│ │ └── 0f260ab0-3631-4c71-b332-c6c7f67f7342 ->
/dev/8b25f3be-7574-4f7a-8851-363129704e52/0f260ab0-3631-4
c71-b332-c6c7f67f7342
│ ├── a5475e57-c6f5-4dc5-a3f2-7fb782d613a7
│ │ └── ae352fab-7477-4376-aa27-04c321b4fbd1 ->
/dev/8b25f3be-7574-4f7a-8851-363129704e52/ae352fab-7477-4
376-aa27-04c321b4fbd1
│ └── bf3bdae1-7318-4443-a19b-7371de30b982
│ └── cbb10cf0-9600-465e-aed9-412f7157706b ->
/dev/8b25f3be-7574-4f7a-8851-363129704e52/cbb10cf0-9600-4
65e-aed9-412f7157706b
├── openSuse.wodel.wd:_nvms
└── _var_lib_ovirt-hosted-engine-setup_tmp2fNoEf
Here I did find some symblic links blinking (not present) like this one
hosted-engine.metadata ->
/var/run/vdsm/storage/8b25f3be-7574-4f7a-8851-363129704e52/8e49032f-680b-40
c2-b422-80d86dc7beda/*f05762e5-e8cd-45e7-ac19-303c1ade79d1*
the lvscan command showed that the lv concerned is inactive, is this
correct?
[root@noveria ~]# lvscan
File descriptor 9 (/dev/dri/card0) leaked on lvscan invocation. Parent PID
2935: bash
ACTIVE '/dev/8b25f3be-7574-4f7a-8851-363129704e52/metadata'
[512,00 MiB] inherit
ACTIVE '/dev/8b25f3be-7574-4f7a-8851-363129704e52/outbox'
[128,00 MiB] inherit
ACTIVE '/dev/8b25f3be-7574-4f7a-8851-363129704e52/leases'
[2,00 GiB] inherit
ACTIVE '/dev/8b25f3be-7574-4f7a-8851-363129704e52/ids' [128,00
MiB] inherit
ACTIVE '/dev/8b25f3be-7574-4f7a-8851-363129704e52/inbox'
[128,00 MiB] inherit
ACTIVE '/dev/8b25f3be-7574-4f7a-8851-363129704e52/master'
[1,00 GiB] inherit
inactive
'/dev/8b25f3be-7574-4f7a-8851-363129704e52/ae352fab-7477-4376-aa27-04c321b4fbd1'
[1,00 GiB] inherit
ACTIVE
'/dev/8b25f3be-7574-4f7a-8851-363129704e52/0f260ab0-3631-4c71-b332-c6c7f67f7342'
[128,00 MiB] inherit
*inactive
'/dev/8b25f3be-7574-4f7a-8851-363129704e52/f05762e5-e8cd-45e7-ac19-303c1ade79d1'
[128,00 MiB] inherit *
inactive
'/dev/8b25f3be-7574-4f7a-8851-363129704e52/cbb10cf0-9600-465e-aed9-412f7157706b'
[40,00 GiB] inherit
and this
[root@noveria ~]# vdsClient -s 0 prepareImage
"00000000-0000-0000-0000-000000000000"
"8b25f3be-7574-4f7a-8851-363129704e52" "
bf3bdae1-7318-4443-a19b-7371de30b982"
"cbb10cf0-9600-465e-aed9-412f7157706b"
{'domainID': '8b25f3be-7574-4f7a-8851-363129704e52',
'imageID': 'bf3bdae1-7318-4443-a19b-7371de30b982',
'leaseOffset': 112197632,
'leasePath': '/dev/8b25f3be-7574-4f7a-8851-363129704e52/leases',
'path':
'/rhev/data-center/mnt/blockSD/8b25f3be-7574-4f7a-8851-363129704e52/images/bf3bdae1-7318-4443-a19b-7371de30b982/cbb1
0cf0-9600-465e-aed9-412f7157706b',
'volType': 'path',
'volumeID': 'cbb10cf0-9600-465e-aed9-412f7157706b'}
and
[root@noveria ~]# vdsClient -s 0 getStorageDomainInfo
8b25f3be-7574-4f7a-8851-363129704e52
uuid = 8b25f3be-7574-4f7a-8851-363129704e52
vguuid = tJKiwH-Cn7v-QCxd-YQrg-MUxA-fbdC-kdga8m
state = OK
version = 3
role = Regular
type = ISCSI
class = Data
pool = []
name = hosted_storage
[root@noveria ~]# lvs
File descriptor 9 (/dev/dri/card0) leaked on lvs invocation. Parent PID
3105: bash
LV VG
Attr LSize Pool Origin Data%
Meta% Move
Log Cpy%Sync Convert
0f260ab0-3631-4c71-b332-c6c7f67f7342 8b25f3be-7574-4f7a-8851-363129704e52
-wi-ao---- 128,00m
ae352fab-7477-4376-aa27-04c321b4fbd1 8b25f3be-7574-4f7a-8851-363129704e52
-wi------- 1,00g
cbb10cf0-9600-465e-aed9-412f7157706b 8b25f3be-7574-4f7a-8851-363129704e52
-wi-a----- 40,00g
f05762e5-e8cd-45e7-ac19-303c1ade79d1 8b25f3be-7574-4f7a-8851-363129704e52
-wi------- 128,00m
ids 8b25f3be-7574-4f7a-8851-363129704e52
-wi-a----- 128,00m
inbox 8b25f3be-7574-4f7a-8851-363129704e52
-wi-a----- 128,00m
leases 8b25f3be-7574-4f7a-8851-363129704e52
-wi-a----- 2,00g
master 8b25f3be-7574-4f7a-8851-363129704e52
-wi-a----- 1,00g
metadata 8b25f3be-7574-4f7a-8851-363129704e52
-wi-a----- 512,00m
outbox 8b25f3be-7574-4f7a-8851-363129704e52
-wi-a----- 128,00m
VDSM logs doesn't show me anything
MainThread::INFO::2015-09-01 23:34:49,551::vdsm::166::vds::(run)
<WorkerThread(Thread-4, started daemon 139990108333824)>
MainThread::INFO::2015-09-01 23:34:49,552::vdsm::166::vds::(run)
<WorkerThread(Thread-3, started daemon 139990116726528)>
MainThread::INFO::2015-09-02 16:07:49,510::vdsm::156::vds::(run) (PID:
1554) I am the actual vdsm 4.17.3-12.git7288ef7.fc22 noveria.wodel.wd
(4.1.6-200.fc22.x86_64)
MainThread::DEBUG::2015-09-02
16:07:49,524::resourceManager::421::Storage.ResourceManager::(registerNamespace)
Registering namespace 'Storage'
MainThread::DEBUG::2015-09-02
16:07:49,524::threadPool::29::Storage.ThreadPool::(__init__) Enter -
numThreads: 10, waitTimeout: 3, maxTasks: 500
MainThread::DEBUG::2015-09-02
16:07:49,526::fileUtils::143::Storage.fileUtils::(createdir) Creating
directory: /rhev/data-center/mnt mode: None
MainThread::WARNING::2015-09-02
16:07:49,526::fileUtils::152::Storage.fileUtils::(createdir) Dir
/rhev/data-center/mnt already exists
MainThread::DEBUG::2015-09-02
16:07:49,564::hsm::403::Storage.Misc.excCmd::(__validateLvmLockingType)
/usr/bin/sudo -n /usr/sbin/lvm dumpconfig global/locking_type (cwd None)
MainThread::DEBUG::2015-09-02
16:07:49,611::hsm::403::Storage.Misc.excCmd::(__validateLvmLockingType)
SUCCESS: <err> = ''; <rc> = 0
MainThread::DEBUG::2015-09-02
16:07:49,611::hsm::427::Storage.HSM::(__cleanStorageRepository) Started
cleaning storage repository at '/rhev/data-center'
MainThread::DEBUG::2015-09-02
16:07:49,614::hsm::459::Storage.HSM::(__cleanStorageRepository) White list:
['/rhev/data-center/hsm-tasks', '/rhev/data-center/hsm-tasks/*',
'/rhev/data-center/mnt']
MainThread::DEBUG::2015-09-02
16:07:49,614::hsm::460::Storage.HSM::(__cleanStorageRepository) Mount list:
[]
MainThread::DEBUG::2015-09-02
16:07:49,614::hsm::462::Storage.HSM::(__cleanStorageRepository) Cleaning
leftovers
MainThread::DEBUG::2015-09-02
16:07:49,615::hsm::505::Storage.HSM::(__cleanStorageRepository) Finished
cleaning storage repository at '/rhev/data-center'
storageRefresh::DEBUG::2015-09-02
16:07:49,616::misc::733::Storage.SamplingMethod::(__call__) Trying to enter
sampling method (storage.sdc.refreshStorage)
MainThread::INFO::2015-09-02
16:07:49,617::dispatcher::46::Storage.Dispatcher::(__init__) Starting
StorageDispatcher...
storageRefresh::DEBUG::2015-09-02
16:07:49,620::misc::736::Storage.SamplingMethod::(__call__) Got in to
sampling method
storageRefresh::DEBUG::2015-09-02
16:07:49,792::misc::733::Storage.SamplingMethod::(__call__) Trying to enter
sampling method (storage.iscsi.rescan)
storageRefresh::DEBUG::2015-09-02
16:07:49,793::misc::736::Storage.SamplingMethod::(__call__) Got in to
sampling method
storageRefresh::DEBUG::2015-09-02
16:07:49,793::iscsi::431::Storage.ISCSI::(rescan) Performing SCSI scan,
this will take up to 30 seconds
storageRefresh::DEBUG::2015-09-02
16:07:49,924::iscsiadm::97::Storage.Misc.excCmd::(_runCmd) /usr/bin/sudo -n
/sbin/iscsiadm -m session -R (cwd None)
MainThread::DEBUG::2015-09-02
16:07:49,924::task::595::Storage.TaskManager.Task::(_updateState)
Task=`68d01d7d-b426-4465-829e-174e2cb47e9e`::moving from state init ->
state preparing
MainThread::INFO::2015-09-02
16:07:49,924::logUtils::48::dispatcher::(wrapper) Run and protect:
registerDomainStateChangeCallback(callbackFunc=<functools.partial object at
0x7fc2f03fa6d8>)
MainThread::INFO::2015-09-02
16:07:49,924::logUtils::51::dispatcher::(wrapper) Run and protect:
registerDomainStateChangeCallback, Return response: None
MainThread::DEBUG::2015-09-02
16:07:49,927::task::1191::Storage.TaskManager.Task::(prepare)
Task=`68d01d7d-b426-4465-829e-174e2cb47e9e`::finished: None
MainThread::DEBUG::2015-09-02
16:07:49,927::task::595::Storage.TaskManager.Task::(_updateState)
Task=`68d01d7d-b426-4465-829e-174e2cb47e9e`::moving from state preparing ->
state finished
MainThread::DEBUG::2015-09-02
16:07:49,927::resourceManager::940::Storage.ResourceManager.Owner::(releaseAll)
Owner.releaseAll requests {} resources {}
MainThread::DEBUG::2015-09-02
16:07:49,927::resourceManager::977::Storage.ResourceManager.Owner::(cancelAll)
Owner.cancelAll requests {}
MainThread::DEBUG::2015-09-02
16:07:49,928::task::993::Storage.TaskManager.Task::(_decref)
Task=`68d01d7d-b426-4465-829e-174e2cb47e9e`::ref 0 aborting False
MainThread::INFO::2015-09-02 16:07:49,928::momIF::46::MOM::(__init__)
Preparing MOM interface
MainThread::INFO::2015-09-02 16:07:49,929::momIF::55::MOM::(__init__)
Using named unix socket /var/run/vdsm/mom-vdsm.sock
MainThread::INFO::2015-09-02 16:07:49,929::secret::90::root::(clear)
Unregistering all secrests
MainThread::DEBUG::2015-09-02
16:07:49,929::libvirtconnection::160::root::(get) trying to connect libvirt
MainThread::INFO::2015-09-02
16:07:49,933::vmchannels::196::vds::(settimeout) Setting channels' timeout
to 30 seconds.
VM Channels Listener::DEBUG::2015-09-02
16:07:49,934::vmchannels::178::vds::(run) Starting VM channels listener
thread.
MainThread::INFO::2015-09-02
16:07:49,935::protocoldetector::172::vds.MultiProtocolAcceptor::(__init__)
Listening at 0.0.0.0:54321
MainThread::DEBUG::2015-09-02
16:07:50,063::protocoldetector::199::vds.MultiProtocolAcceptor::(add_detector)
Adding detector <rpc.bindingxmlrpc.XmlDetector instance at 0x7fc2f00dc440>
storageRefresh::DEBUG::2015-09-02
16:07:50,063::misc::743::Storage.SamplingMethod::(__call__) Returning last
result
storageRefresh::DEBUG::2015-09-02
16:07:50,080::misc::733::Storage.SamplingMethod::(__call__) Trying to enter
sampling method (storage.hba.rescan)
storageRefresh::DEBUG::2015-09-02
16:07:50,081::misc::736::Storage.SamplingMethod::(__call__) Got in to
sampling method
storageRefresh::DEBUG::2015-09-02
16:07:50,081::hba::56::Storage.HBA::(rescan) Starting scan
storageRefresh::DEBUG::2015-09-02
16:07:50,081::supervdsm::76::SuperVdsmProxy::(_connect) Trying to connect
to Super Vdsm
MainThread::DEBUG::2015-09-02
16:07:50,157::protocoldetector::199::vds.MultiProtocolAcceptor::(add_detector)
Adding detector <yajsonrpc.stompreactor.StompDetector instance at
0x7fc2e01332d8>
BindingXMLRPC::INFO::2015-09-02
16:07:50,158::bindingxmlrpc::62::vds::(threaded_start) XMLRPC server running
MainThread::DEBUG::2015-09-02
16:07:50,158::schedule::98::Scheduler::(start) Starting scheduler
periodic-sched
periodic-sched::DEBUG::2015-09-02
16:07:50,159::schedule::142::Scheduler::(_run) started
MainThread::DEBUG::2015-09-02
16:07:50,159::executor::69::Executor::(start) Starting executor
MainThread::DEBUG::2015-09-02
16:07:50,159::executor::157::Executor::(__init__) Starting worker periodic/0
periodic/0::DEBUG::2015-09-02
16:07:50,159::executor::171::Executor::(_run) Worker started
MainThread::DEBUG::2015-09-02
16:07:50,159::executor::157::Executor::(__init__) Starting worker periodic/1
periodic/1::DEBUG::2015-09-02
16:07:50,160::executor::171::Executor::(_run) Worker started
MainThread::DEBUG::2015-09-02
16:07:50,160::executor::157::Executor::(__init__) Starting worker periodic/2
periodic/2::DEBUG::2015-09-02
16:07:50,160::executor::171::Executor::(_run) Worker started
MainThread::DEBUG::2015-09-02
16:07:50,160::executor::157::Executor::(__init__) Starting worker periodic/3
periodic/3::DEBUG::2015-09-02
16:07:50,160::executor::171::Executor::(_run) Worker started
MainThread::DEBUG::2015-09-02
16:07:50,160::libvirtconnection::160::root::(get) trying to connect libvirt
MainThread::DEBUG::2015-09-02
16:07:50,163::periodic::157::virt.periodic.Operation::(start) starting
operation VmDispatcher(<class 'virt.periodic.UpdateVolumes'>)
MainThread::DEBUG::2015-09-02
16:07:50,164::periodic::157::virt.periodic.Operation::(start) starting
operation VmDispatcher(<class 'virt.periodic.NumaInfoMonitor'>)
MainThread::DEBUG::2015-09-02
16:07:50,164::periodic::157::virt.periodic.Operation::(start) starting
operation VmDispatcher(<class 'virt.periodic.BlockjobMonitor'>)
MainThread::DEBUG::2015-09-02
16:07:50,164::periodic::157::virt.periodic.Operation::(start) starting
operation <virt.sampling.VMBulkSampler object at 0x7fc2e0151d10>
MainThread::DEBUG::2015-09-02
16:07:50,164::periodic::157::virt.periodic.Operation::(start) starting
operation VmDispatcher(<class 'virt.periodic.DriveWatermarkMonitor'>)
storageRefresh::DEBUG::2015-09-02
16:07:50,167::hba::62::Storage.HBA::(rescan) Scan finished
storageRefresh::DEBUG::2015-09-02
16:07:50,167::misc::743::Storage.SamplingMethod::(__call__) Returning last
result
storageRefresh::DEBUG::2015-09-02
16:07:50,167::multipath::77::Storage.Misc.excCmd::(rescan) /usr/bin/sudo -n
/usr/sbin/multipath (cwd None)
storageRefresh::DEBUG::2015-09-02
16:07:50,513::multipath::77::Storage.Misc.excCmd::(rescan) SUCCESS: <err> =
''; <rc> = 0
storageRefresh::DEBUG::2015-09-02
16:07:50,513::utils::661::root::(execCmd) /sbin/udevadm settle --timeout=5
(cwd None)
Reactor thread::INFO::2015-09-02
16:07:50,590::protocoldetector::72::ProtocolDetector.AcceptorImpl::(handle_accept)
Accepting connection from 127.0.0.1:56311
Reactor thread::DEBUG::2015-09-02
16:07:50,596::protocoldetector::82::ProtocolDetector.Detector::(__init__)
Using required_size=11
Reactor thread::INFO::2015-09-02
16:07:50,596::protocoldetector::118::ProtocolDetector.Detector::(handle_read)
Detected protocol xml from 127.0.0.1:56311
Reactor thread::DEBUG::2015-09-02
16:07:50,596::bindingxmlrpc::1296::XmlDetector::(handle_socket) xml over
http detected from ('127.0.0.1', 56311)
BindingXMLRPC::INFO::2015-09-02
16:07:50,596::xmlrpc::73::vds.XMLRPCServer::(handle_request) Starting
request handler for 127.0.0.1:56311
Thread-13::INFO::2015-09-02
16:07:50,597::xmlrpc::84::vds.XMLRPCServer::(_process_requests) Request
handler for 127.0.0.1:56311 started
Thread-13::DEBUG::2015-09-02
16:07:50,597::bindingxmlrpc::1256::vds::(wrapper) client [127.0.0.1]::call
getHardwareInfo with () {}
Thread-13::DEBUG::2015-09-02
16:07:50,597::bindingxmlrpc::1263::vds::(wrapper) return getHardwareInfo
with {'status': {'message': 'Recovering from crash or
Initializing',
'code': 99}}
Thread-13::INFO::2015-09-02
16:07:50,599::xmlrpc::92::vds.XMLRPCServer::(_process_requests) Request
handler for 127.0.0.1:56311 stopped
Reactor thread::INFO::2015-09-02
16:07:51,607::protocoldetector::72::ProtocolDetector.AcceptorImpl::(handle_accept)
Accepting connection from 127.0.0.1:56312
Reactor thread::DEBUG::2015-09-02
16:07:51,613::protocoldetector::82::ProtocolDetector.Detector::(__init__)
Using required_size=11
Reactor thread::INFO::2015-09-02
16:07:51,613::protocoldetector::118::ProtocolDetector.Detector::(handle_read)
Detected protocol xml from 127.0.0.1:56312
Reactor thread::DEBUG::2015-09-02
16:07:51,613::bindingxmlrpc::1296::XmlDetector::(handle_socket) xml over
http detected from ('127.0.0.1', 56312)
BindingXMLRPC::INFO::2015-09-02
16:07:51,613::xmlrpc::73::vds.XMLRPCServer::(handle_request) Starting
request handler for 127.0.0.1:56312
Thread-14::INFO::2015-09-02
16:07:51,613::xmlrpc::84::vds.XMLRPCServer::(_process_requests) Request
handler for 127.0.0.1:56312 started
Thread-14::DEBUG::2015-09-02
16:07:51,614::bindingxmlrpc::1256::vds::(wrapper) client [127.0.0.1]::call
getHardwareInfo with () {}
Thread-14::DEBUG::2015-09-02
16:07:51,614::bindingxmlrpc::1263::vds::(wrapper) return getHardwareInfo
with {'status': {'message': 'Recovering from crash or
Initializing',
'code': 99}}
Thread-14::INFO::2015-09-02
16:07:51,615::xmlrpc::92::vds.XMLRPCServer::(_process_requests) Request
handler for 127.0.0.1:56312 stopped
storageRefresh::DEBUG::2015-09-02
16:07:51,924::utils::679::root::(execCmd) SUCCESS: <err> = ''; <rc> =
0
storageRefresh::DEBUG::2015-09-02
16:07:51,926::lvm::498::Storage.OperationMutex::(_invalidateAllPvs)
Operation 'lvm invalidate operation' got the operation mutex
storageRefresh::DEBUG::2015-09-02
16:07:51,926::lvm::500::Storage.OperationMutex::(_invalidateAllPvs)
Operation 'lvm invalidate operation' released the operation mutex
storageRefresh::DEBUG::2015-09-02
16:07:51,926::lvm::509::Storage.OperationMutex::(_invalidateAllVgs)
Operation 'lvm invalidate operation' got the operation mutex
storageRefresh::DEBUG::2015-09-02
16:07:51,926::lvm::511::Storage.OperationMutex::(_invalidateAllVgs)
Operation 'lvm invalidate operation' released the operation mutex
storageRefresh::DEBUG::2015-09-02
16:07:51,926::lvm::529::Storage.OperationMutex::(_invalidateAllLvs)
Operation 'lvm invalidate operation' got the operation mutex
storageRefresh::DEBUG::2015-09-02
16:07:51,926::lvm::531::Storage.OperationMutex::(_invalidateAllLvs)
Operation 'lvm invalidate operation' released the operation mutex
storageRefresh::DEBUG::2015-09-02
16:07:51,926::misc::743::Storage.SamplingMethod::(__call__) Returning last
result
storageRefresh::DEBUG::2015-09-02
16:07:51,926::lvm::320::Storage.OperationMutex::(_reloadpvs) Operation 'lvm
reload operation' got the operation mutex
storageRefresh::DEBUG::2015-09-02
16:07:51,927::lvm::291::Storage.Misc.excCmd::(cmd) /usr/bin/sudo -n
/usr/sbin/lvm pvs --config ' devices { preferred_names = ["^/dev/mapper/"]
ignore_suspended_devices=1 write_cache_state=0 disable_after_error_count=3
obtain_device_list_from_udev=0 filter = [
'\''a|/dev/mapper/Hitachi_HDS721010DLE630_MSK523Y209VK0B|'\'',
'\''r|.*|'\'' ] } global { locking_type=1
prioritise_write_locks=1
wait_for_locks=1 use_lvmetad=0 } backup { retain_min = 50 retain_days =
0 } ' --noheadings --units b --nosuffix --separator '|'
--ignoreskippedcluster -o
uuid,name,size,vg_name,vg_uuid,pe_start,pe_count,pe_alloc_count,mda_count,dev_size
(cwd None)
storageRefresh::DEBUG::2015-09-02
16:07:52,341::lvm::291::Storage.Misc.excCmd::(cmd) SUCCESS: <err> = '
WARNING: lvmetad is running but disabled. Restart lvmetad before enabling
it!\n'; <rc> = 0
storageRefresh::DEBUG::2015-09-02
16:07:52,341::lvm::348::Storage.OperationMutex::(_reloadpvs) Operation 'lvm
reload operation' released the operation mutex
storageRefresh::DEBUG::2015-09-02
16:07:52,341::lvm::371::Storage.OperationMutex::(_reloadvgs) Operation 'lvm
reload operation' got the operation mutex
storageRefresh::DEBUG::2015-09-02
16:07:52,341::lvm::291::Storage.Misc.excCmd::(cmd) /usr/bin/sudo -n
/usr/sbin/lvm vgs --config ' devices { preferred_names = ["^/dev/mapper/"]
ignore_suspended_devices=1 write_cache_state=0 disable_after_error_count=3
obtain_device_list_from_udev=0 filter = [
'\''a|/dev/mapper/Hitachi_HDS721010DLE630_MSK523Y209VK0B|'\'',
'\''r|.*|'\'' ] } global { locking_type=1
prioritise_write_locks=1
wait_for_locks=1 use_lvmetad=0 } backup { retain_min = 50 retain_days =
0 } ' --noheadings --units b --nosuffix --separator '|'
--ignoreskippedcluster -o
uuid,name,attr,size,free,extent_size,extent_count,free_count,tags,vg_mda_size,vg_mda_free,lv_count,pv_count,pv_name
(cwd None)
storageRefresh::DEBUG::2015-09-02
16:07:52,405::lvm::291::Storage.Misc.excCmd::(cmd) SUCCESS: <err> = '
WARNING: lvmetad is running but disabled. Restart lvmetad before enabling
it!\n'; <rc> = 0
storageRefresh::DEBUG::2015-09-02
16:07:52,405::lvm::416::Storage.OperationMutex::(_reloadvgs) Operation 'lvm
reload operation' released the operation mutex
storageRefresh::DEBUG::2015-09-02
16:07:52,406::lvm::291::Storage.Misc.excCmd::(cmd) /usr/bin/sudo -n
/usr/sbin/lvm lvs --config ' devices { preferred_names = ["^/dev/mapper/"]
ignore_suspended_devices=1 write_cache_state=0 disable_after_error_count=3
obtain_device_list_from_udev=0 filter = [
'\''a|/dev/mapper/Hitachi_HDS721010DLE630_MSK523Y209VK0B|'\'',
'\''r|.*|'\'' ] } global { locking_type=1
prioritise_write_locks=1
wait_for_locks=1 use_lvmetad=0 } backup { retain_min = 50 retain_days =
0 } ' --noheadings --units b --nosuffix --separator '|'
--ignoreskippedcluster -o
uuid,name,vg_name,attr,size,seg_start_pe,devices,tags (cwd None)
storageRefresh::DEBUG::2015-09-02
16:07:52,458::lvm::291::Storage.Misc.excCmd::(cmd) SUCCESS: <err> = '
WARNING: lvmetad is running but disabled. Restart lvmetad before enabling
it!\n'; <rc> = 0
storageRefresh::DEBUG::2015-09-02
16:07:52,458::lvm::371::Storage.OperationMutex::(_reloadvgs) Operation 'lvm
reload operation' got the operation mutex
storageRefresh::DEBUG::2015-09-02
16:07:52,459::lvm::291::Storage.Misc.excCmd::(cmd) /usr/bin/sudo -n
/usr/sbin/lvm vgs --config ' devices { preferred_names = ["^/dev/mapper/"]
ignore_suspended_devices=1 write_cache_state=0 disable_after_error_count=3
obtain_device_list_from_udev=0 filter = [
'\''a|/dev/mapper/Hitachi_HDS721010DLE630_MSK523Y209VK0B|'\'',
'\''r|.*|'\'' ] } global { locking_type=1
prioritise_write_locks=1
wait_for_locks=1 use_lvmetad=0 } backup { retain_min = 50 retain_days =
0 } ' --noheadings --units b --nosuffix --separator '|'
--ignoreskippedcluster -o
uuid,name,attr,size,free,extent_size,extent_count,free_count,tags,vg_mda_size,vg_mda_free,lv_count,pv_count,pv_name
(cwd None)
storageRefresh::DEBUG::2015-09-02
16:07:52,491::lvm::291::Storage.Misc.excCmd::(cmd) SUCCESS: <err> = '
WARNING: lvmetad is running but disabled. Restart lvmetad before enabling
it!\n'; <rc> = 0
storageRefresh::DEBUG::2015-09-02
16:07:52,491::lvm::416::Storage.OperationMutex::(_reloadvgs) Operation 'lvm
reload operation' released the operation mutex
storageRefresh::DEBUG::2015-09-02
16:07:52,491::hsm::373::Storage.HSM::(storageRefresh) HSM is ready
Reactor thread::INFO::2015-09-02
16:07:52,624::protocoldetector::72::ProtocolDetector.AcceptorImpl::(handle_accept)
Accepting connection from 127.0.0.1:56313
Reactor thread::DEBUG::2015-09-02
16:07:52,629::protocoldetector::82::ProtocolDetector.Detector::(__init__)
Using required_size=11
Reactor thread::INFO::2015-09-02
16:07:52,629::protocoldetector::118::ProtocolDetector.Detector::(handle_read)
Detected protocol xml from 127.0.0.1:56313
Reactor thread::DEBUG::2015-09-02
16:07:52,629::bindingxmlrpc::1296::XmlDetector::(handle_socket) xml over
http detected from ('127.0.0.1', 56313)
BindingXMLRPC::INFO::2015-09-02
16:07:52,629::xmlrpc::73::vds.XMLRPCServer::(handle_request) Starting
request handler for 127.0.0.1:56313
Thread-15::INFO::2015-09-02
16:07:52,630::xmlrpc::84::vds.XMLRPCServer::(_process_requests) Request
handler for 127.0.0.1:56313 started
Thread-15::DEBUG::2015-09-02
16:07:52,630::bindingxmlrpc::1256::vds::(wrapper) client [127.0.0.1]::call
getHardwareInfo with () {}
Thread-15::DEBUG::2015-09-02
16:07:52,719::bindingxmlrpc::1263::vds::(wrapper) return getHardwareInfo
with {'status': {'message': 'Done', 'code': 0},
'info':
{'systemProductName': 'System Product Name',
'systemSerialNumber': 'System
Serial Number', 'systemFamily': 'To be filled by O.E.M.',
'systemVersion':
'System Version', 'systemUUID':
'267A6B80-D7DA-11DD-81CF-C860009B3CD9',
'systemManufacturer': 'System manufacturer'}}
Thread-15::INFO::2015-09-02
16:07:52,721::xmlrpc::92::vds.XMLRPCServer::(_process_requests) Request
handler for 127.0.0.1:56313 stopped
Reactor thread::INFO::2015-09-02
16:07:52,730::protocoldetector::72::ProtocolDetector.AcceptorImpl::(handle_accept)
Accepting connection from 127.0.0.1:56314
Reactor thread::DEBUG::2015-09-02
16:07:52,735::protocoldetector::82::ProtocolDetector.Detector::(__init__)
Using required_size=11
Reactor thread::INFO::2015-09-02
16:07:52,735::protocoldetector::118::ProtocolDetector.Detector::(handle_read)
Detected protocol xml from 127.0.0.1:56314
Reactor thread::DEBUG::2015-09-02
16:07:52,735::bindingxmlrpc::1296::XmlDetector::(handle_socket) xml over
http detected from ('127.0.0.1', 56314)
BindingXMLRPC::INFO::2015-09-02
16:07:52,735::xmlrpc::73::vds.XMLRPCServer::(handle_request) Starting
request handler for 127.0.0.1:56314
Thread-16::INFO::2015-09-02
16:07:52,735::xmlrpc::84::vds.XMLRPCServer::(_process_requests) Request
handler for 127.0.0.1:56314 started
Thread-16::DEBUG::2015-09-02
16:07:52,736::bindingxmlrpc::325::vds::(wrapper) client [127.0.0.1]
Thread-16::DEBUG::2015-09-02
16:07:52,736::task::595::Storage.TaskManager.Task::(_updateState)
Task=`c4a18001-912b-47dc-9713-7d50e5133b59`::moving from state init ->
state preparing
Thread-16::INFO::2015-09-02
16:07:52,736::logUtils::48::dispatcher::(wrapper) Run and protect:
connectStorageServer(domType=3,
spUUID='00000000-0000-0000-0000-000000000000', conList=[{'id':
'57bc98c0-560f-4e61-9d86-df92ad468d3b', 'connection':
'192.168.1.50',
'iqn': 'iqn.2015-08.openSuse.wodel:target00', 'portal':
'1', 'user':
'iscsiuser', 'password': '********', 'port':
'3260'}], options=None)
Thread-16::DEBUG::2015-09-02
16:07:52,737::iscsiadm::97::Storage.Misc.excCmd::(_runCmd) /usr/bin/sudo -n
/sbin/iscsiadm -m node -T iqn.2015-08.openSuse.wodel:target00 -I default -p
192.168.1.50:3260,1 --op=new (cwd None)
Thread-16::DEBUG::2015-09-02
16:07:52,789::iscsiadm::97::Storage.Misc.excCmd::(_runCmd) SUCCESS: <err> =
''; <rc> = 0
Thread-16::DEBUG::2015-09-02
16:07:52,789::iscsiadm::97::Storage.Misc.excCmd::(_runCmd) /sbin/iscsiadm
-m node -T iqn.2015-08.openSuse.wodel:target00 -I default -p
192.168.1.50:3260,1 -n node.session.auth.authmethod -v '****' --op=update
(cwd None)
Thread-16::DEBUG::2015-09-02
16:07:52,811::iscsiadm::97::Storage.Misc.excCmd::(_runCmd) SUCCESS: <err> =
''; <rc> = 0
Thread-16::DEBUG::2015-09-02
16:07:52,812::iscsiadm::97::Storage.Misc.excCmd::(_runCmd) /sbin/iscsiadm
-m node -T iqn.2015-08.openSuse.wodel:target00 -I default -p
192.168.1.50:3260,1 -n node.session.auth.username -v '****' --op=update
(cwd None)
Thread-16::DEBUG::2015-09-02
16:07:52,846::iscsiadm::97::Storage.Misc.excCmd::(_runCmd) SUCCESS: <err> =
''; <rc> = 0
Thread-16::DEBUG::2015-09-02
16:07:52,847::iscsiadm::97::Storage.Misc.excCmd::(_runCmd) /sbin/iscsiadm
-m node -T iqn.2015-08.openSuse.wodel:target00 -I default -p
192.168.1.50:3260,1 -n node.session.auth.password -v '****' --op=update
(cwd None)
Thread-16::DEBUG::2015-09-02
16:07:52,868::iscsiadm::97::Storage.Misc.excCmd::(_runCmd) SUCCESS: <err> =
''; <rc> = 0
Thread-16::DEBUG::2015-09-02
16:07:52,868::iscsiadm::97::Storage.Misc.excCmd::(_runCmd) /usr/bin/sudo -n
/sbin/iscsiadm -m iface -I default (cwd None)
Thread-16::DEBUG::2015-09-02
16:07:52,905::iscsiadm::97::Storage.Misc.excCmd::(_runCmd) SUCCESS: <err> =
''; <rc> = 0
Thread-16::INFO::2015-09-02
16:07:52,905::iscsi::564::Storage.ISCSI::(setRpFilterIfNeeded) iSCSI
iface.net_ifacename not provided. Skipping.
Thread-16::DEBUG::2015-09-02
16:07:52,906::iscsiadm::97::Storage.Misc.excCmd::(_runCmd) /usr/bin/sudo -n
/sbin/iscsiadm -m node -T iqn.2015-08.openSuse.wodel:target00 -I default -p
192.168.1.50:3260,1 -l (cwd None)
Thread-16::DEBUG::2015-09-02
16:07:53,027::iscsiadm::97::Storage.Misc.excCmd::(_runCmd) SUCCESS: <err> =
''; <rc> = 0
Thread-16::DEBUG::2015-09-02
16:07:53,028::iscsiadm::97::Storage.Misc.excCmd::(_runCmd) /usr/bin/sudo -n
/sbin/iscsiadm -m node -T iqn.2015-08.openSuse.wodel:target00 -I default -p
192.168.1.50:3260,1 -n node.startup -v manual --op=update (cwd None)
Thread-16::DEBUG::2015-09-02
16:07:53,088::iscsiadm::97::Storage.Misc.excCmd::(_runCmd) SUCCESS: <err> =
''; <rc> = 0
Thread-16::DEBUG::2015-09-02 16:07:53,088::utils::661::root::(execCmd)
/sbin/udevadm settle --timeout=5 (cwd None)
Thread-16::DEBUG::2015-09-02 16:07:53,182::utils::679::root::(execCmd)
SUCCESS: <err> = ''; <rc> = 0
Thread-16::DEBUG::2015-09-02
16:07:53,182::misc::733::Storage.SamplingMethod::(__call__) Trying to enter
sampling method (storage.sdc.refreshStorage)
Thread-16::DEBUG::2015-09-02
16:07:53,182::misc::736::Storage.SamplingMethod::(__call__) Got in to
sampling method
Thread-16::DEBUG::2015-09-02
16:07:53,182::misc::733::Storage.SamplingMethod::(__call__) Trying to enter
sampling method (storage.iscsi.rescan)
Thread-16::DEBUG::2015-09-02
16:07:53,182::misc::736::Storage.SamplingMethod::(__call__) Got in to
sampling method
Thread-16::DEBUG::2015-09-02
16:07:53,182::iscsi::431::Storage.ISCSI::(rescan) Performing SCSI scan,
this will take up to 30 seconds
Thread-16::DEBUG::2015-09-02
16:07:53,182::iscsiadm::97::Storage.Misc.excCmd::(_runCmd) /usr/bin/sudo -n
/sbin/iscsiadm -m session -R (cwd None)
Thread-16::DEBUG::2015-09-02
16:07:53,229::misc::743::Storage.SamplingMethod::(__call__) Returning last
result
Thread-16::DEBUG::2015-09-02
16:07:53,229::misc::733::Storage.SamplingMethod::(__call__) Trying to enter
sampling method (storage.hba.rescan)
Thread-16::DEBUG::2015-09-02
16:07:53,229::misc::736::Storage.SamplingMethod::(__call__) Got in to
sampling method
Thread-16::DEBUG::2015-09-02 16:07:53,229::hba::56::Storage.HBA::(rescan)
Starting scan
Thread-16::DEBUG::2015-09-02 16:07:53,300::hba::62::Storage.HBA::(rescan)
Scan finished
Thread-16::DEBUG::2015-09-02
16:07:53,300::misc::743::Storage.SamplingMethod::(__call__) Returning last
result
Thread-16::DEBUG::2015-09-02
16:07:53,300::multipath::77::Storage.Misc.excCmd::(rescan) /usr/bin/sudo -n
/usr/sbin/multipath (cwd None)
Thread-16::DEBUG::2015-09-02
16:07:53,435::multipath::77::Storage.Misc.excCmd::(rescan) SUCCESS: <err> =
''; <rc> = 0
Thread-16::DEBUG::2015-09-02 16:07:53,435::utils::661::root::(execCmd)
/sbin/udevadm settle --timeout=5 (cwd None)
Thread-16::DEBUG::2015-09-02 16:07:53,919::utils::679::root::(execCmd)
SUCCESS: <err> = ''; <rc> = 0
Thread-16::DEBUG::2015-09-02
16:07:53,921::lvm::498::Storage.OperationMutex::(_invalidateAllPvs)
Operation 'lvm invalidate operation' got the operation mutex
Thread-16::DEBUG::2015-09-02
16:07:53,921::lvm::500::Storage.OperationMutex::(_invalidateAllPvs)
Operation 'lvm invalidate operation' released the operation mutex
Thread-16::DEBUG::2015-09-02
16:07:53,922::lvm::509::Storage.OperationMutex::(_invalidateAllVgs)
Operation 'lvm invalidate operation' got the operation mutex
Thread-16::DEBUG::2015-09-02
16:07:53,922::lvm::511::Storage.OperationMutex::(_invalidateAllVgs)
Operation 'lvm invalidate operation' released the operation mutex
Thread-16::DEBUG::2015-09-02
16:07:53,922::lvm::529::Storage.OperationMutex::(_invalidateAllLvs)
Operation 'lvm invalidate operation' got the operation mutex
Thread-16::DEBUG::2015-09-02
16:07:53,922::lvm::531::Storage.OperationMutex::(_invalidateAllLvs)
Operation 'lvm invalidate operation' released the operation mutex
Thread-16::DEBUG::2015-09-02
16:07:53,922::misc::743::Storage.SamplingMethod::(__call__) Returning last
result
Thread-16::DEBUG::2015-09-02
16:07:53,922::lvm::371::Storage.OperationMutex::(_reloadvgs) Operation 'lvm
reload operation' got the operation mutex
Thread-16::DEBUG::2015-09-02
16:07:53,923::lvm::291::Storage.Misc.excCmd::(cmd) /usr/bin/sudo -n
/usr/sbin/lvm vgs --config ' devices { preferred_names = ["^/dev/mapper/"]
ignore_suspended_devices=1 write_cache_state=0 disable_after_error_count=3
obtain_device_list_from_udev=0 filter = [
'\''a|/dev/mapper/33000000100000001|/dev/mapper/Hitachi_HDS721010DLE630_MSK523Y209VK0B|'\'',
'\''r|.*|'\'' ] } global { locking_type=1
prioritise_write_locks=1
wait_for_locks=1 use_lvmetad=0 } backup { retain_min = 50 retain_days =
0 } ' --noheadings --units b --nosuffix --separator '|'
--ignoreskippedcluster -o
uuid,name,attr,size,free,extent_size,extent_count,free_count,tags,vg_mda_size,vg_mda_free,lv_count,pv_count,pv_name
(cwd None)
Thread-16::DEBUG::2015-09-02
16:07:54,058::lvm::291::Storage.Misc.excCmd::(cmd) SUCCESS: <err> = '
WARNING: lvmetad is running but disabled. Restart lvmetad before enabling
it!\n'; <rc> = 0
Thread-16::DEBUG::2015-09-02
16:07:54,059::lvm::416::Storage.OperationMutex::(_reloadvgs) Operation 'lvm
reload operation' released the operation mutex
Thread-16::DEBUG::2015-09-02
16:07:54,059::hsm::2418::Storage.HSM::(__prefetchDomains) Found SD uuids:
('8b25f3be-7574-4f7a-8851-363129704e52',)
Thread-16::DEBUG::2015-09-02
16:07:54,059::hsm::2478::Storage.HSM::(connectStorageServer) knownSDs:
{8b25f3be-7574-4f7a-8851-363129704e52: storage.blockSD.findDomain}
Thread-16::INFO::2015-09-02
16:07:54,059::logUtils::51::dispatcher::(wrapper) Run and protect:
connectStorageServer, Return response: {'statuslist': [{'status': 0,
'id':
'57bc98c0-560f-4e61-9d86-df92ad468d3b'}]}
Thread-16::DEBUG::2015-09-02
16:07:54,059::task::1191::Storage.TaskManager.Task::(prepare)
Task=`c4a18001-912b-47dc-9713-7d50e5133b59`::finished: {'statuslist':
[{'status': 0, 'id': '57bc98c0-560f-4e61-9d86-df92ad468d3b'}]}
Thread-16::DEBUG::2015-09-02
16:07:54,059::task::595::Storage.TaskManager.Task::(_updateState)
Task=`c4a18001-912b-47dc-9713-7d50e5133b59`::moving from state preparing ->
state finished
Thread-16::DEBUG::2015-09-02
16:07:54,060::resourceManager::940::Storage.ResourceManager.Owner::(releaseAll)
Owner.releaseAll requests {} resources {}
Thread-16::DEBUG::2015-09-02
16:07:54,060::resourceManager::977::Storage.ResourceManager.Owner::(cancelAll)
Owner.cancelAll requests {}
Thread-16::DEBUG::2015-09-02
16:07:54,060::task::993::Storage.TaskManager.Task::(_decref)
Task=`c4a18001-912b-47dc-9713-7d50e5133b59`::ref 0 aborting False
Thread-16::INFO::2015-09-02
16:07:54,062::xmlrpc::92::vds.XMLRPCServer::(_process_requests) Request
handler for 127.0.0.1:56314 stopped
Reactor thread::INFO::2015-09-02
16:07:54,070::protocoldetector::72::ProtocolDetector.AcceptorImpl::(handle_accept)
Accepting connection from 127.0.0.1:56316
Reactor thread::DEBUG::2015-09-02
16:07:54,075::protocoldetector::82::ProtocolDetector.Detector::(__init__)
Using required_size=11
Reactor thread::INFO::2015-09-02
16:07:54,076::protocoldetector::118::ProtocolDetector.Detector::(handle_read)
Detected protocol xml from 127.0.0.1:56316
Reactor thread::DEBUG::2015-09-02
16:07:54,076::bindingxmlrpc::1296::XmlDetector::(handle_socket) xml over
http detected from ('127.0.0.1', 56316)
BindingXMLRPC::INFO::2015-09-02
16:07:54,076::xmlrpc::73::vds.XMLRPCServer::(handle_request) Starting
request handler for 127.0.0.1:56316
Thread-17::INFO::2015-09-02
16:07:54,076::xmlrpc::84::vds.XMLRPCServer::(_process_requests) Request
handler for 127.0.0.1:56316 started
Thread-17::DEBUG::2015-09-02
16:07:54,077::bindingxmlrpc::325::vds::(wrapper) client [127.0.0.1]
Thread-17::DEBUG::2015-09-02
16:07:54,077::task::595::Storage.TaskManager.Task::(_updateState)
Task=`7936300e-8a1a-47f5-83c4-16ed19853e36`::moving from state init ->
state preparing
Thread-17::INFO::2015-09-02
16:07:54,077::logUtils::48::dispatcher::(wrapper) Run and protect:
prepareImage(sdUUID='8b25f3be-7574-4f7a-8851-363129704e52',
spUUID='00000000-0000-0000-0000-000000000000',
imgUUID='bf3bdae1-7318-4443-a19b-7371de30b982',
leafUUID='cbb10cf0-9600-465e-aed9-412f7157706b')
Thread-17::DEBUG::2015-09-02
16:07:54,077::resourceManager::198::Storage.ResourceManager.Request::(__init__)
ResName=`Storage.8b25f3be-7574-4f7a-8851-363129704e52`ReqID=`fc59b8b4-51c5-4a15-9716-aedbb6de62e6`::Request
was made in '/usr/share/vdsm/storage/hsm.py' line '3194' at
'prepareImage'
Thread-17::DEBUG::2015-09-02
16:07:54,078::resourceManager::542::Storage.ResourceManager::(registerResource)
Trying to register resource 'Storage.8b25f3be-7574-4f7a-8851-363129704e52'
for lock type 'shared'
Thread-17::DEBUG::2015-09-02
16:07:54,078::resourceManager::601::Storage.ResourceManager::(registerResource)
Resource 'Storage.8b25f3be-7574-4f7a-8851-363129704e52' is free. Now
locking as 'shared' (1 active user)
Thread-17::DEBUG::2015-09-02
16:07:54,078::resourceManager::238::Storage.ResourceManager.Request::(grant)
ResName=`Storage.8b25f3be-7574-4f7a-8851-363129704e52`ReqID=`fc59b8b4-51c5-4a15-9716-aedbb6de62e6`::Granted
request
Thread-17::DEBUG::2015-09-02
16:07:54,078::task::827::Storage.TaskManager.Task::(resourceAcquired)
Task=`7936300e-8a1a-47f5-83c4-16ed19853e36`::_resourcesAcquired:
Storage.8b25f3be-7574-4f7a-8851-363129704e52 (shared)
Thread-17::DEBUG::2015-09-02
16:07:54,078::task::993::Storage.TaskManager.Task::(_decref)
Task=`7936300e-8a1a-47f5-83c4-16ed19853e36`::ref 1 aborting False
Thread-17::DEBUG::2015-09-02
16:07:54,078::misc::733::Storage.SamplingMethod::(__call__) Trying to enter
sampling method (storage.sdc.refreshStorage)
Thread-17::DEBUG::2015-09-02
16:07:54,078::misc::736::Storage.SamplingMethod::(__call__) Got in to
sampling method
Thread-17::DEBUG::2015-09-02
16:07:54,078::misc::733::Storage.SamplingMethod::(__call__) Trying to enter
sampling method (storage.iscsi.rescan)
Thread-17::DEBUG::2015-09-02
16:07:54,078::misc::736::Storage.SamplingMethod::(__call__) Got in to
sampling method
Thread-17::DEBUG::2015-09-02
16:07:54,078::iscsi::431::Storage.ISCSI::(rescan) Performing SCSI scan,
this will take up to 30 seconds
Thread-17::DEBUG::2015-09-02
16:07:54,078::iscsiadm::97::Storage.Misc.excCmd::(_runCmd) /usr/bin/sudo -n
/sbin/iscsiadm -m session -R (cwd None)
Thread-17::DEBUG::2015-09-02
16:07:54,130::misc::743::Storage.SamplingMethod::(__call__) Returning last
result
Thread-17::DEBUG::2015-09-02
16:07:54,130::misc::733::Storage.SamplingMethod::(__call__) Trying to enter
sampling method (storage.hba.rescan)
Thread-17::DEBUG::2015-09-02
16:07:54,130::misc::736::Storage.SamplingMethod::(__call__) Got in to
sampling method
Thread-17::DEBUG::2015-09-02 16:07:54,130::hba::56::Storage.HBA::(rescan)
Starting scan
Thread-17::DEBUG::2015-09-02 16:07:54,197::hba::62::Storage.HBA::(rescan)
Scan finished
Thread-17::DEBUG::2015-09-02
16:07:54,197::misc::743::Storage.SamplingMethod::(__call__) Returning last
result
Thread-17::DEBUG::2015-09-02
16:07:54,197::multipath::77::Storage.Misc.excCmd::(rescan) /usr/bin/sudo -n
/usr/sbin/multipath (cwd None)
Thread-17::DEBUG::2015-09-02
16:07:54,298::multipath::77::Storage.Misc.excCmd::(rescan) SUCCESS: <err> =
''; <rc> = 0
Thread-17::DEBUG::2015-09-02 16:07:54,299::utils::661::root::(execCmd)
/sbin/udevadm settle --timeout=5 (cwd None)
Thread-17::DEBUG::2015-09-02 16:07:54,307::utils::679::root::(execCmd)
SUCCESS: <err> = ''; <rc> = 0
Thread-17::DEBUG::2015-09-02
16:07:54,309::lvm::498::Storage.OperationMutex::(_invalidateAllPvs)
Operation 'lvm invalidate operation' got the operation mutex
Thread-17::DEBUG::2015-09-02
16:07:54,310::lvm::500::Storage.OperationMutex::(_invalidateAllPvs)
Operation 'lvm invalidate operation' released the operation mutex
Thread-17::DEBUG::2015-09-02
16:07:54,310::lvm::509::Storage.OperationMutex::(_invalidateAllVgs)
Operation 'lvm invalidate operation' got the operation mutex
Thread-17::DEBUG::2015-09-02
16:07:54,310::lvm::511::Storage.OperationMutex::(_invalidateAllVgs)
Operation 'lvm invalidate operation' released the operation mutex
Thread-17::DEBUG::2015-09-02
16:07:54,310::lvm::529::Storage.OperationMutex::(_invalidateAllLvs)
Operation 'lvm invalidate operation' got the operation mutex
Thread-17::DEBUG::2015-09-02
16:07:54,310::lvm::531::Storage.OperationMutex::(_invalidateAllLvs)
Operation 'lvm invalidate operation' released the operation mutex
Thread-17::DEBUG::2015-09-02
16:07:54,310::misc::743::Storage.SamplingMethod::(__call__) Returning last
result
Thread-17::DEBUG::2015-09-02
16:07:54,310::lvm::371::Storage.OperationMutex::(_reloadvgs) Operation 'lvm
reload operation' got the operation mutex
Thread-17::DEBUG::2015-09-02
16:07:54,312::lvm::291::Storage.Misc.excCmd::(cmd) /usr/bin/sudo -n
/usr/sbin/lvm vgs --config ' devices { preferred_names = ["^/dev/mapper/"]
ignore_suspended_devices=1 write_cache_state=0 disable_after_error_count=3
obtain_device_list_from_udev=0 filter = [
'\''a|/dev/mapper/33000000100000001|/dev/mapper/Hitachi_HDS721010DLE630_MSK523Y209VK0B|'\'',
'\''r|.*|'\'' ] } global { locking_type=1
prioritise_write_locks=1
wait_for_locks=1 use_lvmetad=0 } backup { retain_min = 50 retain_days =
0 } ' --noheadings --units b --nosuffix --separator '|'
--ignoreskippedcluster -o
uuid,name,attr,size,free,extent_size,extent_count,free_count,tags,vg_mda_size,vg_mda_free,lv_count,pv_count,pv_name
8b25f3be-7574-4f7a-8851-363129704e52 (cwd None)
Thread-17::DEBUG::2015-09-02
16:07:54,478::lvm::291::Storage.Misc.excCmd::(cmd) SUCCESS: <err> = '
WARNING: lvmetad is running but disabled. Restart lvmetad before enabling
it!\n'; <rc> = 0
Thread-17::DEBUG::2015-09-02
16:07:54,478::lvm::416::Storage.OperationMutex::(_reloadvgs) Operation 'lvm
reload operation' released the operation mutex
Thread-17::DEBUG::2015-09-02
16:07:54,479::persistentDict::192::Storage.PersistentDict::(__init__)
Created a persistent dict with LvMetadataRW backend
Thread-17::DEBUG::2015-09-02
16:07:54,479::blockSD::337::Storage.Misc.excCmd::(readlines) /usr/bin/dd
iflag=direct skip=0 bs=2048
if=/dev/8b25f3be-7574-4f7a-8851-363129704e52/metadata count=1 (cwd None)
Thread-17::DEBUG::2015-09-02
16:07:54,553::blockSD::337::Storage.Misc.excCmd::(readlines) SUCCESS: <err>
= '1+0 records in\n1+0 records out\n2048 bytes (2.0 kB) copied, 0.00107202
s, 1.9 MB/s\n'; <rc> = 0
Thread-17::DEBUG::2015-09-02
16:07:54,553::misc::260::Storage.Misc::(validateDDBytes) err: ['1+0 records
in', '1+0 records out', '2048 bytes (2.0 kB) copied, 0.00107202 s, 1.9
MB/s'], size: 2048
Thread-17::DEBUG::2015-09-02
16:07:54,553::persistentDict::234::Storage.PersistentDict::(refresh) read
lines (LvMetadataRW)=[]
Thread-17::DEBUG::2015-09-02
16:07:54,553::persistentDict::252::Storage.PersistentDict::(refresh) Empty
metadata
Thread-17::DEBUG::2015-09-02
16:07:54,553::persistentDict::192::Storage.PersistentDict::(__init__)
Created a persistent dict with VGTagMetadataRW backend
Thread-17::DEBUG::2015-09-02
16:07:54,554::lvm::504::Storage.OperationMutex::(_invalidatevgs) Operation
'lvm invalidate operation' got the operation mutex
Thread-17::DEBUG::2015-09-02
16:07:54,554::lvm::506::Storage.OperationMutex::(_invalidatevgs) Operation
'lvm invalidate operation' released the operation mutex
Thread-17::DEBUG::2015-09-02
16:07:54,554::lvm::514::Storage.OperationMutex::(_invalidatelvs) Operation
'lvm invalidate operation' got the operation mutex
Thread-17::DEBUG::2015-09-02
16:07:54,554::lvm::526::Storage.OperationMutex::(_invalidatelvs) Operation
'lvm invalidate operation' released the operation mutex
Thread-17::DEBUG::2015-09-02
16:07:54,554::lvm::371::Storage.OperationMutex::(_reloadvgs) Operation 'lvm
reload operation' got the operation mutex
Thread-17::DEBUG::2015-09-02
16:07:54,554::lvm::291::Storage.Misc.excCmd::(cmd) /usr/bin/sudo -n
/usr/sbin/lvm vgs --config ' devices { preferred_names = ["^/dev/mapper/"]
ignore_suspended_devices=1 write_cache_state=0 disable_after_error_count=3
obtain_device_list_from_udev=0 filter = [
'\''a|/dev/mapper/33000000100000001|/dev/mapper/Hitachi_HDS721010DLE630_MSK523Y209VK0B|'\'',
'\''r|.*|'\'' ] } global { locking_type=1
prioritise_write_locks=1
wait_for_locks=1 use_lvmetad=0 } backup { retain_min = 50 retain_days =
0 } ' --noheadings --units b --nosuffix --separator '|'
--ignoreskippedcluster -o
uuid,name,attr,size,free,extent_size,extent_count,free_count,tags,vg_mda_size,vg_mda_free,lv_count,pv_count,pv_name
8b25f3be-7574-4f7a-8851-363129704e52 (cwd None)
Thread-17::DEBUG::2015-09-02
16:07:54,685::lvm::291::Storage.Misc.excCmd::(cmd) SUCCESS: <err> = '
WARNING: lvmetad is running but disabled. Restart lvmetad before enabling
it!\n'; <rc> = 0
Thread-17::DEBUG::2015-09-02
16:07:54,686::lvm::416::Storage.OperationMutex::(_reloadvgs) Operation 'lvm
reload operation' released the operation mutex
Thread-17::DEBUG::2015-09-02
16:07:54,686::persistentDict::234::Storage.PersistentDict::(refresh) read
lines (VGTagMetadataRW)=['CLASS=Data', 'DESCRIPTION=hosted_storage',
'IOOPTIMEOUTSEC=10', 'LEASERETRIES=3', 'LEASETIMESEC=60',
'LOCKPOLICY=',
'LOCKRENEWALINTERVALSEC=5', 'LOGBLKSIZE=512', 'PHYBLKSIZE=4096',
'POOL_UUID=',
u'PV0=pv:33000000100000001,uuid:kTaQQh-4LCD-OghQ-cP5D-R7MM-aj6e-kTdQf0,pestart:0,pecount:397,mapoffset:0',
'ROLE=Regular', 'SDUUID=8b25f3be-7574-4f7a-8851-363129704e52',
'TYPE=ISCSI', 'VERSION=3',
'VGUUID=tJKiwH-Cn7v-QCxd-YQrg-MUxA-fbdC-kdga8m',
'_SHA_CKSUM=4a100ce5195650f43971d849835a8b3d8c0343da']
Thread-17::DEBUG::2015-09-02
16:07:54,687::resourceManager::421::Storage.ResourceManager::(registerNamespace)
Registering namespace '8b25f3be-7574-4f7a-8851-363129704e52_imageNS'
Thread-17::DEBUG::2015-09-02
16:07:54,687::resourceManager::421::Storage.ResourceManager::(registerNamespace)
Registering namespace '8b25f3be-7574-4f7a-8851-363129704e52_volumeNS'
Thread-17::DEBUG::2015-09-02
16:07:54,687::resourceManager::421::Storage.ResourceManager::(registerNamespace)
Registering namespace '8b25f3be-7574-4f7a-8851-363129704e52_lvmActivationNS'
Thread-17::DEBUG::2015-09-02
16:07:54,687::lvm::428::Storage.OperationMutex::(_reloadlvs) Operation 'lvm
reload operation' got the operation mutex
What should I do to bring the VM engine back?
2015-09-02 16:24 GMT+01:00 Simone Tiraboschi <stirabos(a)redhat.com>:
>
>
> On Wed, Sep 2, 2015 at 10:49 AM, wodel youchi <wodel.youchi(a)gmail.com>
> wrote:
>
>> I will try this afternoon to do this, but just to clarify something.
>>
>> the hosted_engine setup creates it's own DC the hosted_DC, which
>> contains the hosted engine storage domain, I am correct?
>>
>
> No, ovirt-hosted-engine-setup doesn't create a special datacenter. The
> default is to add the host to the Default datacenter in the default cluster.
> You could choose a different one from ovirt-hosted-engine-setup, simply
> import the hosted-engine storage domain in the datacenter of the cluster
> you selected.
>
> In setup there is a question like this:
> Local storage datacenter name is an internal name
> and currently will not be shown in engine's admin UI.
> Please enter local datacenter name
> which ask about 'Local storage datacenter' which is basically the
> description we were using for the storage pool.
>
>
>
>> if yes, where will I import the hostedengine storage domain, into the
>> default DC?
>>
>> 2015-09-02 8:47 GMT+01:00 Roy Golan <rgolan(a)redhat.com>:
>>
>>>
>>>
>>> On Wed, Sep 2, 2015 at 12:51 AM, wodel youchi <wodel.youchi(a)gmail.com>
>>> wrote:
>>>
>>>> I could finally terminate the installation, but still no vm engine on
>>>> webui
>>>>
>>>> I added a data domain, the default DC is up, but no engine VM.
>>>>
>>>
>>>
>>>
>>> Good now you need to import the HostedEngine storage domain. Try to go
>>> to
>>>
>>> *Storage -> Import Domain and put the path to the domain which you used
>>> in the hosted-engine setup.*
>>>
>>>
>>> *After the domain is imported, the engine will be imported
>>> automatically. *
>>>
>>> *This whole process will become automatic eventually. (patch is written
>>> currently)*
>>>
>>>
>>>>
>>>> 2015-09-01 21:22 GMT+01:00 wodel youchi <wodel.youchi(a)gmail.com>:
>>>>
>>>>> Something mounted on /rhev/data-center/mnt I'm not sure.
>>>>>
>>>>> there were directories, and under these directories there were other
>>>>> directories (dom_md, ha_agent, images), and under them there were
symbolic
>>>>> links to devices under /dev
>>>>> (ids, inbox, leases, etc...) the devices pointed to the lvm
>>>>> partitions created by the setup.
>>>>>
>>>>> but the mount command didn't show anything, unlike nfs, when I
used
>>>>> it the mount and df commands showed up the engine's VM mount
point.
>>>>>
>>>>>
>>>>> 2015-09-01 20:16 GMT+01:00 Simone Tiraboschi
<stirabos(a)redhat.com>:
>>>>>
>>>>>>
>>>>>>
>>>>>> On Tue, Sep 1, 2015 at 7:29 PM, wodel youchi
<wodel.youchi(a)gmail.com
>>>>>> > wrote:
>>>>>>
>>>>>>> Hi,
>>>>>>>
>>>>>>> After removing the -x from the sql files, the installation
>>>>>>> terminated successfully, but ...
>>>>>>>
>>>>>>> I had a problem with vdsm, and error about permission denied
with
>>>>>>> KVM module, so I restarted my machine.
>>>>>>> After the reboot the ovirt-ha-agent stops complaining about
the
>>>>>>> vm.conf file not present in /var/rum/ovirt-hosted-engine-ha
>>>>>>>
>>>>>>> And the mount command doesn't show any iscsi mount, the
disk is
>>>>>>> detected via fdisk -l
>>>>>>> the lvs command returns all logical volumes created.
>>>>>>>
>>>>>>> I think it's a mount problem, but since there are many
lv, I don't
>>>>>>> how to mount them manually.
>>>>>>>
>>>>>>
>>>>>> Do you have something mounted under /rhev/data-center/mnt ?
>>>>>> If not you probably hit this bug:
>>>>>>
https://bugzilla.redhat.com/1258465
>>>>>>
>>>>>>
>>>>>>
>>>>>>> LV VG
>>>>>>> Attr LSize Pool
Origin Data%
>>>>>>> Meta% Move
>>>>>>> Log Cpy%Sync Convert
>>>>>>>
>>>>>>> 3b894e23-429d-43bf-b6cd-6427a387799a
>>>>>>> 5445bbee-bb3a-4e6d-9614-a0c9378fe078 -wi-ao---- 128,00m
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> be78c0fd-52bf-445a-9555-64061029c2d9
>>>>>>> 5445bbee-bb3a-4e6d-9614-a0c9378fe078 -wi-a----- 1,00g
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> c9f74ffc-2eba-40a9-9c1c-f3b6d8e12657
>>>>>>> 5445bbee-bb3a-4e6d-9614-a0c9378fe078 -wi-a----- 40,00g
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> feede664-5754-4ca2-aeb3-af7aff32ed42
>>>>>>> 5445bbee-bb3a-4e6d-9614-a0c9378fe078 -wi-a----- 128,00m
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> ids
>>>>>>>
5445bbee-bb3a-4e6d-9614-a0c9378fe078
>>>>>>> -wi-ao---- 128,00m
>>>>>>>
>>>>>>>
>>>>>>> inbox
>>>>>>>
5445bbee-bb3a-4e6d-9614-a0c9378fe078
>>>>>>> -wi-a----- 128,00m
>>>>>>>
>>>>>>>
>>>>>>> leases
>>>>>>>
5445bbee-bb3a-4e6d-9614-a0c9378fe078
>>>>>>> -wi-a----- 2,00g
>>>>>>>
>>>>>>>
>>>>>>> master
>>>>>>>
5445bbee-bb3a-4e6d-9614-a0c9378fe078
>>>>>>> -wi-a----- 1,00g
>>>>>>>
>>>>>>>
>>>>>>> metadata
>>>>>>>
5445bbee-bb3a-4e6d-9614-a0c9378fe078 -wi-a-----
>>>>>>> 512,00m
>>>>>>>
>>>>>>>
>>>>>>> outbox
>>>>>>>
5445bbee-bb3a-4e6d-9614-a0c9378fe078
>>>>>>> -wi-a----- 128,00m
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> 2015-09-01 16:57 GMT+01:00 Simone Tiraboschi
<stirabos(a)redhat.com>:
>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> On Tue, Sep 1, 2015 at 5:08 PM, wodel youchi <
>>>>>>>> wodel.youchi(a)gmail.com> wrote:
>>>>>>>>
>>>>>>>>> Hi again,
>>>>>>>>>
>>>>>>>>> I tried with the snapshot repository, but I am having
this error
>>>>>>>>> while executing engine-setup
>>>>>>>>>
>>>>>>>>> [ INFO ] Creating/refreshing Engine database schema
>>>>>>>>> [ ERROR ] Failed to execute stage 'Misc
configuration': Command
>>>>>>>>> '/usr/share/ovirt-engine/dbscripts/schema.sh'
failed to execu
>>>>>>>>> te
>>>>>>>>> [ INFO ] DNF Performing DNF transaction rollback
>>>>>>>>> [ INFO ] Rolling back database schema
>>>>>>>>> [ INFO ] Clearing Engine database engine
>>>>>>>>> [ ERROR ] Engine database rollback failed: must be
owner of
>>>>>>>>> schema pg_catalog
>>>>>>>>> [ INFO ] Stage: Clean up
>>>>>>>>> Log file is located at
>>>>>>>>>
/var/log/ovirt-engine/setup/ovirt-engine-setup-20150901153202-w0ds25.log
>>>>>>>>> [ INFO ] Generating answer file
>>>>>>>>>
'/var/lib/ovirt-engine/setup/answers/20150901153939-setup.conf'
>>>>>>>>> [ INFO ] Stage: Pre-termination
>>>>>>>>> [ INFO ] Stage: Termination
>>>>>>>>> [ ERROR ] Execution of setup failed
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> and in the deployement log I have these errors
>>>>>>>>>
>>>>>>>>> Saving custom users permissions on database
objects...
>>>>>>>>> upgrade script detected a change in Config, View or
Stored
>>>>>>>>> Procedure...
>>>>>>>>> Running upgrade shell script
>>>>>>>>>
'/usr/share/ovirt-engine/dbscripts/upgrade/pre_upgrade/0000_config.sql'...
>>>>>>>>> Running upgrade shell script
>>>>>>>>>
'/usr/share/ovirt-engine/dbscripts/upgrade/pre_upgrade/0010_custom.sql'...
>>>>>>>>> Running upgrade shell script
>>>>>>>>>
'/usr/share/ovirt-engine/dbscripts/upgrade/pre_upgrade/0020_add_materialized_views_table.sql'...
>>>>>>>>> Running upgrade shell script
>>>>>>>>>
'/usr/share/ovirt-engine/dbscripts/upgrade/pre_upgrade/0030_materialized_views_extensions.sql'...
>>>>>>>>> Running upgrade shell script
>>>>>>>>>
'/usr/share/ovirt-engine/dbscripts/pre_upgrade/0040_extend_installed_by_column.sql'...
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> 2015-09-01 15:39:35 DEBUG
>>>>>>>>>
otopi.plugins.ovirt_engine_setup.ovirt_engine.db.schema plugin.execute:941
>>>>>>>>> execute-output:
['/usr/share/ovirt-engine/dbscripts/schema.sh', '-s',
>>>>>>>>> 'localhost', '-p', '5432',
'-u', 'engine', '-d', 'engine', '-l',
'/var/log
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
*/ovirt-engine/setup/ovirt-engine-setup-20150901153202-w0ds25.log',
>>>>>>>>> '-c', 'apply']
>>>>>>>>>
stderr:/usr/share/ovirt-engine/dbscripts/upgrade/pre_upgrade/0000_config.sql:
>>>>>>>>> ligne 1: /bin : is a
>>>>>>>>>
directory/usr/share/ovirt-engine/dbscripts/upgrade/pre_upgrade/0000_config.sql:
>>>>>>>>> ligne 2: DATABASE : command not
>>>>>>>>>
found/usr/share/ovirt-engine/dbscripts/upgrade/pre_upgrade/0000_config.sql:
>>>>>>>>> ligne 4: This : command not
>>>>>>>>>
found/usr/share/ovirt-engine/dbscripts/upgrade/pre_upgrade/0000_config.sql:
>>>>>>>>> ligne 5: The : command not
>>>>>>>>>
found/usr/share/ovirt-engine/dbscripts/upgrade/pre_upgrade/0000_config.sql:
>>>>>>>>> ligne 6: Add : command not
>>>>>>>>>
found/usr/share/ovirt-engine/dbscripts/upgrade/pre_upgrade/0000_config.sql:
>>>>>>>>> ligne 7: syntax error near the symbole unexpected «
(
>>>>>>>>>
»/usr/share/ovirt-engine/dbscripts/upgrade/pre_upgrade/0000_config.sql:
>>>>>>>>> ligne 7: ` Update section (w/o overriding current
>>>>>>>>>
value)'/usr/share/ovirt-engine/dbscripts/upgrade/pre_upgrade/0010_custom.sql:
>>>>>>>>> ligne 1: /bin : is a
>>>>>>>>>
directory/usr/share/ovirt-engine/dbscripts/upgrade/pre_upgrade/0010_custom.sql:
>>>>>>>>> ligne 2: Currently : command not
>>>>>>>>>
found/usr/share/ovirt-engine/dbscripts/upgrade/pre_upgrade/0010_custom.sql:
>>>>>>>>> ligne 3: This : command not
>>>>>>>>>
found/usr/share/ovirt-engine/dbscripts/upgrade/pre_upgrade/0010_custom.sql:
>>>>>>>>> ligne 4: This : command not
>>>>>>>>>
found/usr/share/ovirt-engine/dbscripts/upgrade/pre_upgrade/0010_custom.sql:
>>>>>>>>> ligne 5: So, : command not
>>>>>>>>>
found/usr/share/ovirt-engine/dbscripts/upgrade/pre_upgrade/0010_custom.sql:
>>>>>>>>> ligne 6: Since : command not
>>>>>>>>>
found/usr/share/ovirt-engine/dbscripts/upgrade/pre_upgrade/0010_custom.sql:
>>>>>>>>> ligne 7: bin/ : is a
>>>>>>>>>
directory/usr/share/ovirt-engine/dbscripts/upgrade/pre_upgrade/0010_custom.sql:
>>>>>>>>> ligne 9: update : command not
>>>>>>>>>
found/usr/share/ovirt-engine/dbscripts/upgrade/pre_upgrade/0010_custom.sql:
>>>>>>>>> ligne 10: *
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> *syntax error near the symbole unexpected « (
>>>>>>>>>
»/usr/share/ovirt-engine/dbscripts/upgrade/pre_upgrade/0010_custom.sql:
>>>>>>>>> ligne 10: `and exists(select 1 from schema_version
where version =
>>>>>>>>> '03010250' and current =
>>>>>>>>>
true);'/usr/share/ovirt-engine/dbscripts/upgrade/pre_upgrade/0020_add_materialized_views_table.sql:
>>>>>>>>> ligne 1: -- : command not
>>>>>>>>>
found/usr/share/ovirt-engine/dbscripts/upgrade/pre_upgrade/0020_add_materialized_views_table.sql:
>>>>>>>>> ligne 2: **syntax error near the symbole unexpected «
( »*
>>>>>>>>>
/usr/share/ovirt-engine/dbscripts/upgrade/pre_upgrade/0020_add_materialized_views_table.sql:
>>>>>>>>> ligne 2: `CREATE FUNCTION
__temp__0030_add_materialized_views_table()'
>>>>>>>>>
/usr/share/ovirt-engine/dbscripts/upgrade/pre_upgrade/0030_materialized_views_extensions.sql:
>>>>>>>>> ligne 1: -- : commande introuvable
>>>>>>>>>
/usr/share/ovirt-engine/dbscripts/upgrade/pre_upgrade/0030_materialized_views_extensions.sql:
>>>>>>>>> ligne 2: erreur de syntaxe près du symbole inattendu
« ( »
>>>>>>>>>
/usr/share/ovirt-engine/dbscripts/upgrade/pre_upgrade/0030_materialized_views_extensions.sql:
>>>>>>>>> ligne 2: `select
fn_db_add_column('materialized_views',
>>>>>>>>> 'min_refresh_rate_in_sec', 'int default
0');'
>>>>>>>>>
/usr/share/ovirt-engine/dbscripts/upgrade/pre_upgrade/0040_extend_installed_by_column.sql:
>>>>>>>>> ligne 1: erreur de syntaxe près du symbole inattendu
« ( »
>>>>>>>>>
/usr/share/ovirt-engine/dbscripts/upgrade/pre_upgrade/0040_extend_installed_by_column.sql:
>>>>>>>>> ligne 1: `ALTER TABLE schema_version ALTER COLUMN
installed_by TYPE
>>>>>>>>> varchar(63);'
>>>>>>>>> 2015-09-01 15:39:35 DEBUG otopi.context
>>>>>>>>> context._executeMethod:156 method exception
>>>>>>>>> Traceback (most recent call last):
>>>>>>>>> File
"/usr/lib/python2.7/site-packages/otopi/context.py", line
>>>>>>>>> 146, in _executeMethod
>>>>>>>>> method['method']()
>>>>>>>>> File
>>>>>>>>>
"/usr/share/ovirt-engine/setup/bin/../plugins/ovirt-engine-setup/ovirt-engine/db/schema.py",
>>>>>>>>> line 291, in _misc
>>>>>>>>> oenginecons.EngineDBEnv.PGPASS_FILE
>>>>>>>>> File
"/usr/lib/python2.7/site-packages/otopi/plugin.py", line
>>>>>>>>> 946, in execute
>>>>>>>>> command=args[0],
>>>>>>>>> RuntimeError: Command
>>>>>>>>> '/usr/share/ovirt-engine/dbscripts/schema.sh'
failed to execute
>>>>>>>>> 2015-09-01 15:39:35 ERROR otopi.context
>>>>>>>>> context._executeMethod:165 Failed to execute stage
'Misc configuration':
>>>>>>>>> Command
'/usr/share/ovirt-engine/dbscripts/schema.sh' failed to execute
>>>>>>>>> 2015-09-01 15:39:35 DEBUG otopi.transaction
transaction.abort:134
>>>>>>>>> aborting 'DNF Transaction'
>>>>>>>>> 2015-09-01 15:39:35 DEBUG
>>>>>>>>> otopi.plugins.otopi.packagers.dnfpackager
dnfpackager.verbose:90 DNF
>>>>>>>>> Closing transaction with rollback
>>>>>>>>> 2015-09-01 15:39:35 INFO
>>>>>>>>> otopi.plugins.otopi.packagers.dnfpackager
dnfpackager.info:94
>>>>>>>>> DNF Performing DNF transaction rollback
>>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> It was an issue with package building: all the sql files
where
>>>>>>>> executable and so the issue.
>>>>>>>> We fixed it and tomorrow build should be OK. If you
prefer to
>>>>>>>> continue right now simply recursively remove the x
attribute on each sql
>>>>>>>> file under /usr/share/ovirt-engine/dbscripts
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>>>
>>>>>>>>> 2015-09-01 13:04 GMT+01:00 Simone Tiraboschi
<stirabos(a)redhat.com
>>>>>>>>> >:
>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> On Tue, Sep 1, 2015 at 12:40 PM, Yedidyah Bar
David <
>>>>>>>>>> didi(a)redhat.com> wrote:
>>>>>>>>>>
>>>>>>>>>>> On Tue, Sep 1, 2015 at 1:25 PM, wodel youchi
<
>>>>>>>>>>> wodel.youchi(a)gmail.com> wrote:
>>>>>>>>>>> > Hi,
>>>>>>>>>>> >
>>>>>>>>>>> > I am using the repo of the 3.6 version
>>>>>>>>>>> >
(
http://plain.resources.ovirt.org/pub/ovirt-3.6-pre/rpm/fc22/
>>>>>>>>>>> )
>>>>>>>>>>> >
>>>>>>>>>>> > I installed the
ovirt-hosted-engine-setup with it's
>>>>>>>>>>> dependencies,and the
>>>>>>>>>>> > ovirt-hosted-engine-ha package is one of
them.
>>>>>>>>>>> >
>>>>>>>>>>> > Correction: The problem with this
version
>>>>>>>>>>> >
>>>>>>>>>>>
ovirt-hosted-engine-ha-1.3.0-0.0.master.20150819082341.20150819082338.git183a4ff.fc22.noarch.rpm,
>>>>>>>>>>> > is that after the installation is done,
the service
>>>>>>>>>>> ovirt-ha-agent crashes
>>>>>>>>>>> > after being started, see the bug :
>>>>>>>>>>> >
>>>>>>>>>>> >
https://bugzilla.redhat.com/show_bug.cgi?id=1254745
>>>>>>>>>>> >
>>>>>>>>>>> > A new version was produced, I downloaded
it manually a few
>>>>>>>>>>> days ago, this is
>>>>>>>>>>> > it :
>>>>>>>>>>> >
>>>>>>>>>>>
ovirt-hosted-engine-ha-1.3.0-0.0.master.20150820064645.20150820064642.git02529e0.fc22.noarch.rpm
>>>>>>>>>>> >
>>>>>>>>>>> > This one did correct the problem, but
it's not present
>>>>>>>>>>> anymore on the
>>>>>>>>>>> > repository.
>>>>>>>>>>>
>>>>>>>>>>> Was this on ovirt-3.6-pre?
>>>>>>>>>>>
>>>>>>>>>>> ovirt-3.6-snapshot has a newer version.
>>>>>>>>>>>
>>>>>>>>>>> >
>>>>>>>>>>> > For Simone: yes I did added an NFS4 data
domain, but no
>>>>>>>>>>> success so far, no
>>>>>>>>>>> > VM engine present.
>>>>>>>>>>> >
>>>>>>>>>>> > Regards.
>>>>>>>>>>> >
>>>>>>>>>>> > 2015-09-01 11:15 GMT+01:00 Simone
Tiraboschi <
>>>>>>>>>>> stirabos(a)redhat.com>:
>>>>>>>>>>> >>
>>>>>>>>>>> >>
>>>>>>>>>>> >>
>>>>>>>>>>> >> On Tue, Sep 1, 2015 at 11:46 AM,
Yedidyah Bar David <
>>>>>>>>>>> didi(a)redhat.com>
>>>>>>>>>>> >> wrote:
>>>>>>>>>>> >>>
>>>>>>>>>>> >>> On Tue, Sep 1, 2015 at 11:25 AM,
wodel youchi <
>>>>>>>>>>> wodel.youchi(a)gmail.com>
>>>>>>>>>>> >>> wrote:
>>>>>>>>>>> >>> > Hi,
>>>>>>>>>>> >>> >
>>>>>>>>>>> >>> > Another test of ovirt
hosted-engine on FC22 using ovirt
>>>>>>>>>>> 3.6 Beta3.
>>>>>>>>>>> >>> > VM engine is also a FC22
>>>>>>>>>>> >>> >
>>>>>>>>>>> >>> > Problem:
>>>>>>>>>>> >>> > - No VM engine on webui
>>>>>>>>>>> >>>
>>>>>>>>>>> >>> This is still not supported,
see/follow [1].
>>>>>>>>>>> >>>
>>>>>>>>>>> >>> [1]
https://bugzilla.redhat.com/show_bug.cgi?id=1224889
>>>>>>>>>>> >>
>>>>>>>>>>> >>
>>>>>>>>>>> >> ? :-)
>>>>>>>>>>>
>>>>>>>>>>> Sorry :-(
>>>>>>>>>>>
>>>>>>>>>>>
https://bugzilla.redhat.com/show_bug.cgi?id=1160094
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> This is just about editing the VM from the web
GUI but in order
>>>>>>>>>> to be able to edit the engine VM you should be
able at least to find the
>>>>>>>>>> engine VM in the engine as it was also in 3.5
>>>>>>>>>>
>>>>>>>>>> I'll try to reproduce verifying another
patch
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>> I see that all patches there are merged, but
bug is in POST.
>>>>>>>>>>>
>>>>>>>>>>> >>
>>>>>>>>>>> >> Did you also try adding an
additional storage domain for
>>>>>>>>>>> regular VMs?
>>>>>>>>>>> >> engine-VM will be shown on the
engine only when you add at
>>>>>>>>>>> least one
>>>>>>>>>>> >> additional storage domain for
regulars VM and the whole
>>>>>>>>>>> datacenter goes up:
>>>>>>>>>>> >>
https://bugzilla.redhat.com/show_bug.cgi?id=1222010#c1
>>>>>>>>>>> >>
>>>>>>>>>>> >>
>>>>>>>>>>> >>>
>>>>>>>>>>> >>> > Test Environment
>>>>>>>>>>> >>> > Just two machines:
>>>>>>>>>>> >>> >
>>>>>>>>>>> >>> > 1 - Machine 1 used as
storage :
>>>>>>>>>>> >>> > - iscsi target with a
raw file for the VM engine
>>>>>>>>>>> storage
>>>>>>>>>>> >>> > - NFS4 for other data
domains
>>>>>>>>>>> >>> >
>>>>>>>>>>> >>> > 2 - Machine 2 used as
hypervisor
>>>>>>>>>>> >>> >
>>>>>>>>>>> >>> >
>>>>>>>>>>> >>> > The installation went
without problem, but as always, the
>>>>>>>>>>> VM engine is
>>>>>>>>>>> >>> > not
>>>>>>>>>>> >>> > present on the webui.
>>>>>>>>>>> >>> >
>>>>>>>>>>> >>> >
>>>>>>>>>>> >>> > PS:
>>>>>>>>>>> >>> > 1- I gave the VM engine
just 2Gb of memory since I don't
>>>>>>>>>>> have too much
>>>>>>>>>>> >>> > RAM
>>>>>>>>>>> >>> > on hypervisor, could that
be the cause of the problem?
>>>>>>>>>>> >>>
>>>>>>>>>>> >>> Shouldn't be related
>>>>>>>>>>> >>>
>>>>>>>>>>> >>> > 2- This version of
>>>>>>>>>>> >>> >
>>>>>>>>>>> >>> >
>>>>>>>>>>>
ovirt-hosted-engine-ha-1.3.0-0.0.master.20150424113926.20150424113923.git7c14f4c.fc22.noarch.rpm
>>>>>>>>>>> >>> > package is causing the
ovirt-ha-agent to crash, it was
>>>>>>>>>>> replaced with
>>>>>>>>>>> >>> > another
>>>>>>>>>>> >>> > which I still have
>>>>>>>>>>> >>> >
>>>>>>>>>>> >>> >
>>>>>>>>>>>
ovirt-hosted-engine-ha-1.3.0-0.0.master.20150820064645.20150820064642.git02529e0.fc22.noarch.rpm,
>>>>>>>>>>> >>> > but it's not present on
the repository, I had to update
>>>>>>>>>>> the package
>>>>>>>>>>> >>> > manually
>>>>>>>>>>> >>> > at the end of
ovirt-hosted-engine-setup installation.
>>>>>>>>>>> >>>
>>>>>>>>>>> >>> Not sure I follow.
>>>>>>>>>>> >>>
>>>>>>>>>>> >>> What exact repo was used?
>>>>>>>>>>> >>>
>>>>>>>>>>> >>> hosted-engine --deploy does not
update/install packages for
>>>>>>>>>>> you (as
>>>>>>>>>>> >>> does engine-setup),
>>>>>>>>>>> >>> it's up to you to make sure
what you want/need is installed
>>>>>>>>>>> prior to
>>>>>>>>>>> >>> running it.
>>>>>>>>>>> >>>
>>>>>>>>>>> >>> Best,
>>>>>>>>>>> >>> --
>>>>>>>>>>> >>> Didi
>>>>>>>>>>> >>>
_______________________________________________
>>>>>>>>>>> >>> Users mailing list
>>>>>>>>>>> >>> Users(a)ovirt.org
>>>>>>>>>>> >>>
http://lists.ovirt.org/mailman/listinfo/users
>>>>>>>>>>> >>
>>>>>>>>>>> >>
>>>>>>>>>>> >
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> --
>>>>>>>>>>> Didi
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>>
>>>>>>>
>>>>>>
>>>>>
>>>>
>>>
>>
>