Thanks,
but before that I stuck again with the storage of the VM engine not
detected after reboot.
the /rhev is populated, but ovirt-ha-agent crashes with
MainThread::INFO::2015-09-02
16:12:20,261::brokerlink::129::ovirt_hosted_engine_ha.lib.brokerlink.BrokerLink::(start_monitor)
Starting monitor engine-health, options {'use_ssl': 'true',
'vm_uuid':
'ab1dc1a9-b6e9-4890-8485-1019da2f328f', 'address': '0
'}
MainThread::INFO::2015-09-02
16:12:20,283::brokerlink::140::ovirt_hosted_engine_ha.lib.brokerlink.BrokerLink::(start_monitor)
Success, id 139994237094736
MainThread::INFO::2015-09-02
16:12:20,702::brokerlink::178::ovirt_hosted_engine_ha.lib.brokerlink.BrokerLink::(set_storage_do
main) Success, id 139994236985168
MainThread::INFO::2015-09-02
16:12:20,702::hosted_engine::574::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_ini
tialize_broker) Broker initialized, all submonitors started
MainThread::INFO::2015-09-02
16:12:20,799::hosted_engine::678::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_ini
tialize_sanlock) Ensuring lease for lockspace hosted-engine, host id 1 is
acquired (file: /var/run/vdsm/storage/8b25f3be-7574
-4f7a-8851-363129704e52/a44d1302-3165-4632-9d99-3e035dfc3ac7/0f260ab0-3631-4c71-b332-c6c7f67f7342)
MainThread::INFO::2015-09-02
16:12:20,800::hosted_engine::401::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(star
t_monitoring) Reloading vm.conf from the shared storage domain
MainThread::ERROR::2015-09-02
16:12:20,927::agent::201::ovirt_hosted_engine_ha.agent.agent.Agent::(_run_agent)
*Error: ''Configuration value not found:
file=/var/run/ovirt-hosted-engine-ha/vm.conf, key=memSize'' - trying to
restart agent *
MainThread::WARNING::2015-09-02
16:12:25,932::agent::204::ovirt_hosted_engine_ha.agent.agent.Agent::(_run_agent)
Restarting a
gent, attempt '9'
MainThread::ERROR::2015-09-02
16:12:25,933::agent::206::ovirt_hosted_engine_ha.agent.agent.Agent::(_run_agent)
Too many error
s occurred, giving up. Please review the log and consider filing a bug.
MainThread::INFO::2015-09-02
16:12:25,933::agent::143::ovirt_hosted_engine_ha.agent.agent.Agent::(run)
Agent shutting down
I restared vdsm ha-agent and broker-agent wihtout success
When executed
[root@noveria ~]# hosted-engine --vm-status
You must run deploy first
I got this
[root@noveria ~]# tree /var/run/vdsm/storage/
/var/run/vdsm/storage/
└── 8b25f3be-7574-4f7a-8851-363129704e52
├── 8e49032f-680b-40c2-b422-80d86dc7beda
│ └── f05762e5-e8cd-45e7-ac19-303c1ade79d1 ->
/dev/8b25f3be-7574-4f7a-8851-363129704e52/f05762e5-e8cd-45e7-ac19-303c1
ade79d1
├── a44d1302-3165-4632-9d99-3e035dfc3ac7
│ └── 0f260ab0-3631-4c71-b332-c6c7f67f7342 ->
/dev/8b25f3be-7574-4f7a-8851-363129704e52/0f260ab0-3631-4c71-b332-c6c7f
67f7342
├── a5475e57-c6f5-4dc5-a3f2-7fb782d613a7
│ └── ae352fab-7477-4376-aa27-04c321b4fbd1 ->
/dev/8b25f3be-7574-4f7a-8851-363129704e52/ae352fab-7477-4376-aa27-04c32
1b4fbd1
└── bf3bdae1-7318-4443-a19b-7371de30b982
└── cbb10cf0-9600-465e-aed9-412f7157706b ->
/dev/8b25f3be-7574-4f7a-8851-363129704e52/cbb10cf0-9600-465e-aed9-412f7
157706b
and this
[root@noveria rhev]# tree
.
└── data-center
├── 00000001-0001-0001-0001-000000000221
└── mnt
├── blockSD
│ └── 8b25f3be-7574-4f7a-8851-363129704e52
│ ├── dom_md
│ │ ├── ids -> /dev/8b25f3be-7574-4f7a-8851-363129704e52/ids
│ │ ├── inbox ->
/dev/8b25f3be-7574-4f7a-8851-363129704e52/inbox
│ │ ├── leases ->
/dev/8b25f3be-7574-4f7a-8851-363129704e52/leases
│ │ ├── master ->
/dev/8b25f3be-7574-4f7a-8851-363129704e52/master
│ │ ├── metadata ->
/dev/8b25f3be-7574-4f7a-8851-363129704e52/metadata
│ │ └── outbox ->
/dev/8b25f3be-7574-4f7a-8851-363129704e52/outbox
│ ├── ha_agent
│ │ ├── hosted-engine.lockspace ->
/var/run/vdsm/storage/8b25f3be-7574-4f7a-8851-363129704e52/a44d1302-3165-4
632-9d99-3e035dfc3ac7/0f260ab0-3631-4c71-b332-c6c7f67f7342
│ │ └── hosted-engine.metadata ->
/var/run/vdsm/storage/8b25f3be-7574-4f7a-8851-363129704e52/8e49032f-680b-40
c2-b422-80d86dc7beda/f05762e5-e8cd-45e7-ac19-303c1ade79d1
│ └── images
│ ├── 8e49032f-680b-40c2-b422-80d86dc7beda
│ │ └── f05762e5-e8cd-45e7-ac19-303c1ade79d1 ->
/dev/8b25f3be-7574-4f7a-8851-363129704e52/f05762e5-e8cd-4
5e7-ac19-303c1ade79d1
│ ├── a44d1302-3165-4632-9d99-3e035dfc3ac7
│ │ └── 0f260ab0-3631-4c71-b332-c6c7f67f7342 ->
/dev/8b25f3be-7574-4f7a-8851-363129704e52/0f260ab0-3631-4
c71-b332-c6c7f67f7342
│ ├── a5475e57-c6f5-4dc5-a3f2-7fb782d613a7
│ │ └── ae352fab-7477-4376-aa27-04c321b4fbd1 ->
/dev/8b25f3be-7574-4f7a-8851-363129704e52/ae352fab-7477-4
376-aa27-04c321b4fbd1
│ └── bf3bdae1-7318-4443-a19b-7371de30b982
│ └── cbb10cf0-9600-465e-aed9-412f7157706b ->
/dev/8b25f3be-7574-4f7a-8851-363129704e52/cbb10cf0-9600-4
65e-aed9-412f7157706b
├── openSuse.wodel.wd:_nvms
└── _var_lib_ovirt-hosted-engine-setup_tmp2fNoEf
Here I did find some symblic links blinking (not present) like this one
hosted-engine.metadata ->
/var/run/vdsm/storage/8b25f3be-7574-4f7a-8851-363129704e52/8e49032f-680b-40
c2-b422-80d86dc7beda/*f05762e5-e8cd-45e7-ac19-303c1ade79d1*
the lvscan command showed that the lv concerned is inactive, is this
correct?
[root@noveria ~]# lvscan
File descriptor 9 (/dev/dri/card0) leaked on lvscan invocation. Parent PID
2935: bash
ACTIVE '/dev/8b25f3be-7574-4f7a-8851-363129704e52/metadata'
[512,00 MiB] inherit
ACTIVE '/dev/8b25f3be-7574-4f7a-8851-363129704e52/outbox'
[128,00 MiB] inherit
ACTIVE '/dev/8b25f3be-7574-4f7a-8851-363129704e52/leases' [2,00
GiB] inherit
ACTIVE '/dev/8b25f3be-7574-4f7a-8851-363129704e52/ids' [128,00
MiB] inherit
ACTIVE '/dev/8b25f3be-7574-4f7a-8851-363129704e52/inbox'
[128,00 MiB] inherit
ACTIVE '/dev/8b25f3be-7574-4f7a-8851-363129704e52/master' [1,00
GiB] inherit
inactive
'/dev/8b25f3be-7574-4f7a-8851-363129704e52/ae352fab-7477-4376-aa27-04c321b4fbd1'
[1,00 GiB] inherit
ACTIVE
'/dev/8b25f3be-7574-4f7a-8851-363129704e52/0f260ab0-3631-4c71-b332-c6c7f67f7342'
[128,00 MiB] inherit
*inactive
'/dev/8b25f3be-7574-4f7a-8851-363129704e52/f05762e5-e8cd-45e7-ac19-303c1ade79d1'
[128,00 MiB] inherit *
inactive
'/dev/8b25f3be-7574-4f7a-8851-363129704e52/cbb10cf0-9600-465e-aed9-412f7157706b'
[40,00 GiB] inherit
and this
[root@noveria ~]# vdsClient -s 0 prepareImage
"00000000-0000-0000-0000-000000000000"
"8b25f3be-7574-4f7a-8851-363129704e52" "
bf3bdae1-7318-4443-a19b-7371de30b982"
"cbb10cf0-9600-465e-aed9-412f7157706b"
{'domainID': '8b25f3be-7574-4f7a-8851-363129704e52',
'imageID': 'bf3bdae1-7318-4443-a19b-7371de30b982',
'leaseOffset': 112197632,
'leasePath': '/dev/8b25f3be-7574-4f7a-8851-363129704e52/leases',
'path':
'/rhev/data-center/mnt/blockSD/8b25f3be-7574-4f7a-8851-363129704e52/images/bf3bdae1-7318-4443-a19b-7371de30b982/cbb1
0cf0-9600-465e-aed9-412f7157706b',
'volType': 'path',
'volumeID': 'cbb10cf0-9600-465e-aed9-412f7157706b'}
and
[root@noveria ~]# vdsClient -s 0 getStorageDomainInfo
8b25f3be-7574-4f7a-8851-363129704e52
uuid = 8b25f3be-7574-4f7a-8851-363129704e52
vguuid = tJKiwH-Cn7v-QCxd-YQrg-MUxA-fbdC-kdga8m
state = OK
version = 3
role = Regular
type = ISCSI
class = Data
pool = []
name = hosted_storage
[root@noveria ~]# lvs
File descriptor 9 (/dev/dri/card0) leaked on lvs invocation. Parent PID
3105: bash
LV VG
Attr LSize Pool Origin Data%
Meta% Move
Log Cpy%Sync Convert
0f260ab0-3631-4c71-b332-c6c7f67f7342 8b25f3be-7574-4f7a-8851-363129704e52
-wi-ao---- 128,00m
ae352fab-7477-4376-aa27-04c321b4fbd1 8b25f3be-7574-4f7a-8851-363129704e52
-wi------- 1,00g
cbb10cf0-9600-465e-aed9-412f7157706b 8b25f3be-7574-4f7a-8851-363129704e52
-wi-a----- 40,00g
f05762e5-e8cd-45e7-ac19-303c1ade79d1 8b25f3be-7574-4f7a-8851-363129704e52
-wi------- 128,00m
ids 8b25f3be-7574-4f7a-8851-363129704e52
-wi-a----- 128,00m
inbox 8b25f3be-7574-4f7a-8851-363129704e52
-wi-a----- 128,00m
leases 8b25f3be-7574-4f7a-8851-363129704e52
-wi-a----- 2,00g
master 8b25f3be-7574-4f7a-8851-363129704e52
-wi-a----- 1,00g
metadata 8b25f3be-7574-4f7a-8851-363129704e52
-wi-a----- 512,00m
outbox 8b25f3be-7574-4f7a-8851-363129704e52
-wi-a----- 128,00m
VDSM logs doesn't show me anything
MainThread::INFO::2015-09-01 23:34:49,551::vdsm::166::vds::(run)
<WorkerThread(Thread-4, started daemon 139990108333824)>
MainThread::INFO::2015-09-01 23:34:49,552::vdsm::166::vds::(run)
<WorkerThread(Thread-3, started daemon 139990116726528)>
MainThread::INFO::2015-09-02 16:07:49,510::vdsm::156::vds::(run) (PID:
1554) I am the actual vdsm 4.17.3-12.git7288ef7.fc22 noveria.wodel.wd
(4.1.6-200.fc22.x86_64)
MainThread::DEBUG::2015-09-02
16:07:49,524::resourceManager::421::Storage.ResourceManager::(registerNamespace)
Registering namespace 'Storage'
MainThread::DEBUG::2015-09-02
16:07:49,524::threadPool::29::Storage.ThreadPool::(__init__) Enter -
numThreads: 10, waitTimeout: 3, maxTasks: 500
MainThread::DEBUG::2015-09-02
16:07:49,526::fileUtils::143::Storage.fileUtils::(createdir) Creating
directory: /rhev/data-center/mnt mode: None
MainThread::WARNING::2015-09-02
16:07:49,526::fileUtils::152::Storage.fileUtils::(createdir) Dir
/rhev/data-center/mnt already exists
MainThread::DEBUG::2015-09-02
16:07:49,564::hsm::403::Storage.Misc.excCmd::(__validateLvmLockingType)
/usr/bin/sudo -n /usr/sbin/lvm dumpconfig global/locking_type (cwd None)
MainThread::DEBUG::2015-09-02
16:07:49,611::hsm::403::Storage.Misc.excCmd::(__validateLvmLockingType)
SUCCESS: <err> = ''; <rc> = 0
MainThread::DEBUG::2015-09-02
16:07:49,611::hsm::427::Storage.HSM::(__cleanStorageRepository) Started
cleaning storage repository at '/rhev/data-center'
MainThread::DEBUG::2015-09-02
16:07:49,614::hsm::459::Storage.HSM::(__cleanStorageRepository) White list:
['/rhev/data-center/hsm-tasks', '/rhev/data-center/hsm-tasks/*',
'/rhev/data-center/mnt']
MainThread::DEBUG::2015-09-02
16:07:49,614::hsm::460::Storage.HSM::(__cleanStorageRepository) Mount list:
[]
MainThread::DEBUG::2015-09-02
16:07:49,614::hsm::462::Storage.HSM::(__cleanStorageRepository) Cleaning
leftovers
MainThread::DEBUG::2015-09-02
16:07:49,615::hsm::505::Storage.HSM::(__cleanStorageRepository) Finished
cleaning storage repository at '/rhev/data-center'
storageRefresh::DEBUG::2015-09-02
16:07:49,616::misc::733::Storage.SamplingMethod::(__call__) Trying to enter
sampling method (storage.sdc.refreshStorage)
MainThread::INFO::2015-09-02
16:07:49,617::dispatcher::46::Storage.Dispatcher::(__init__) Starting
StorageDispatcher...
storageRefresh::DEBUG::2015-09-02
16:07:49,620::misc::736::Storage.SamplingMethod::(__call__) Got in to
sampling method
storageRefresh::DEBUG::2015-09-02
16:07:49,792::misc::733::Storage.SamplingMethod::(__call__) Trying to enter
sampling method (storage.iscsi.rescan)
storageRefresh::DEBUG::2015-09-02
16:07:49,793::misc::736::Storage.SamplingMethod::(__call__) Got in to
sampling method
storageRefresh::DEBUG::2015-09-02
16:07:49,793::iscsi::431::Storage.ISCSI::(rescan) Performing SCSI scan,
this will take up to 30 seconds
storageRefresh::DEBUG::2015-09-02
16:07:49,924::iscsiadm::97::Storage.Misc.excCmd::(_runCmd) /usr/bin/sudo -n
/sbin/iscsiadm -m session -R (cwd None)
MainThread::DEBUG::2015-09-02
16:07:49,924::task::595::Storage.TaskManager.Task::(_updateState)
Task=`68d01d7d-b426-4465-829e-174e2cb47e9e`::moving from state init ->
state preparing
MainThread::INFO::2015-09-02
16:07:49,924::logUtils::48::dispatcher::(wrapper) Run and protect:
registerDomainStateChangeCallback(callbackFunc=<functools.partial object at
0x7fc2f03fa6d8>)
MainThread::INFO::2015-09-02
16:07:49,924::logUtils::51::dispatcher::(wrapper) Run and protect:
registerDomainStateChangeCallback, Return response: None
MainThread::DEBUG::2015-09-02
16:07:49,927::task::1191::Storage.TaskManager.Task::(prepare)
Task=`68d01d7d-b426-4465-829e-174e2cb47e9e`::finished: None
MainThread::DEBUG::2015-09-02
16:07:49,927::task::595::Storage.TaskManager.Task::(_updateState)
Task=`68d01d7d-b426-4465-829e-174e2cb47e9e`::moving from state preparing ->
state finished
MainThread::DEBUG::2015-09-02
16:07:49,927::resourceManager::940::Storage.ResourceManager.Owner::(releaseAll)
Owner.releaseAll requests {} resources {}
MainThread::DEBUG::2015-09-02
16:07:49,927::resourceManager::977::Storage.ResourceManager.Owner::(cancelAll)
Owner.cancelAll requests {}
MainThread::DEBUG::2015-09-02
16:07:49,928::task::993::Storage.TaskManager.Task::(_decref)
Task=`68d01d7d-b426-4465-829e-174e2cb47e9e`::ref 0 aborting False
MainThread::INFO::2015-09-02 16:07:49,928::momIF::46::MOM::(__init__)
Preparing MOM interface
MainThread::INFO::2015-09-02 16:07:49,929::momIF::55::MOM::(__init__) Using
named unix socket /var/run/vdsm/mom-vdsm.sock
MainThread::INFO::2015-09-02 16:07:49,929::secret::90::root::(clear)
Unregistering all secrests
MainThread::DEBUG::2015-09-02
16:07:49,929::libvirtconnection::160::root::(get) trying to connect libvirt
MainThread::INFO::2015-09-02
16:07:49,933::vmchannels::196::vds::(settimeout) Setting channels' timeout
to 30 seconds.
VM Channels Listener::DEBUG::2015-09-02
16:07:49,934::vmchannels::178::vds::(run) Starting VM channels listener
thread.
MainThread::INFO::2015-09-02
16:07:49,935::protocoldetector::172::vds.MultiProtocolAcceptor::(__init__)
Listening at 0.0.0.0:54321
MainThread::DEBUG::2015-09-02
16:07:50,063::protocoldetector::199::vds.MultiProtocolAcceptor::(add_detector)
Adding detector <rpc.bindingxmlrpc.XmlDetector instance at 0x7fc2f00dc440>
storageRefresh::DEBUG::2015-09-02
16:07:50,063::misc::743::Storage.SamplingMethod::(__call__) Returning last
result
storageRefresh::DEBUG::2015-09-02
16:07:50,080::misc::733::Storage.SamplingMethod::(__call__) Trying to enter
sampling method (storage.hba.rescan)
storageRefresh::DEBUG::2015-09-02
16:07:50,081::misc::736::Storage.SamplingMethod::(__call__) Got in to
sampling method
storageRefresh::DEBUG::2015-09-02
16:07:50,081::hba::56::Storage.HBA::(rescan) Starting scan
storageRefresh::DEBUG::2015-09-02
16:07:50,081::supervdsm::76::SuperVdsmProxy::(_connect) Trying to connect
to Super Vdsm
MainThread::DEBUG::2015-09-02
16:07:50,157::protocoldetector::199::vds.MultiProtocolAcceptor::(add_detector)
Adding detector <yajsonrpc.stompreactor.StompDetector instance at
0x7fc2e01332d8>
BindingXMLRPC::INFO::2015-09-02
16:07:50,158::bindingxmlrpc::62::vds::(threaded_start) XMLRPC server running
MainThread::DEBUG::2015-09-02
16:07:50,158::schedule::98::Scheduler::(start) Starting scheduler
periodic-sched
periodic-sched::DEBUG::2015-09-02
16:07:50,159::schedule::142::Scheduler::(_run) started
MainThread::DEBUG::2015-09-02 16:07:50,159::executor::69::Executor::(start)
Starting executor
MainThread::DEBUG::2015-09-02
16:07:50,159::executor::157::Executor::(__init__) Starting worker periodic/0
periodic/0::DEBUG::2015-09-02 16:07:50,159::executor::171::Executor::(_run)
Worker started
MainThread::DEBUG::2015-09-02
16:07:50,159::executor::157::Executor::(__init__) Starting worker periodic/1
periodic/1::DEBUG::2015-09-02 16:07:50,160::executor::171::Executor::(_run)
Worker started
MainThread::DEBUG::2015-09-02
16:07:50,160::executor::157::Executor::(__init__) Starting worker periodic/2
periodic/2::DEBUG::2015-09-02 16:07:50,160::executor::171::Executor::(_run)
Worker started
MainThread::DEBUG::2015-09-02
16:07:50,160::executor::157::Executor::(__init__) Starting worker periodic/3
periodic/3::DEBUG::2015-09-02 16:07:50,160::executor::171::Executor::(_run)
Worker started
MainThread::DEBUG::2015-09-02
16:07:50,160::libvirtconnection::160::root::(get) trying to connect libvirt
MainThread::DEBUG::2015-09-02
16:07:50,163::periodic::157::virt.periodic.Operation::(start) starting
operation VmDispatcher(<class 'virt.periodic.UpdateVolumes'>)
MainThread::DEBUG::2015-09-02
16:07:50,164::periodic::157::virt.periodic.Operation::(start) starting
operation VmDispatcher(<class 'virt.periodic.NumaInfoMonitor'>)
MainThread::DEBUG::2015-09-02
16:07:50,164::periodic::157::virt.periodic.Operation::(start) starting
operation VmDispatcher(<class 'virt.periodic.BlockjobMonitor'>)
MainThread::DEBUG::2015-09-02
16:07:50,164::periodic::157::virt.periodic.Operation::(start) starting
operation <virt.sampling.VMBulkSampler object at 0x7fc2e0151d10>
MainThread::DEBUG::2015-09-02
16:07:50,164::periodic::157::virt.periodic.Operation::(start) starting
operation VmDispatcher(<class 'virt.periodic.DriveWatermarkMonitor'>)
storageRefresh::DEBUG::2015-09-02
16:07:50,167::hba::62::Storage.HBA::(rescan) Scan finished
storageRefresh::DEBUG::2015-09-02
16:07:50,167::misc::743::Storage.SamplingMethod::(__call__) Returning last
result
storageRefresh::DEBUG::2015-09-02
16:07:50,167::multipath::77::Storage.Misc.excCmd::(rescan) /usr/bin/sudo -n
/usr/sbin/multipath (cwd None)
storageRefresh::DEBUG::2015-09-02
16:07:50,513::multipath::77::Storage.Misc.excCmd::(rescan) SUCCESS: <err> =
''; <rc> = 0
storageRefresh::DEBUG::2015-09-02 16:07:50,513::utils::661::root::(execCmd)
/sbin/udevadm settle --timeout=5 (cwd None)
Reactor thread::INFO::2015-09-02
16:07:50,590::protocoldetector::72::ProtocolDetector.AcceptorImpl::(handle_accept)
Accepting connection from 127.0.0.1:56311
Reactor thread::DEBUG::2015-09-02
16:07:50,596::protocoldetector::82::ProtocolDetector.Detector::(__init__)
Using required_size=11
Reactor thread::INFO::2015-09-02
16:07:50,596::protocoldetector::118::ProtocolDetector.Detector::(handle_read)
Detected protocol xml from 127.0.0.1:56311
Reactor thread::DEBUG::2015-09-02
16:07:50,596::bindingxmlrpc::1296::XmlDetector::(handle_socket) xml over
http detected from ('127.0.0.1', 56311)
BindingXMLRPC::INFO::2015-09-02
16:07:50,596::xmlrpc::73::vds.XMLRPCServer::(handle_request) Starting
request handler for 127.0.0.1:56311
Thread-13::INFO::2015-09-02
16:07:50,597::xmlrpc::84::vds.XMLRPCServer::(_process_requests) Request
handler for 127.0.0.1:56311 started
Thread-13::DEBUG::2015-09-02
16:07:50,597::bindingxmlrpc::1256::vds::(wrapper) client [127.0.0.1]::call
getHardwareInfo with () {}
Thread-13::DEBUG::2015-09-02
16:07:50,597::bindingxmlrpc::1263::vds::(wrapper) return getHardwareInfo
with {'status': {'message': 'Recovering from crash or
Initializing',
'code': 99}}
Thread-13::INFO::2015-09-02
16:07:50,599::xmlrpc::92::vds.XMLRPCServer::(_process_requests) Request
handler for 127.0.0.1:56311 stopped
Reactor thread::INFO::2015-09-02
16:07:51,607::protocoldetector::72::ProtocolDetector.AcceptorImpl::(handle_accept)
Accepting connection from 127.0.0.1:56312
Reactor thread::DEBUG::2015-09-02
16:07:51,613::protocoldetector::82::ProtocolDetector.Detector::(__init__)
Using required_size=11
Reactor thread::INFO::2015-09-02
16:07:51,613::protocoldetector::118::ProtocolDetector.Detector::(handle_read)
Detected protocol xml from 127.0.0.1:56312
Reactor thread::DEBUG::2015-09-02
16:07:51,613::bindingxmlrpc::1296::XmlDetector::(handle_socket) xml over
http detected from ('127.0.0.1', 56312)
BindingXMLRPC::INFO::2015-09-02
16:07:51,613::xmlrpc::73::vds.XMLRPCServer::(handle_request) Starting
request handler for 127.0.0.1:56312
Thread-14::INFO::2015-09-02
16:07:51,613::xmlrpc::84::vds.XMLRPCServer::(_process_requests) Request
handler for 127.0.0.1:56312 started
Thread-14::DEBUG::2015-09-02
16:07:51,614::bindingxmlrpc::1256::vds::(wrapper) client [127.0.0.1]::call
getHardwareInfo with () {}
Thread-14::DEBUG::2015-09-02
16:07:51,614::bindingxmlrpc::1263::vds::(wrapper) return getHardwareInfo
with {'status': {'message': 'Recovering from crash or
Initializing',
'code': 99}}
Thread-14::INFO::2015-09-02
16:07:51,615::xmlrpc::92::vds.XMLRPCServer::(_process_requests) Request
handler for 127.0.0.1:56312 stopped
storageRefresh::DEBUG::2015-09-02 16:07:51,924::utils::679::root::(execCmd)
SUCCESS: <err> = ''; <rc> = 0
storageRefresh::DEBUG::2015-09-02
16:07:51,926::lvm::498::Storage.OperationMutex::(_invalidateAllPvs)
Operation 'lvm invalidate operation' got the operation mutex
storageRefresh::DEBUG::2015-09-02
16:07:51,926::lvm::500::Storage.OperationMutex::(_invalidateAllPvs)
Operation 'lvm invalidate operation' released the operation mutex
storageRefresh::DEBUG::2015-09-02
16:07:51,926::lvm::509::Storage.OperationMutex::(_invalidateAllVgs)
Operation 'lvm invalidate operation' got the operation mutex
storageRefresh::DEBUG::2015-09-02
16:07:51,926::lvm::511::Storage.OperationMutex::(_invalidateAllVgs)
Operation 'lvm invalidate operation' released the operation mutex
storageRefresh::DEBUG::2015-09-02
16:07:51,926::lvm::529::Storage.OperationMutex::(_invalidateAllLvs)
Operation 'lvm invalidate operation' got the operation mutex
storageRefresh::DEBUG::2015-09-02
16:07:51,926::lvm::531::Storage.OperationMutex::(_invalidateAllLvs)
Operation 'lvm invalidate operation' released the operation mutex
storageRefresh::DEBUG::2015-09-02
16:07:51,926::misc::743::Storage.SamplingMethod::(__call__) Returning last
result
storageRefresh::DEBUG::2015-09-02
16:07:51,926::lvm::320::Storage.OperationMutex::(_reloadpvs) Operation 'lvm
reload operation' got the operation mutex
storageRefresh::DEBUG::2015-09-02
16:07:51,927::lvm::291::Storage.Misc.excCmd::(cmd) /usr/bin/sudo -n
/usr/sbin/lvm pvs --config ' devices { preferred_names = ["^/dev/mapper/"]
ignore_suspended_devices=1 write_cache_state=0 disable_after_error_count=3
obtain_device_list_from_udev=0 filter = [
'\''a|/dev/mapper/Hitachi_HDS721010DLE630_MSK523Y209VK0B|'\'',
'\''r|.*|'\'' ] } global { locking_type=1
prioritise_write_locks=1
wait_for_locks=1 use_lvmetad=0 } backup { retain_min = 50 retain_days =
0 } ' --noheadings --units b --nosuffix --separator '|'
--ignoreskippedcluster -o
uuid,name,size,vg_name,vg_uuid,pe_start,pe_count,pe_alloc_count,mda_count,dev_size
(cwd None)
storageRefresh::DEBUG::2015-09-02
16:07:52,341::lvm::291::Storage.Misc.excCmd::(cmd) SUCCESS: <err> = '
WARNING: lvmetad is running but disabled. Restart lvmetad before enabling
it!\n'; <rc> = 0
storageRefresh::DEBUG::2015-09-02
16:07:52,341::lvm::348::Storage.OperationMutex::(_reloadpvs) Operation 'lvm
reload operation' released the operation mutex
storageRefresh::DEBUG::2015-09-02
16:07:52,341::lvm::371::Storage.OperationMutex::(_reloadvgs) Operation 'lvm
reload operation' got the operation mutex
storageRefresh::DEBUG::2015-09-02
16:07:52,341::lvm::291::Storage.Misc.excCmd::(cmd) /usr/bin/sudo -n
/usr/sbin/lvm vgs --config ' devices { preferred_names = ["^/dev/mapper/"]
ignore_suspended_devices=1 write_cache_state=0 disable_after_error_count=3
obtain_device_list_from_udev=0 filter = [
'\''a|/dev/mapper/Hitachi_HDS721010DLE630_MSK523Y209VK0B|'\'',
'\''r|.*|'\'' ] } global { locking_type=1
prioritise_write_locks=1
wait_for_locks=1 use_lvmetad=0 } backup { retain_min = 50 retain_days =
0 } ' --noheadings --units b --nosuffix --separator '|'
--ignoreskippedcluster -o
uuid,name,attr,size,free,extent_size,extent_count,free_count,tags,vg_mda_size,vg_mda_free,lv_count,pv_count,pv_name
(cwd None)
storageRefresh::DEBUG::2015-09-02
16:07:52,405::lvm::291::Storage.Misc.excCmd::(cmd) SUCCESS: <err> = '
WARNING: lvmetad is running but disabled. Restart lvmetad before enabling
it!\n'; <rc> = 0
storageRefresh::DEBUG::2015-09-02
16:07:52,405::lvm::416::Storage.OperationMutex::(_reloadvgs) Operation 'lvm
reload operation' released the operation mutex
storageRefresh::DEBUG::2015-09-02
16:07:52,406::lvm::291::Storage.Misc.excCmd::(cmd) /usr/bin/sudo -n
/usr/sbin/lvm lvs --config ' devices { preferred_names = ["^/dev/mapper/"]
ignore_suspended_devices=1 write_cache_state=0 disable_after_error_count=3
obtain_device_list_from_udev=0 filter = [
'\''a|/dev/mapper/Hitachi_HDS721010DLE630_MSK523Y209VK0B|'\'',
'\''r|.*|'\'' ] } global { locking_type=1
prioritise_write_locks=1
wait_for_locks=1 use_lvmetad=0 } backup { retain_min = 50 retain_days =
0 } ' --noheadings --units b --nosuffix --separator '|'
--ignoreskippedcluster -o
uuid,name,vg_name,attr,size,seg_start_pe,devices,tags (cwd None)
storageRefresh::DEBUG::2015-09-02
16:07:52,458::lvm::291::Storage.Misc.excCmd::(cmd) SUCCESS: <err> = '
WARNING: lvmetad is running but disabled. Restart lvmetad before enabling
it!\n'; <rc> = 0
storageRefresh::DEBUG::2015-09-02
16:07:52,458::lvm::371::Storage.OperationMutex::(_reloadvgs) Operation 'lvm
reload operation' got the operation mutex
storageRefresh::DEBUG::2015-09-02
16:07:52,459::lvm::291::Storage.Misc.excCmd::(cmd) /usr/bin/sudo -n
/usr/sbin/lvm vgs --config ' devices { preferred_names = ["^/dev/mapper/"]
ignore_suspended_devices=1 write_cache_state=0 disable_after_error_count=3
obtain_device_list_from_udev=0 filter = [
'\''a|/dev/mapper/Hitachi_HDS721010DLE630_MSK523Y209VK0B|'\'',
'\''r|.*|'\'' ] } global { locking_type=1
prioritise_write_locks=1
wait_for_locks=1 use_lvmetad=0 } backup { retain_min = 50 retain_days =
0 } ' --noheadings --units b --nosuffix --separator '|'
--ignoreskippedcluster -o
uuid,name,attr,size,free,extent_size,extent_count,free_count,tags,vg_mda_size,vg_mda_free,lv_count,pv_count,pv_name
(cwd None)
storageRefresh::DEBUG::2015-09-02
16:07:52,491::lvm::291::Storage.Misc.excCmd::(cmd) SUCCESS: <err> = '
WARNING: lvmetad is running but disabled. Restart lvmetad before enabling
it!\n'; <rc> = 0
storageRefresh::DEBUG::2015-09-02
16:07:52,491::lvm::416::Storage.OperationMutex::(_reloadvgs) Operation 'lvm
reload operation' released the operation mutex
storageRefresh::DEBUG::2015-09-02
16:07:52,491::hsm::373::Storage.HSM::(storageRefresh) HSM is ready
Reactor thread::INFO::2015-09-02
16:07:52,624::protocoldetector::72::ProtocolDetector.AcceptorImpl::(handle_accept)
Accepting connection from 127.0.0.1:56313
Reactor thread::DEBUG::2015-09-02
16:07:52,629::protocoldetector::82::ProtocolDetector.Detector::(__init__)
Using required_size=11
Reactor thread::INFO::2015-09-02
16:07:52,629::protocoldetector::118::ProtocolDetector.Detector::(handle_read)
Detected protocol xml from 127.0.0.1:56313
Reactor thread::DEBUG::2015-09-02
16:07:52,629::bindingxmlrpc::1296::XmlDetector::(handle_socket) xml over
http detected from ('127.0.0.1', 56313)
BindingXMLRPC::INFO::2015-09-02
16:07:52,629::xmlrpc::73::vds.XMLRPCServer::(handle_request) Starting
request handler for 127.0.0.1:56313
Thread-15::INFO::2015-09-02
16:07:52,630::xmlrpc::84::vds.XMLRPCServer::(_process_requests) Request
handler for 127.0.0.1:56313 started
Thread-15::DEBUG::2015-09-02
16:07:52,630::bindingxmlrpc::1256::vds::(wrapper) client [127.0.0.1]::call
getHardwareInfo with () {}
Thread-15::DEBUG::2015-09-02
16:07:52,719::bindingxmlrpc::1263::vds::(wrapper) return getHardwareInfo
with {'status': {'message': 'Done', 'code': 0},
'info':
{'systemProductName': 'System Product Name', 'systemSerialNumber':
'System
Serial Number', 'systemFamily': 'To be filled by O.E.M.',
'systemVersion':
'System Version', 'systemUUID':
'267A6B80-D7DA-11DD-81CF-C860009B3CD9',
'systemManufacturer': 'System manufacturer'}}
Thread-15::INFO::2015-09-02
16:07:52,721::xmlrpc::92::vds.XMLRPCServer::(_process_requests) Request
handler for 127.0.0.1:56313 stopped
Reactor thread::INFO::2015-09-02
16:07:52,730::protocoldetector::72::ProtocolDetector.AcceptorImpl::(handle_accept)
Accepting connection from 127.0.0.1:56314
Reactor thread::DEBUG::2015-09-02
16:07:52,735::protocoldetector::82::ProtocolDetector.Detector::(__init__)
Using required_size=11
Reactor thread::INFO::2015-09-02
16:07:52,735::protocoldetector::118::ProtocolDetector.Detector::(handle_read)
Detected protocol xml from 127.0.0.1:56314
Reactor thread::DEBUG::2015-09-02
16:07:52,735::bindingxmlrpc::1296::XmlDetector::(handle_socket) xml over
http detected from ('127.0.0.1', 56314)
BindingXMLRPC::INFO::2015-09-02
16:07:52,735::xmlrpc::73::vds.XMLRPCServer::(handle_request) Starting
request handler for 127.0.0.1:56314
Thread-16::INFO::2015-09-02
16:07:52,735::xmlrpc::84::vds.XMLRPCServer::(_process_requests) Request
handler for 127.0.0.1:56314 started
Thread-16::DEBUG::2015-09-02
16:07:52,736::bindingxmlrpc::325::vds::(wrapper) client [127.0.0.1]
Thread-16::DEBUG::2015-09-02
16:07:52,736::task::595::Storage.TaskManager.Task::(_updateState)
Task=`c4a18001-912b-47dc-9713-7d50e5133b59`::moving from state init ->
state preparing
Thread-16::INFO::2015-09-02
16:07:52,736::logUtils::48::dispatcher::(wrapper) Run and protect:
connectStorageServer(domType=3,
spUUID='00000000-0000-0000-0000-000000000000', conList=[{'id':
'57bc98c0-560f-4e61-9d86-df92ad468d3b', 'connection':
'192.168.1.50',
'iqn': 'iqn.2015-08.openSuse.wodel:target00', 'portal':
'1', 'user':
'iscsiuser', 'password': '********', 'port':
'3260'}], options=None)
Thread-16::DEBUG::2015-09-02
16:07:52,737::iscsiadm::97::Storage.Misc.excCmd::(_runCmd) /usr/bin/sudo -n
/sbin/iscsiadm -m node -T iqn.2015-08.openSuse.wodel:target00 -I default -p
192.168.1.50:3260,1 --op=new (cwd None)
Thread-16::DEBUG::2015-09-02
16:07:52,789::iscsiadm::97::Storage.Misc.excCmd::(_runCmd) SUCCESS: <err> =
''; <rc> = 0
Thread-16::DEBUG::2015-09-02
16:07:52,789::iscsiadm::97::Storage.Misc.excCmd::(_runCmd) /sbin/iscsiadm
-m node -T iqn.2015-08.openSuse.wodel:target00 -I default -p
192.168.1.50:3260,1 -n node.session.auth.authmethod -v '****' --op=update
(cwd None)
Thread-16::DEBUG::2015-09-02
16:07:52,811::iscsiadm::97::Storage.Misc.excCmd::(_runCmd) SUCCESS: <err> =
''; <rc> = 0
Thread-16::DEBUG::2015-09-02
16:07:52,812::iscsiadm::97::Storage.Misc.excCmd::(_runCmd) /sbin/iscsiadm
-m node -T iqn.2015-08.openSuse.wodel:target00 -I default -p
192.168.1.50:3260,1 -n node.session.auth.username -v '****' --op=update
(cwd None)
Thread-16::DEBUG::2015-09-02
16:07:52,846::iscsiadm::97::Storage.Misc.excCmd::(_runCmd) SUCCESS: <err> =
''; <rc> = 0
Thread-16::DEBUG::2015-09-02
16:07:52,847::iscsiadm::97::Storage.Misc.excCmd::(_runCmd) /sbin/iscsiadm
-m node -T iqn.2015-08.openSuse.wodel:target00 -I default -p
192.168.1.50:3260,1 -n node.session.auth.password -v '****' --op=update
(cwd None)
Thread-16::DEBUG::2015-09-02
16:07:52,868::iscsiadm::97::Storage.Misc.excCmd::(_runCmd) SUCCESS: <err> =
''; <rc> = 0
Thread-16::DEBUG::2015-09-02
16:07:52,868::iscsiadm::97::Storage.Misc.excCmd::(_runCmd) /usr/bin/sudo -n
/sbin/iscsiadm -m iface -I default (cwd None)
Thread-16::DEBUG::2015-09-02
16:07:52,905::iscsiadm::97::Storage.Misc.excCmd::(_runCmd) SUCCESS: <err> =
''; <rc> = 0
Thread-16::INFO::2015-09-02
16:07:52,905::iscsi::564::Storage.ISCSI::(setRpFilterIfNeeded) iSCSI
iface.net_ifacename not provided. Skipping.
Thread-16::DEBUG::2015-09-02
16:07:52,906::iscsiadm::97::Storage.Misc.excCmd::(_runCmd) /usr/bin/sudo -n
/sbin/iscsiadm -m node -T iqn.2015-08.openSuse.wodel:target00 -I default -p
192.168.1.50:3260,1 -l (cwd None)
Thread-16::DEBUG::2015-09-02
16:07:53,027::iscsiadm::97::Storage.Misc.excCmd::(_runCmd) SUCCESS: <err> =
''; <rc> = 0
Thread-16::DEBUG::2015-09-02
16:07:53,028::iscsiadm::97::Storage.Misc.excCmd::(_runCmd) /usr/bin/sudo -n
/sbin/iscsiadm -m node -T iqn.2015-08.openSuse.wodel:target00 -I default -p
192.168.1.50:3260,1 -n node.startup -v manual --op=update (cwd None)
Thread-16::DEBUG::2015-09-02
16:07:53,088::iscsiadm::97::Storage.Misc.excCmd::(_runCmd) SUCCESS: <err> =
''; <rc> = 0
Thread-16::DEBUG::2015-09-02 16:07:53,088::utils::661::root::(execCmd)
/sbin/udevadm settle --timeout=5 (cwd None)
Thread-16::DEBUG::2015-09-02 16:07:53,182::utils::679::root::(execCmd)
SUCCESS: <err> = ''; <rc> = 0
Thread-16::DEBUG::2015-09-02
16:07:53,182::misc::733::Storage.SamplingMethod::(__call__) Trying to enter
sampling method (storage.sdc.refreshStorage)
Thread-16::DEBUG::2015-09-02
16:07:53,182::misc::736::Storage.SamplingMethod::(__call__) Got in to
sampling method
Thread-16::DEBUG::2015-09-02
16:07:53,182::misc::733::Storage.SamplingMethod::(__call__) Trying to enter
sampling method (storage.iscsi.rescan)
Thread-16::DEBUG::2015-09-02
16:07:53,182::misc::736::Storage.SamplingMethod::(__call__) Got in to
sampling method
Thread-16::DEBUG::2015-09-02
16:07:53,182::iscsi::431::Storage.ISCSI::(rescan) Performing SCSI scan,
this will take up to 30 seconds
Thread-16::DEBUG::2015-09-02
16:07:53,182::iscsiadm::97::Storage.Misc.excCmd::(_runCmd) /usr/bin/sudo -n
/sbin/iscsiadm -m session -R (cwd None)
Thread-16::DEBUG::2015-09-02
16:07:53,229::misc::743::Storage.SamplingMethod::(__call__) Returning last
result
Thread-16::DEBUG::2015-09-02
16:07:53,229::misc::733::Storage.SamplingMethod::(__call__) Trying to enter
sampling method (storage.hba.rescan)
Thread-16::DEBUG::2015-09-02
16:07:53,229::misc::736::Storage.SamplingMethod::(__call__) Got in to
sampling method
Thread-16::DEBUG::2015-09-02 16:07:53,229::hba::56::Storage.HBA::(rescan)
Starting scan
Thread-16::DEBUG::2015-09-02 16:07:53,300::hba::62::Storage.HBA::(rescan)
Scan finished
Thread-16::DEBUG::2015-09-02
16:07:53,300::misc::743::Storage.SamplingMethod::(__call__) Returning last
result
Thread-16::DEBUG::2015-09-02
16:07:53,300::multipath::77::Storage.Misc.excCmd::(rescan) /usr/bin/sudo -n
/usr/sbin/multipath (cwd None)
Thread-16::DEBUG::2015-09-02
16:07:53,435::multipath::77::Storage.Misc.excCmd::(rescan) SUCCESS: <err> =
''; <rc> = 0
Thread-16::DEBUG::2015-09-02 16:07:53,435::utils::661::root::(execCmd)
/sbin/udevadm settle --timeout=5 (cwd None)
Thread-16::DEBUG::2015-09-02 16:07:53,919::utils::679::root::(execCmd)
SUCCESS: <err> = ''; <rc> = 0
Thread-16::DEBUG::2015-09-02
16:07:53,921::lvm::498::Storage.OperationMutex::(_invalidateAllPvs)
Operation 'lvm invalidate operation' got the operation mutex
Thread-16::DEBUG::2015-09-02
16:07:53,921::lvm::500::Storage.OperationMutex::(_invalidateAllPvs)
Operation 'lvm invalidate operation' released the operation mutex
Thread-16::DEBUG::2015-09-02
16:07:53,922::lvm::509::Storage.OperationMutex::(_invalidateAllVgs)
Operation 'lvm invalidate operation' got the operation mutex
Thread-16::DEBUG::2015-09-02
16:07:53,922::lvm::511::Storage.OperationMutex::(_invalidateAllVgs)
Operation 'lvm invalidate operation' released the operation mutex
Thread-16::DEBUG::2015-09-02
16:07:53,922::lvm::529::Storage.OperationMutex::(_invalidateAllLvs)
Operation 'lvm invalidate operation' got the operation mutex
Thread-16::DEBUG::2015-09-02
16:07:53,922::lvm::531::Storage.OperationMutex::(_invalidateAllLvs)
Operation 'lvm invalidate operation' released the operation mutex
Thread-16::DEBUG::2015-09-02
16:07:53,922::misc::743::Storage.SamplingMethod::(__call__) Returning last
result
Thread-16::DEBUG::2015-09-02
16:07:53,922::lvm::371::Storage.OperationMutex::(_reloadvgs) Operation 'lvm
reload operation' got the operation mutex
Thread-16::DEBUG::2015-09-02
16:07:53,923::lvm::291::Storage.Misc.excCmd::(cmd) /usr/bin/sudo -n
/usr/sbin/lvm vgs --config ' devices { preferred_names = ["^/dev/mapper/"]
ignore_suspended_devices=1 write_cache_state=0 disable_after_error_count=3
obtain_device_list_from_udev=0 filter = [
'\''a|/dev/mapper/33000000100000001|/dev/mapper/Hitachi_HDS721010DLE630_MSK523Y209VK0B|'\'',
'\''r|.*|'\'' ] } global { locking_type=1
prioritise_write_locks=1
wait_for_locks=1 use_lvmetad=0 } backup { retain_min = 50 retain_days =
0 } ' --noheadings --units b --nosuffix --separator '|'
--ignoreskippedcluster -o
uuid,name,attr,size,free,extent_size,extent_count,free_count,tags,vg_mda_size,vg_mda_free,lv_count,pv_count,pv_name
(cwd None)
Thread-16::DEBUG::2015-09-02
16:07:54,058::lvm::291::Storage.Misc.excCmd::(cmd) SUCCESS: <err> = '
WARNING: lvmetad is running but disabled. Restart lvmetad before enabling
it!\n'; <rc> = 0
Thread-16::DEBUG::2015-09-02
16:07:54,059::lvm::416::Storage.OperationMutex::(_reloadvgs) Operation 'lvm
reload operation' released the operation mutex
Thread-16::DEBUG::2015-09-02
16:07:54,059::hsm::2418::Storage.HSM::(__prefetchDomains) Found SD uuids:
('8b25f3be-7574-4f7a-8851-363129704e52',)
Thread-16::DEBUG::2015-09-02
16:07:54,059::hsm::2478::Storage.HSM::(connectStorageServer) knownSDs:
{8b25f3be-7574-4f7a-8851-363129704e52: storage.blockSD.findDomain}
Thread-16::INFO::2015-09-02
16:07:54,059::logUtils::51::dispatcher::(wrapper) Run and protect:
connectStorageServer, Return response: {'statuslist': [{'status': 0,
'id':
'57bc98c0-560f-4e61-9d86-df92ad468d3b'}]}
Thread-16::DEBUG::2015-09-02
16:07:54,059::task::1191::Storage.TaskManager.Task::(prepare)
Task=`c4a18001-912b-47dc-9713-7d50e5133b59`::finished: {'statuslist':
[{'status': 0, 'id': '57bc98c0-560f-4e61-9d86-df92ad468d3b'}]}
Thread-16::DEBUG::2015-09-02
16:07:54,059::task::595::Storage.TaskManager.Task::(_updateState)
Task=`c4a18001-912b-47dc-9713-7d50e5133b59`::moving from state preparing ->
state finished
Thread-16::DEBUG::2015-09-02
16:07:54,060::resourceManager::940::Storage.ResourceManager.Owner::(releaseAll)
Owner.releaseAll requests {} resources {}
Thread-16::DEBUG::2015-09-02
16:07:54,060::resourceManager::977::Storage.ResourceManager.Owner::(cancelAll)
Owner.cancelAll requests {}
Thread-16::DEBUG::2015-09-02
16:07:54,060::task::993::Storage.TaskManager.Task::(_decref)
Task=`c4a18001-912b-47dc-9713-7d50e5133b59`::ref 0 aborting False
Thread-16::INFO::2015-09-02
16:07:54,062::xmlrpc::92::vds.XMLRPCServer::(_process_requests) Request
handler for 127.0.0.1:56314 stopped
Reactor thread::INFO::2015-09-02
16:07:54,070::protocoldetector::72::ProtocolDetector.AcceptorImpl::(handle_accept)
Accepting connection from 127.0.0.1:56316
Reactor thread::DEBUG::2015-09-02
16:07:54,075::protocoldetector::82::ProtocolDetector.Detector::(__init__)
Using required_size=11
Reactor thread::INFO::2015-09-02
16:07:54,076::protocoldetector::118::ProtocolDetector.Detector::(handle_read)
Detected protocol xml from 127.0.0.1:56316
Reactor thread::DEBUG::2015-09-02
16:07:54,076::bindingxmlrpc::1296::XmlDetector::(handle_socket) xml over
http detected from ('127.0.0.1', 56316)
BindingXMLRPC::INFO::2015-09-02
16:07:54,076::xmlrpc::73::vds.XMLRPCServer::(handle_request) Starting
request handler for 127.0.0.1:56316
Thread-17::INFO::2015-09-02
16:07:54,076::xmlrpc::84::vds.XMLRPCServer::(_process_requests) Request
handler for 127.0.0.1:56316 started
Thread-17::DEBUG::2015-09-02
16:07:54,077::bindingxmlrpc::325::vds::(wrapper) client [127.0.0.1]
Thread-17::DEBUG::2015-09-02
16:07:54,077::task::595::Storage.TaskManager.Task::(_updateState)
Task=`7936300e-8a1a-47f5-83c4-16ed19853e36`::moving from state init ->
state preparing
Thread-17::INFO::2015-09-02
16:07:54,077::logUtils::48::dispatcher::(wrapper) Run and protect:
prepareImage(sdUUID='8b25f3be-7574-4f7a-8851-363129704e52',
spUUID='00000000-0000-0000-0000-000000000000',
imgUUID='bf3bdae1-7318-4443-a19b-7371de30b982',
leafUUID='cbb10cf0-9600-465e-aed9-412f7157706b')
Thread-17::DEBUG::2015-09-02
16:07:54,077::resourceManager::198::Storage.ResourceManager.Request::(__init__)
ResName=`Storage.8b25f3be-7574-4f7a-8851-363129704e52`ReqID=`fc59b8b4-51c5-4a15-9716-aedbb6de62e6`::Request
was made in '/usr/share/vdsm/storage/hsm.py' line '3194' at
'prepareImage'
Thread-17::DEBUG::2015-09-02
16:07:54,078::resourceManager::542::Storage.ResourceManager::(registerResource)
Trying to register resource 'Storage.8b25f3be-7574-4f7a-8851-363129704e52'
for lock type 'shared'
Thread-17::DEBUG::2015-09-02
16:07:54,078::resourceManager::601::Storage.ResourceManager::(registerResource)
Resource 'Storage.8b25f3be-7574-4f7a-8851-363129704e52' is free. Now
locking as 'shared' (1 active user)
Thread-17::DEBUG::2015-09-02
16:07:54,078::resourceManager::238::Storage.ResourceManager.Request::(grant)
ResName=`Storage.8b25f3be-7574-4f7a-8851-363129704e52`ReqID=`fc59b8b4-51c5-4a15-9716-aedbb6de62e6`::Granted
request
Thread-17::DEBUG::2015-09-02
16:07:54,078::task::827::Storage.TaskManager.Task::(resourceAcquired)
Task=`7936300e-8a1a-47f5-83c4-16ed19853e36`::_resourcesAcquired:
Storage.8b25f3be-7574-4f7a-8851-363129704e52 (shared)
Thread-17::DEBUG::2015-09-02
16:07:54,078::task::993::Storage.TaskManager.Task::(_decref)
Task=`7936300e-8a1a-47f5-83c4-16ed19853e36`::ref 1 aborting False
Thread-17::DEBUG::2015-09-02
16:07:54,078::misc::733::Storage.SamplingMethod::(__call__) Trying to enter
sampling method (storage.sdc.refreshStorage)
Thread-17::DEBUG::2015-09-02
16:07:54,078::misc::736::Storage.SamplingMethod::(__call__) Got in to
sampling method
Thread-17::DEBUG::2015-09-02
16:07:54,078::misc::733::Storage.SamplingMethod::(__call__) Trying to enter
sampling method (storage.iscsi.rescan)
Thread-17::DEBUG::2015-09-02
16:07:54,078::misc::736::Storage.SamplingMethod::(__call__) Got in to
sampling method
Thread-17::DEBUG::2015-09-02
16:07:54,078::iscsi::431::Storage.ISCSI::(rescan) Performing SCSI scan,
this will take up to 30 seconds
Thread-17::DEBUG::2015-09-02
16:07:54,078::iscsiadm::97::Storage.Misc.excCmd::(_runCmd) /usr/bin/sudo -n
/sbin/iscsiadm -m session -R (cwd None)
Thread-17::DEBUG::2015-09-02
16:07:54,130::misc::743::Storage.SamplingMethod::(__call__) Returning last
result
Thread-17::DEBUG::2015-09-02
16:07:54,130::misc::733::Storage.SamplingMethod::(__call__) Trying to enter
sampling method (storage.hba.rescan)
Thread-17::DEBUG::2015-09-02
16:07:54,130::misc::736::Storage.SamplingMethod::(__call__) Got in to
sampling method
Thread-17::DEBUG::2015-09-02 16:07:54,130::hba::56::Storage.HBA::(rescan)
Starting scan
Thread-17::DEBUG::2015-09-02 16:07:54,197::hba::62::Storage.HBA::(rescan)
Scan finished
Thread-17::DEBUG::2015-09-02
16:07:54,197::misc::743::Storage.SamplingMethod::(__call__) Returning last
result
Thread-17::DEBUG::2015-09-02
16:07:54,197::multipath::77::Storage.Misc.excCmd::(rescan) /usr/bin/sudo -n
/usr/sbin/multipath (cwd None)
Thread-17::DEBUG::2015-09-02
16:07:54,298::multipath::77::Storage.Misc.excCmd::(rescan) SUCCESS: <err> =
''; <rc> = 0
Thread-17::DEBUG::2015-09-02 16:07:54,299::utils::661::root::(execCmd)
/sbin/udevadm settle --timeout=5 (cwd None)
Thread-17::DEBUG::2015-09-02 16:07:54,307::utils::679::root::(execCmd)
SUCCESS: <err> = ''; <rc> = 0
Thread-17::DEBUG::2015-09-02
16:07:54,309::lvm::498::Storage.OperationMutex::(_invalidateAllPvs)
Operation 'lvm invalidate operation' got the operation mutex
Thread-17::DEBUG::2015-09-02
16:07:54,310::lvm::500::Storage.OperationMutex::(_invalidateAllPvs)
Operation 'lvm invalidate operation' released the operation mutex
Thread-17::DEBUG::2015-09-02
16:07:54,310::lvm::509::Storage.OperationMutex::(_invalidateAllVgs)
Operation 'lvm invalidate operation' got the operation mutex
Thread-17::DEBUG::2015-09-02
16:07:54,310::lvm::511::Storage.OperationMutex::(_invalidateAllVgs)
Operation 'lvm invalidate operation' released the operation mutex
Thread-17::DEBUG::2015-09-02
16:07:54,310::lvm::529::Storage.OperationMutex::(_invalidateAllLvs)
Operation 'lvm invalidate operation' got the operation mutex
Thread-17::DEBUG::2015-09-02
16:07:54,310::lvm::531::Storage.OperationMutex::(_invalidateAllLvs)
Operation 'lvm invalidate operation' released the operation mutex
Thread-17::DEBUG::2015-09-02
16:07:54,310::misc::743::Storage.SamplingMethod::(__call__) Returning last
result
Thread-17::DEBUG::2015-09-02
16:07:54,310::lvm::371::Storage.OperationMutex::(_reloadvgs) Operation 'lvm
reload operation' got the operation mutex
Thread-17::DEBUG::2015-09-02
16:07:54,312::lvm::291::Storage.Misc.excCmd::(cmd) /usr/bin/sudo -n
/usr/sbin/lvm vgs --config ' devices { preferred_names = ["^/dev/mapper/"]
ignore_suspended_devices=1 write_cache_state=0 disable_after_error_count=3
obtain_device_list_from_udev=0 filter = [
'\''a|/dev/mapper/33000000100000001|/dev/mapper/Hitachi_HDS721010DLE630_MSK523Y209VK0B|'\'',
'\''r|.*|'\'' ] } global { locking_type=1
prioritise_write_locks=1
wait_for_locks=1 use_lvmetad=0 } backup { retain_min = 50 retain_days =
0 } ' --noheadings --units b --nosuffix --separator '|'
--ignoreskippedcluster -o
uuid,name,attr,size,free,extent_size,extent_count,free_count,tags,vg_mda_size,vg_mda_free,lv_count,pv_count,pv_name
8b25f3be-7574-4f7a-8851-363129704e52 (cwd None)
Thread-17::DEBUG::2015-09-02
16:07:54,478::lvm::291::Storage.Misc.excCmd::(cmd) SUCCESS: <err> = '
WARNING: lvmetad is running but disabled. Restart lvmetad before enabling
it!\n'; <rc> = 0
Thread-17::DEBUG::2015-09-02
16:07:54,478::lvm::416::Storage.OperationMutex::(_reloadvgs) Operation 'lvm
reload operation' released the operation mutex
Thread-17::DEBUG::2015-09-02
16:07:54,479::persistentDict::192::Storage.PersistentDict::(__init__)
Created a persistent dict with LvMetadataRW backend
Thread-17::DEBUG::2015-09-02
16:07:54,479::blockSD::337::Storage.Misc.excCmd::(readlines) /usr/bin/dd
iflag=direct skip=0 bs=2048
if=/dev/8b25f3be-7574-4f7a-8851-363129704e52/metadata count=1 (cwd None)
Thread-17::DEBUG::2015-09-02
16:07:54,553::blockSD::337::Storage.Misc.excCmd::(readlines) SUCCESS: <err>
= '1+0 records in\n1+0 records out\n2048 bytes (2.0 kB) copied, 0.00107202
s, 1.9 MB/s\n'; <rc> = 0
Thread-17::DEBUG::2015-09-02
16:07:54,553::misc::260::Storage.Misc::(validateDDBytes) err: ['1+0 records
in', '1+0 records out', '2048 bytes (2.0 kB) copied, 0.00107202 s, 1.9
MB/s'], size: 2048
Thread-17::DEBUG::2015-09-02
16:07:54,553::persistentDict::234::Storage.PersistentDict::(refresh) read
lines (LvMetadataRW)=[]
Thread-17::DEBUG::2015-09-02
16:07:54,553::persistentDict::252::Storage.PersistentDict::(refresh) Empty
metadata
Thread-17::DEBUG::2015-09-02
16:07:54,553::persistentDict::192::Storage.PersistentDict::(__init__)
Created a persistent dict with VGTagMetadataRW backend
Thread-17::DEBUG::2015-09-02
16:07:54,554::lvm::504::Storage.OperationMutex::(_invalidatevgs) Operation
'lvm invalidate operation' got the operation mutex
Thread-17::DEBUG::2015-09-02
16:07:54,554::lvm::506::Storage.OperationMutex::(_invalidatevgs) Operation
'lvm invalidate operation' released the operation mutex
Thread-17::DEBUG::2015-09-02
16:07:54,554::lvm::514::Storage.OperationMutex::(_invalidatelvs) Operation
'lvm invalidate operation' got the operation mutex
Thread-17::DEBUG::2015-09-02
16:07:54,554::lvm::526::Storage.OperationMutex::(_invalidatelvs) Operation
'lvm invalidate operation' released the operation mutex
Thread-17::DEBUG::2015-09-02
16:07:54,554::lvm::371::Storage.OperationMutex::(_reloadvgs) Operation 'lvm
reload operation' got the operation mutex
Thread-17::DEBUG::2015-09-02
16:07:54,554::lvm::291::Storage.Misc.excCmd::(cmd) /usr/bin/sudo -n
/usr/sbin/lvm vgs --config ' devices { preferred_names = ["^/dev/mapper/"]
ignore_suspended_devices=1 write_cache_state=0 disable_after_error_count=3
obtain_device_list_from_udev=0 filter = [
'\''a|/dev/mapper/33000000100000001|/dev/mapper/Hitachi_HDS721010DLE630_MSK523Y209VK0B|'\'',
'\''r|.*|'\'' ] } global { locking_type=1
prioritise_write_locks=1
wait_for_locks=1 use_lvmetad=0 } backup { retain_min = 50 retain_days =
0 } ' --noheadings --units b --nosuffix --separator '|'
--ignoreskippedcluster -o
uuid,name,attr,size,free,extent_size,extent_count,free_count,tags,vg_mda_size,vg_mda_free,lv_count,pv_count,pv_name
8b25f3be-7574-4f7a-8851-363129704e52 (cwd None)
Thread-17::DEBUG::2015-09-02
16:07:54,685::lvm::291::Storage.Misc.excCmd::(cmd) SUCCESS: <err> = '
WARNING: lvmetad is running but disabled. Restart lvmetad before enabling
it!\n'; <rc> = 0
Thread-17::DEBUG::2015-09-02
16:07:54,686::lvm::416::Storage.OperationMutex::(_reloadvgs) Operation 'lvm
reload operation' released the operation mutex
Thread-17::DEBUG::2015-09-02
16:07:54,686::persistentDict::234::Storage.PersistentDict::(refresh) read
lines (VGTagMetadataRW)=['CLASS=Data', 'DESCRIPTION=hosted_storage',
'IOOPTIMEOUTSEC=10', 'LEASERETRIES=3', 'LEASETIMESEC=60',
'LOCKPOLICY=',
'LOCKRENEWALINTERVALSEC=5', 'LOGBLKSIZE=512', 'PHYBLKSIZE=4096',
'POOL_UUID=',
u'PV0=pv:33000000100000001,uuid:kTaQQh-4LCD-OghQ-cP5D-R7MM-aj6e-kTdQf0,pestart:0,pecount:397,mapoffset:0',
'ROLE=Regular', 'SDUUID=8b25f3be-7574-4f7a-8851-363129704e52',
'TYPE=ISCSI', 'VERSION=3',
'VGUUID=tJKiwH-Cn7v-QCxd-YQrg-MUxA-fbdC-kdga8m',
'_SHA_CKSUM=4a100ce5195650f43971d849835a8b3d8c0343da']
Thread-17::DEBUG::2015-09-02
16:07:54,687::resourceManager::421::Storage.ResourceManager::(registerNamespace)
Registering namespace '8b25f3be-7574-4f7a-8851-363129704e52_imageNS'
Thread-17::DEBUG::2015-09-02
16:07:54,687::resourceManager::421::Storage.ResourceManager::(registerNamespace)
Registering namespace '8b25f3be-7574-4f7a-8851-363129704e52_volumeNS'
Thread-17::DEBUG::2015-09-02
16:07:54,687::resourceManager::421::Storage.ResourceManager::(registerNamespace)
Registering namespace '8b25f3be-7574-4f7a-8851-363129704e52_lvmActivationNS'
Thread-17::DEBUG::2015-09-02
16:07:54,687::lvm::428::Storage.OperationMutex::(_reloadlvs) Operation 'lvm
reload operation' got the operation mutex
What should I do to bring the VM engine back?
2015-09-02 16:24 GMT+01:00 Simone Tiraboschi <stirabos(a)redhat.com>:
On Wed, Sep 2, 2015 at 10:49 AM, wodel youchi <wodel.youchi(a)gmail.com>
wrote:
> I will try this afternoon to do this, but just to clarify something.
>
> the hosted_engine setup creates it's own DC the hosted_DC, which contains
> the hosted engine storage domain, I am correct?
>
No, ovirt-hosted-engine-setup doesn't create a special datacenter. The
default is to add the host to the Default datacenter in the default cluster.
You could choose a different one from ovirt-hosted-engine-setup, simply
import the hosted-engine storage domain in the datacenter of the cluster
you selected.
In setup there is a question like this:
Local storage datacenter name is an internal name
and currently will not be shown in engine's admin UI.
Please enter local datacenter name
which ask about 'Local storage datacenter' which is basically the
description we were using for the storage pool.
> if yes, where will I import the hostedengine storage domain, into the
> default DC?
>
> 2015-09-02 8:47 GMT+01:00 Roy Golan <rgolan(a)redhat.com>:
>
>>
>>
>> On Wed, Sep 2, 2015 at 12:51 AM, wodel youchi <wodel.youchi(a)gmail.com>
>> wrote:
>>
>>> I could finally terminate the installation, but still no vm engine on
>>> webui
>>>
>>> I added a data domain, the default DC is up, but no engine VM.
>>>
>>
>>
>>
>> Good now you need to import the HostedEngine storage domain. Try to go
>> to
>>
>> *Storage -> Import Domain and put the path to the domain which you used
>> in the hosted-engine setup.*
>>
>>
>> *After the domain is imported, the engine will be imported
>> automatically. *
>>
>> *This whole process will become automatic eventually. (patch is written
>> currently)*
>>
>>
>>>
>>> 2015-09-01 21:22 GMT+01:00 wodel youchi <wodel.youchi(a)gmail.com>:
>>>
>>>> Something mounted on /rhev/data-center/mnt I'm not sure.
>>>>
>>>> there were directories, and under these directories there were other
>>>> directories (dom_md, ha_agent, images), and under them there were
symbolic
>>>> links to devices under /dev
>>>> (ids, inbox, leases, etc...) the devices pointed to the lvm partitions
>>>> created by the setup.
>>>>
>>>> but the mount command didn't show anything, unlike nfs, when I used
it
>>>> the mount and df commands showed up the engine's VM mount point.
>>>>
>>>>
>>>> 2015-09-01 20:16 GMT+01:00 Simone Tiraboschi
<stirabos(a)redhat.com>:
>>>>
>>>>>
>>>>>
>>>>> On Tue, Sep 1, 2015 at 7:29 PM, wodel youchi
<wodel.youchi(a)gmail.com>
>>>>> wrote:
>>>>>
>>>>>> Hi,
>>>>>>
>>>>>> After removing the -x from the sql files, the installation
>>>>>> terminated successfully, but ...
>>>>>>
>>>>>> I had a problem with vdsm, and error about permission denied
with
>>>>>> KVM module, so I restarted my machine.
>>>>>> After the reboot the ovirt-ha-agent stops complaining about the
>>>>>> vm.conf file not present in /var/rum/ovirt-hosted-engine-ha
>>>>>>
>>>>>> And the mount command doesn't show any iscsi mount, the disk
is
>>>>>> detected via fdisk -l
>>>>>> the lvs command returns all logical volumes created.
>>>>>>
>>>>>> I think it's a mount problem, but since there are many lv, I
don't
>>>>>> how to mount them manually.
>>>>>>
>>>>>
>>>>> Do you have something mounted under /rhev/data-center/mnt ?
>>>>> If not you probably hit this bug:
https://bugzilla.redhat.com/1258465
>>>>>
>>>>>
>>>>>
>>>>>> LV VG
>>>>>> Attr LSize Pool Origin
Data%
>>>>>> Meta% Move
>>>>>> Log Cpy%Sync Convert
>>>>>>
>>>>>> 3b894e23-429d-43bf-b6cd-6427a387799a
>>>>>> 5445bbee-bb3a-4e6d-9614-a0c9378fe078 -wi-ao---- 128,00m
>>>>>>
>>>>>>
>>>>>>
>>>>>> be78c0fd-52bf-445a-9555-64061029c2d9
>>>>>> 5445bbee-bb3a-4e6d-9614-a0c9378fe078 -wi-a----- 1,00g
>>>>>>
>>>>>>
>>>>>>
>>>>>> c9f74ffc-2eba-40a9-9c1c-f3b6d8e12657
>>>>>> 5445bbee-bb3a-4e6d-9614-a0c9378fe078 -wi-a----- 40,00g
>>>>>>
>>>>>>
>>>>>>
>>>>>> feede664-5754-4ca2-aeb3-af7aff32ed42
>>>>>> 5445bbee-bb3a-4e6d-9614-a0c9378fe078 -wi-a----- 128,00m
>>>>>>
>>>>>>
>>>>>>
>>>>>> ids
>>>>>>
5445bbee-bb3a-4e6d-9614-a0c9378fe078
>>>>>> -wi-ao---- 128,00m
>>>>>>
>>>>>>
>>>>>> inbox
>>>>>>
5445bbee-bb3a-4e6d-9614-a0c9378fe078
>>>>>> -wi-a----- 128,00m
>>>>>>
>>>>>>
>>>>>> leases
>>>>>>
5445bbee-bb3a-4e6d-9614-a0c9378fe078
>>>>>> -wi-a----- 2,00g
>>>>>>
>>>>>>
>>>>>> master
>>>>>>
5445bbee-bb3a-4e6d-9614-a0c9378fe078
>>>>>> -wi-a----- 1,00g
>>>>>>
>>>>>>
>>>>>> metadata
>>>>>> 5445bbee-bb3a-4e6d-9614-a0c9378fe078
-wi-a-----
>>>>>> 512,00m
>>>>>>
>>>>>>
>>>>>> outbox
>>>>>>
5445bbee-bb3a-4e6d-9614-a0c9378fe078
>>>>>> -wi-a----- 128,00m
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>> 2015-09-01 16:57 GMT+01:00 Simone Tiraboschi
<stirabos(a)redhat.com>:
>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> On Tue, Sep 1, 2015 at 5:08 PM, wodel youchi <
>>>>>>> wodel.youchi(a)gmail.com> wrote:
>>>>>>>
>>>>>>>> Hi again,
>>>>>>>>
>>>>>>>> I tried with the snapshot repository, but I am having
this error
>>>>>>>> while executing engine-setup
>>>>>>>>
>>>>>>>> [ INFO ] Creating/refreshing Engine database schema
>>>>>>>> [ ERROR ] Failed to execute stage 'Misc
configuration': Command
>>>>>>>> '/usr/share/ovirt-engine/dbscripts/schema.sh'
failed to execu
>>>>>>>> te
>>>>>>>> [ INFO ] DNF Performing DNF transaction rollback
>>>>>>>> [ INFO ] Rolling back database schema
>>>>>>>> [ INFO ] Clearing Engine database engine
>>>>>>>> [ ERROR ] Engine database rollback failed: must be owner
of
>>>>>>>> schema pg_catalog
>>>>>>>> [ INFO ] Stage: Clean up
>>>>>>>> Log file is located at
>>>>>>>>
/var/log/ovirt-engine/setup/ovirt-engine-setup-20150901153202-w0ds25.log
>>>>>>>> [ INFO ] Generating answer file
>>>>>>>>
'/var/lib/ovirt-engine/setup/answers/20150901153939-setup.conf'
>>>>>>>> [ INFO ] Stage: Pre-termination
>>>>>>>> [ INFO ] Stage: Termination
>>>>>>>> [ ERROR ] Execution of setup failed
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>> and in the deployement log I have these errors
>>>>>>>>
>>>>>>>> Saving custom users permissions on database objects...
>>>>>>>> upgrade script detected a change in Config, View or
Stored
>>>>>>>> Procedure...
>>>>>>>> Running upgrade shell script
>>>>>>>>
'/usr/share/ovirt-engine/dbscripts/upgrade/pre_upgrade/0000_config.sql'...
>>>>>>>> Running upgrade shell script
>>>>>>>>
'/usr/share/ovirt-engine/dbscripts/upgrade/pre_upgrade/0010_custom.sql'...
>>>>>>>> Running upgrade shell script
>>>>>>>>
'/usr/share/ovirt-engine/dbscripts/upgrade/pre_upgrade/0020_add_materialized_views_table.sql'...
>>>>>>>> Running upgrade shell script
>>>>>>>>
'/usr/share/ovirt-engine/dbscripts/upgrade/pre_upgrade/0030_materialized_views_extensions.sql'...
>>>>>>>> Running upgrade shell script
>>>>>>>>
'/usr/share/ovirt-engine/dbscripts/pre_upgrade/0040_extend_installed_by_column.sql'...
>>>>>>>>
>>>>>>>>
>>>>>>>> 2015-09-01 15:39:35 DEBUG
>>>>>>>> otopi.plugins.ovirt_engine_setup.ovirt_engine.db.schema
plugin.execute:941
>>>>>>>> execute-output:
['/usr/share/ovirt-engine/dbscripts/schema.sh', '-s',
>>>>>>>> 'localhost', '-p', '5432',
'-u', 'engine', '-d', 'engine', '-l',
'/var/log
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>>
>>>>>>>>
*/ovirt-engine/setup/ovirt-engine-setup-20150901153202-w0ds25.log',
>>>>>>>> '-c', 'apply']
>>>>>>>>
stderr:/usr/share/ovirt-engine/dbscripts/upgrade/pre_upgrade/0000_config.sql:
>>>>>>>> ligne 1: /bin : is a
>>>>>>>>
directory/usr/share/ovirt-engine/dbscripts/upgrade/pre_upgrade/0000_config.sql:
>>>>>>>> ligne 2: DATABASE : command not
>>>>>>>>
found/usr/share/ovirt-engine/dbscripts/upgrade/pre_upgrade/0000_config.sql:
>>>>>>>> ligne 4: This : command not
>>>>>>>>
found/usr/share/ovirt-engine/dbscripts/upgrade/pre_upgrade/0000_config.sql:
>>>>>>>> ligne 5: The : command not
>>>>>>>>
found/usr/share/ovirt-engine/dbscripts/upgrade/pre_upgrade/0000_config.sql:
>>>>>>>> ligne 6: Add : command not
>>>>>>>>
found/usr/share/ovirt-engine/dbscripts/upgrade/pre_upgrade/0000_config.sql:
>>>>>>>> ligne 7: syntax error near the symbole unexpected « (
>>>>>>>>
»/usr/share/ovirt-engine/dbscripts/upgrade/pre_upgrade/0000_config.sql:
>>>>>>>> ligne 7: ` Update section (w/o overriding current
>>>>>>>>
value)'/usr/share/ovirt-engine/dbscripts/upgrade/pre_upgrade/0010_custom.sql:
>>>>>>>> ligne 1: /bin : is a
>>>>>>>>
directory/usr/share/ovirt-engine/dbscripts/upgrade/pre_upgrade/0010_custom.sql:
>>>>>>>> ligne 2: Currently : command not
>>>>>>>>
found/usr/share/ovirt-engine/dbscripts/upgrade/pre_upgrade/0010_custom.sql:
>>>>>>>> ligne 3: This : command not
>>>>>>>>
found/usr/share/ovirt-engine/dbscripts/upgrade/pre_upgrade/0010_custom.sql:
>>>>>>>> ligne 4: This : command not
>>>>>>>>
found/usr/share/ovirt-engine/dbscripts/upgrade/pre_upgrade/0010_custom.sql:
>>>>>>>> ligne 5: So, : command not
>>>>>>>>
found/usr/share/ovirt-engine/dbscripts/upgrade/pre_upgrade/0010_custom.sql:
>>>>>>>> ligne 6: Since : command not
>>>>>>>>
found/usr/share/ovirt-engine/dbscripts/upgrade/pre_upgrade/0010_custom.sql:
>>>>>>>> ligne 7: bin/ : is a
>>>>>>>>
directory/usr/share/ovirt-engine/dbscripts/upgrade/pre_upgrade/0010_custom.sql:
>>>>>>>> ligne 9: update : command not
>>>>>>>>
found/usr/share/ovirt-engine/dbscripts/upgrade/pre_upgrade/0010_custom.sql:
>>>>>>>> ligne 10: *
>>>>>>>>
>>>>>>>>
>>>>>>>> *syntax error near the symbole unexpected « (
>>>>>>>>
»/usr/share/ovirt-engine/dbscripts/upgrade/pre_upgrade/0010_custom.sql:
>>>>>>>> ligne 10: `and exists(select 1 from schema_version where
version =
>>>>>>>> '03010250' and current =
>>>>>>>>
true);'/usr/share/ovirt-engine/dbscripts/upgrade/pre_upgrade/0020_add_materialized_views_table.sql:
>>>>>>>> ligne 1: -- : command not
>>>>>>>>
found/usr/share/ovirt-engine/dbscripts/upgrade/pre_upgrade/0020_add_materialized_views_table.sql:
>>>>>>>> ligne 2: **syntax error near the symbole unexpected « (
»*
>>>>>>>>
/usr/share/ovirt-engine/dbscripts/upgrade/pre_upgrade/0020_add_materialized_views_table.sql:
>>>>>>>> ligne 2: `CREATE FUNCTION
__temp__0030_add_materialized_views_table()'
>>>>>>>>
/usr/share/ovirt-engine/dbscripts/upgrade/pre_upgrade/0030_materialized_views_extensions.sql:
>>>>>>>> ligne 1: -- : commande introuvable
>>>>>>>>
/usr/share/ovirt-engine/dbscripts/upgrade/pre_upgrade/0030_materialized_views_extensions.sql:
>>>>>>>> ligne 2: erreur de syntaxe près du symbole inattendu « (
»
>>>>>>>>
/usr/share/ovirt-engine/dbscripts/upgrade/pre_upgrade/0030_materialized_views_extensions.sql:
>>>>>>>> ligne 2: `select
fn_db_add_column('materialized_views',
>>>>>>>> 'min_refresh_rate_in_sec', 'int default
0');'
>>>>>>>>
/usr/share/ovirt-engine/dbscripts/upgrade/pre_upgrade/0040_extend_installed_by_column.sql:
>>>>>>>> ligne 1: erreur de syntaxe près du symbole inattendu « (
»
>>>>>>>>
/usr/share/ovirt-engine/dbscripts/upgrade/pre_upgrade/0040_extend_installed_by_column.sql:
>>>>>>>> ligne 1: `ALTER TABLE schema_version ALTER COLUMN
installed_by TYPE
>>>>>>>> varchar(63);'
>>>>>>>> 2015-09-01 15:39:35 DEBUG otopi.context
context._executeMethod:156
>>>>>>>> method exception
>>>>>>>> Traceback (most recent call last):
>>>>>>>> File
"/usr/lib/python2.7/site-packages/otopi/context.py", line
>>>>>>>> 146, in _executeMethod
>>>>>>>> method['method']()
>>>>>>>> File
>>>>>>>>
"/usr/share/ovirt-engine/setup/bin/../plugins/ovirt-engine-setup/ovirt-engine/db/schema.py",
>>>>>>>> line 291, in _misc
>>>>>>>> oenginecons.EngineDBEnv.PGPASS_FILE
>>>>>>>> File
"/usr/lib/python2.7/site-packages/otopi/plugin.py", line
>>>>>>>> 946, in execute
>>>>>>>> command=args[0],
>>>>>>>> RuntimeError: Command
>>>>>>>> '/usr/share/ovirt-engine/dbscripts/schema.sh'
failed to execute
>>>>>>>> 2015-09-01 15:39:35 ERROR otopi.context
context._executeMethod:165
>>>>>>>> Failed to execute stage 'Misc configuration':
Command
>>>>>>>> '/usr/share/ovirt-engine/dbscripts/schema.sh'
failed to execute
>>>>>>>> 2015-09-01 15:39:35 DEBUG otopi.transaction
transaction.abort:134
>>>>>>>> aborting 'DNF Transaction'
>>>>>>>> 2015-09-01 15:39:35 DEBUG
>>>>>>>> otopi.plugins.otopi.packagers.dnfpackager
dnfpackager.verbose:90 DNF
>>>>>>>> Closing transaction with rollback
>>>>>>>> 2015-09-01 15:39:35 INFO
otopi.plugins.otopi.packagers.dnfpackager
>>>>>>>> dnfpackager.info:94 DNF Performing DNF transaction
rollback
>>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>> It was an issue with package building: all the sql files
where
>>>>>>> executable and so the issue.
>>>>>>> We fixed it and tomorrow build should be OK. If you prefer
to
>>>>>>> continue right now simply recursively remove the x attribute
on each sql
>>>>>>> file under /usr/share/ovirt-engine/dbscripts
>>>>>>>
>>>>>>>
>>>>>>>
>>>>>>>>
>>>>>>>> 2015-09-01 13:04 GMT+01:00 Simone Tiraboschi
<stirabos(a)redhat.com>
>>>>>>>> :
>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> On Tue, Sep 1, 2015 at 12:40 PM, Yedidyah Bar David
<
>>>>>>>>> didi(a)redhat.com> wrote:
>>>>>>>>>
>>>>>>>>>> On Tue, Sep 1, 2015 at 1:25 PM, wodel youchi
<
>>>>>>>>>> wodel.youchi(a)gmail.com> wrote:
>>>>>>>>>> > Hi,
>>>>>>>>>> >
>>>>>>>>>> > I am using the repo of the 3.6 version
>>>>>>>>>> >
(
http://plain.resources.ovirt.org/pub/ovirt-3.6-pre/rpm/fc22/)
>>>>>>>>>> >
>>>>>>>>>> > I installed the ovirt-hosted-engine-setup
with it's
>>>>>>>>>> dependencies,and the
>>>>>>>>>> > ovirt-hosted-engine-ha package is one of
them.
>>>>>>>>>> >
>>>>>>>>>> > Correction: The problem with this version
>>>>>>>>>> >
>>>>>>>>>>
ovirt-hosted-engine-ha-1.3.0-0.0.master.20150819082341.20150819082338.git183a4ff.fc22.noarch.rpm,
>>>>>>>>>> > is that after the installation is done, the
service
>>>>>>>>>> ovirt-ha-agent crashes
>>>>>>>>>> > after being started, see the bug :
>>>>>>>>>> >
>>>>>>>>>> >
https://bugzilla.redhat.com/show_bug.cgi?id=1254745
>>>>>>>>>> >
>>>>>>>>>> > A new version was produced, I downloaded it
manually a few
>>>>>>>>>> days ago, this is
>>>>>>>>>> > it :
>>>>>>>>>> >
>>>>>>>>>>
ovirt-hosted-engine-ha-1.3.0-0.0.master.20150820064645.20150820064642.git02529e0.fc22.noarch.rpm
>>>>>>>>>> >
>>>>>>>>>> > This one did correct the problem, but
it's not present anymore
>>>>>>>>>> on the
>>>>>>>>>> > repository.
>>>>>>>>>>
>>>>>>>>>> Was this on ovirt-3.6-pre?
>>>>>>>>>>
>>>>>>>>>> ovirt-3.6-snapshot has a newer version.
>>>>>>>>>>
>>>>>>>>>> >
>>>>>>>>>> > For Simone: yes I did added an NFS4 data
domain, but no
>>>>>>>>>> success so far, no
>>>>>>>>>> > VM engine present.
>>>>>>>>>> >
>>>>>>>>>> > Regards.
>>>>>>>>>> >
>>>>>>>>>> > 2015-09-01 11:15 GMT+01:00 Simone Tiraboschi
<
>>>>>>>>>> stirabos(a)redhat.com>:
>>>>>>>>>> >>
>>>>>>>>>> >>
>>>>>>>>>> >>
>>>>>>>>>> >> On Tue, Sep 1, 2015 at 11:46 AM,
Yedidyah Bar David <
>>>>>>>>>> didi(a)redhat.com>
>>>>>>>>>> >> wrote:
>>>>>>>>>> >>>
>>>>>>>>>> >>> On Tue, Sep 1, 2015 at 11:25 AM,
wodel youchi <
>>>>>>>>>> wodel.youchi(a)gmail.com>
>>>>>>>>>> >>> wrote:
>>>>>>>>>> >>> > Hi,
>>>>>>>>>> >>> >
>>>>>>>>>> >>> > Another test of ovirt
hosted-engine on FC22 using ovirt
>>>>>>>>>> 3.6 Beta3.
>>>>>>>>>> >>> > VM engine is also a FC22
>>>>>>>>>> >>> >
>>>>>>>>>> >>> > Problem:
>>>>>>>>>> >>> > - No VM engine on webui
>>>>>>>>>> >>>
>>>>>>>>>> >>> This is still not supported,
see/follow [1].
>>>>>>>>>> >>>
>>>>>>>>>> >>> [1]
https://bugzilla.redhat.com/show_bug.cgi?id=1224889
>>>>>>>>>> >>
>>>>>>>>>> >>
>>>>>>>>>> >> ? :-)
>>>>>>>>>>
>>>>>>>>>> Sorry :-(
>>>>>>>>>>
>>>>>>>>>>
https://bugzilla.redhat.com/show_bug.cgi?id=1160094
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> This is just about editing the VM from the web GUI
but in order
>>>>>>>>> to be able to edit the engine VM you should be able
at least to find the
>>>>>>>>> engine VM in the engine as it was also in 3.5
>>>>>>>>>
>>>>>>>>> I'll try to reproduce verifying another patch
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>> I see that all patches there are merged, but bug
is in POST.
>>>>>>>>>>
>>>>>>>>>> >>
>>>>>>>>>> >> Did you also try adding an additional
storage domain for
>>>>>>>>>> regular VMs?
>>>>>>>>>> >> engine-VM will be shown on the engine
only when you add at
>>>>>>>>>> least one
>>>>>>>>>> >> additional storage domain for regulars
VM and the whole
>>>>>>>>>> datacenter goes up:
>>>>>>>>>> >>
https://bugzilla.redhat.com/show_bug.cgi?id=1222010#c1
>>>>>>>>>> >>
>>>>>>>>>> >>
>>>>>>>>>> >>>
>>>>>>>>>> >>> > Test Environment
>>>>>>>>>> >>> > Just two machines:
>>>>>>>>>> >>> >
>>>>>>>>>> >>> > 1 - Machine 1 used as storage
:
>>>>>>>>>> >>> > - iscsi target with a raw
file for the VM engine storage
>>>>>>>>>> >>> > - NFS4 for other data
domains
>>>>>>>>>> >>> >
>>>>>>>>>> >>> > 2 - Machine 2 used as
hypervisor
>>>>>>>>>> >>> >
>>>>>>>>>> >>> >
>>>>>>>>>> >>> > The installation went without
problem, but as always, the
>>>>>>>>>> VM engine is
>>>>>>>>>> >>> > not
>>>>>>>>>> >>> > present on the webui.
>>>>>>>>>> >>> >
>>>>>>>>>> >>> >
>>>>>>>>>> >>> > PS:
>>>>>>>>>> >>> > 1- I gave the VM engine just
2Gb of memory since I don't
>>>>>>>>>> have too much
>>>>>>>>>> >>> > RAM
>>>>>>>>>> >>> > on hypervisor, could that be
the cause of the problem?
>>>>>>>>>> >>>
>>>>>>>>>> >>> Shouldn't be related
>>>>>>>>>> >>>
>>>>>>>>>> >>> > 2- This version of
>>>>>>>>>> >>> >
>>>>>>>>>> >>> >
>>>>>>>>>>
ovirt-hosted-engine-ha-1.3.0-0.0.master.20150424113926.20150424113923.git7c14f4c.fc22.noarch.rpm
>>>>>>>>>> >>> > package is causing the
ovirt-ha-agent to crash, it was
>>>>>>>>>> replaced with
>>>>>>>>>> >>> > another
>>>>>>>>>> >>> > which I still have
>>>>>>>>>> >>> >
>>>>>>>>>> >>> >
>>>>>>>>>>
ovirt-hosted-engine-ha-1.3.0-0.0.master.20150820064645.20150820064642.git02529e0.fc22.noarch.rpm,
>>>>>>>>>> >>> > but it's not present on the
repository, I had to update
>>>>>>>>>> the package
>>>>>>>>>> >>> > manually
>>>>>>>>>> >>> > at the end of
ovirt-hosted-engine-setup installation.
>>>>>>>>>> >>>
>>>>>>>>>> >>> Not sure I follow.
>>>>>>>>>> >>>
>>>>>>>>>> >>> What exact repo was used?
>>>>>>>>>> >>>
>>>>>>>>>> >>> hosted-engine --deploy does not
update/install packages for
>>>>>>>>>> you (as
>>>>>>>>>> >>> does engine-setup),
>>>>>>>>>> >>> it's up to you to make sure what
you want/need is installed
>>>>>>>>>> prior to
>>>>>>>>>> >>> running it.
>>>>>>>>>> >>>
>>>>>>>>>> >>> Best,
>>>>>>>>>> >>> --
>>>>>>>>>> >>> Didi
>>>>>>>>>> >>>
_______________________________________________
>>>>>>>>>> >>> Users mailing list
>>>>>>>>>> >>> Users(a)ovirt.org
>>>>>>>>>> >>>
http://lists.ovirt.org/mailman/listinfo/users
>>>>>>>>>> >>
>>>>>>>>>> >>
>>>>>>>>>> >
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> --
>>>>>>>>>> Didi
>>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>
>>>>>>>
>>>>>>
>>>>>
>>>>
>>>
>>
>