On Thu, Sep 3, 2015 at 2:07 AM, wodel youchi <wodel.youchi(a)gmail.com> wrote:
Hi again,
I had to restart the installation all over, I used the freshly pushed new
packages.
I had two problems:
1 - the engine's setup didn't terminate correctly when I chose to use
ovirt-vmconsole with this error
[ INFO ] Restarting ovirt-vmconsole proxy service
*[ ERROR ] **Failed to execute stage 'Closing up': Failed to stop service
'ovirt-vmconsole-proxy-sshd'*
[ INFO ] Stage: Clean up
Log file is located at
/var/log/ovirt-engine/setup/ovirt-engine-setup-20150903000415-6egi46.log
[ INFO ] Generating answer file
'/var/lib/ovirt-engine/setup/answers/20150903001209-setup.conf'
[ INFO ] Stage: Pre-termination
[ INFO ] Stage: Termination
[ ERROR ] Execution of setup failed
So I executed engine-cleanup which terminate with this error
[ INFO ] Clearing Engine database engine
*[ ERROR ] Failed to execute stage 'Misc configuration': must be owner of
schema pg_catalog *
[ INFO ] Stage: Clean up
Log file is located at
/var/log/ovirt-engine/setup/ovirt-engine-remove-20150903001440-da1u76.log
[ INFO ] Generating answer file
'/var/lib/ovirt-engine/setup/answers/20150903001513-cleanup.conf'
[ INFO ] Stage: Pre-termination
[ INFO ] Stage: Termination
[ ERROR ] Execution of cleanup failed
And then, I executed again engine-setup without ovirt-vmconsole
This time the setup completed.
2 - I added a NFS4 storage domain to the default DC (Default), the DC went
up, and then I tried to import the hosted-engine storage domain, but
without success.
click on import, choose iscsi, connect to the target, scan, login, but no
device chown !!! (iscsi.jpeg)
Can you please attach the relevant VDSM logs from the host you used to were
using to import that storage domain?
The only new thing I had, is the disk of the VM engine being shown under
> disks tab.
>
> 2015-09-02 19:50 GMT+01:00 wodel youchi <wodel.youchi(a)gmail.com>:
>
>> I found this on vdsm log
>>
>> Thread-432::DEBUG::2015-09-02
>> 19:37:30,854::bindingxmlrpc::1256::vds::(wrapper) client [127.0.0.1]::call
>> vmGetStats with ('ab
>> 1dc1a9-b6e9-4890-8485-1019da2f328f',) {}
>> Thread-432::DEBUG::2015-09-02
>> 19:37:30,854::bindingxmlrpc::1263::vds::(wrapper) return vmGetStats with
>> {'status': {'message':
>> '*Virtual machine does not exist*', 'code': 1}}
>>
>>
>> I really don't understand anything
>>
>> 2015-09-02 17:01 GMT+01:00 wodel youchi <wodel.youchi(a)gmail.com>:
>>
>>> Thanks,
>>>
>>> but before that I stuck again with the storage of the VM engine not
>>> detected after reboot.
>>>
>>> the /rhev is populated, but ovirt-ha-agent crashes with
>>>
>>> MainThread::INFO::2015-09-02
>>>
16:12:20,261::brokerlink::129::ovirt_hosted_engine_ha.lib.brokerlink.BrokerLink::(start_monitor)
>>> Starting monitor engine-health, options {'use_ssl': 'true',
'vm_uuid':
>>> 'ab1dc1a9-b6e9-4890-8485-1019da2f328f', 'address': '0
>>> '}
>>> MainThread::INFO::2015-09-02
>>>
16:12:20,283::brokerlink::140::ovirt_hosted_engine_ha.lib.brokerlink.BrokerLink::(start_monitor)
>>> Success, id 139994237094736
>>> MainThread::INFO::2015-09-02
>>>
16:12:20,702::brokerlink::178::ovirt_hosted_engine_ha.lib.brokerlink.BrokerLink::(set_storage_do
>>> main) Success, id 139994236985168
>>> MainThread::INFO::2015-09-02
>>>
16:12:20,702::hosted_engine::574::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_ini
>>> tialize_broker) Broker initialized, all submonitors started
>>> MainThread::INFO::2015-09-02
>>>
16:12:20,799::hosted_engine::678::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(_ini
>>> tialize_sanlock) Ensuring lease for lockspace hosted-engine, host id 1
>>> is acquired (file: /var/run/vdsm/storage/8b25f3be-7574
>>>
-4f7a-8851-363129704e52/a44d1302-3165-4632-9d99-3e035dfc3ac7/0f260ab0-3631-4c71-b332-c6c7f67f7342)
>>>
>>> MainThread::INFO::2015-09-02
>>>
16:12:20,800::hosted_engine::401::ovirt_hosted_engine_ha.agent.hosted_engine.HostedEngine::(star
>>> t_monitoring) Reloading vm.conf from the shared storage domain
>>> MainThread::ERROR::2015-09-02
>>>
16:12:20,927::agent::201::ovirt_hosted_engine_ha.agent.agent.Agent::(_run_agent)
>>>
>>> *Error: ''Configuration value not found:
>>> file=/var/run/ovirt-hosted-engine-ha/vm.conf, key=memSize'' - trying
to
>>> restart agent *
>>> MainThread::WARNING::2015-09-02
>>>
16:12:25,932::agent::204::ovirt_hosted_engine_ha.agent.agent.Agent::(_run_agent)
>>> Restarting a
>>> gent, attempt '9'
>>> MainThread::ERROR::2015-09-02
>>>
16:12:25,933::agent::206::ovirt_hosted_engine_ha.agent.agent.Agent::(_run_agent)
>>> Too many error
>>> s occurred, giving up. Please review the log and consider filing a bug.
>>> MainThread::INFO::2015-09-02
>>> 16:12:25,933::agent::143::ovirt_hosted_engine_ha.agent.agent.Agent::(run)
>>> Agent shutting down
>>>
>>>
>>> I restared vdsm ha-agent and broker-agent wihtout success
>>>
>>> When executed
>>> [root@noveria ~]# hosted-engine --vm-status
>>> You must run deploy first
>>>
>>>
>>> I got this
>>>
>>> [root@noveria ~]# tree /var/run/vdsm/storage/
>>> /var/run/vdsm/storage/
>>> └── 8b25f3be-7574-4f7a-8851-363129704e52
>>> ├── 8e49032f-680b-40c2-b422-80d86dc7beda
>>> │ └── f05762e5-e8cd-45e7-ac19-303c1ade79d1 ->
>>> /dev/8b25f3be-7574-4f7a-8851-363129704e52/f05762e5-e8cd-45e7-ac19-303c1
>>> ade79d1
>>> ├── a44d1302-3165-4632-9d99-3e035dfc3ac7
>>> │ └── 0f260ab0-3631-4c71-b332-c6c7f67f7342 ->
>>> /dev/8b25f3be-7574-4f7a-8851-363129704e52/0f260ab0-3631-4c71-b332-c6c7f
>>> 67f7342
>>> ├── a5475e57-c6f5-4dc5-a3f2-7fb782d613a7
>>> │ └── ae352fab-7477-4376-aa27-04c321b4fbd1 ->
>>> /dev/8b25f3be-7574-4f7a-8851-363129704e52/ae352fab-7477-4376-aa27-04c32
>>> 1b4fbd1
>>> └── bf3bdae1-7318-4443-a19b-7371de30b982
>>> └── cbb10cf0-9600-465e-aed9-412f7157706b ->
>>> /dev/8b25f3be-7574-4f7a-8851-363129704e52/cbb10cf0-9600-465e-aed9-412f7
>>> 157706b
>>>
>>> and this
>>> [root@noveria rhev]# tree
>>> .
>>> └── data-center
>>> ├── 00000001-0001-0001-0001-000000000221
>>> └── mnt
>>> ├── blockSD
>>> │ └── 8b25f3be-7574-4f7a-8851-363129704e52
>>> │ ├── dom_md
>>> │ │ ├── ids ->
>>> /dev/8b25f3be-7574-4f7a-8851-363129704e52/ids
>>> │ │ ├── inbox ->
>>> /dev/8b25f3be-7574-4f7a-8851-363129704e52/inbox
>>> │ │ ├── leases ->
>>> /dev/8b25f3be-7574-4f7a-8851-363129704e52/leases
>>> │ │ ├── master ->
>>> /dev/8b25f3be-7574-4f7a-8851-363129704e52/master
>>> │ │ ├── metadata ->
>>> /dev/8b25f3be-7574-4f7a-8851-363129704e52/metadata
>>> │ │ └── outbox ->
>>> /dev/8b25f3be-7574-4f7a-8851-363129704e52/outbox
>>> │ ├── ha_agent
>>> │ │ ├── hosted-engine.lockspace ->
>>> /var/run/vdsm/storage/8b25f3be-7574-4f7a-8851-363129704e52/a44d1302-3165-4
>>> 632-9d99-3e035dfc3ac7/0f260ab0-3631-4c71-b332-c6c7f67f7342
>>> │ │ └── hosted-engine.metadata ->
>>> /var/run/vdsm/storage/8b25f3be-7574-4f7a-8851-363129704e52/8e49032f-680b-40
>>> c2-b422-80d86dc7beda/f05762e5-e8cd-45e7-ac19-303c1ade79d1
>>> │ └── images
>>> │ ├── 8e49032f-680b-40c2-b422-80d86dc7beda
>>> │ │ └── f05762e5-e8cd-45e7-ac19-303c1ade79d1 ->
>>> /dev/8b25f3be-7574-4f7a-8851-363129704e52/f05762e5-e8cd-4
>>> 5e7-ac19-303c1ade79d1
>>> │ ├── a44d1302-3165-4632-9d99-3e035dfc3ac7
>>> │ │ └── 0f260ab0-3631-4c71-b332-c6c7f67f7342 ->
>>> /dev/8b25f3be-7574-4f7a-8851-363129704e52/0f260ab0-3631-4
>>> c71-b332-c6c7f67f7342
>>> │ ├── a5475e57-c6f5-4dc5-a3f2-7fb782d613a7
>>> │ │ └── ae352fab-7477-4376-aa27-04c321b4fbd1 ->
>>> /dev/8b25f3be-7574-4f7a-8851-363129704e52/ae352fab-7477-4
>>> 376-aa27-04c321b4fbd1
>>> │ └── bf3bdae1-7318-4443-a19b-7371de30b982
>>> │ └── cbb10cf0-9600-465e-aed9-412f7157706b ->
>>> /dev/8b25f3be-7574-4f7a-8851-363129704e52/cbb10cf0-9600-4
>>> 65e-aed9-412f7157706b
>>> ├── openSuse.wodel.wd:_nvms
>>> └── _var_lib_ovirt-hosted-engine-setup_tmp2fNoEf
>>>
>>>
>>> Here I did find some symblic links blinking (not present) like this one
>>> hosted-engine.metadata ->
>>> /var/run/vdsm/storage/8b25f3be-7574-4f7a-8851-363129704e52/8e49032f-680b-40
>>> c2-b422-80d86dc7beda/*f05762e5-e8cd-45e7-ac19-303c1ade79d1*
>>>
>>>
>>> the lvscan command showed that the lv concerned is inactive, is this
>>> correct?
>>> [root@noveria ~]# lvscan
>>> File descriptor 9 (/dev/dri/card0) leaked on lvscan invocation. Parent
>>> PID 2935: bash
>>> ACTIVE
'/dev/8b25f3be-7574-4f7a-8851-363129704e52/metadata'
>>> [512,00 MiB] inherit
>>> ACTIVE
'/dev/8b25f3be-7574-4f7a-8851-363129704e52/outbox'
>>> [128,00 MiB] inherit
>>> ACTIVE
'/dev/8b25f3be-7574-4f7a-8851-363129704e52/leases'
>>> [2,00 GiB] inherit
>>> ACTIVE '/dev/8b25f3be-7574-4f7a-8851-363129704e52/ids'
>>> [128,00 MiB] inherit
>>> ACTIVE '/dev/8b25f3be-7574-4f7a-8851-363129704e52/inbox'
>>> [128,00 MiB] inherit
>>> ACTIVE
'/dev/8b25f3be-7574-4f7a-8851-363129704e52/master'
>>> [1,00 GiB] inherit
>>> inactive
>>>
'/dev/8b25f3be-7574-4f7a-8851-363129704e52/ae352fab-7477-4376-aa27-04c321b4fbd1'
>>> [1,00 GiB] inherit
>>> ACTIVE
>>>
'/dev/8b25f3be-7574-4f7a-8851-363129704e52/0f260ab0-3631-4c71-b332-c6c7f67f7342'
>>> [128,00 MiB] inherit
>>> *inactive
>>>
'/dev/8b25f3be-7574-4f7a-8851-363129704e52/f05762e5-e8cd-45e7-ac19-303c1ade79d1'
>>> [128,00 MiB] inherit *
>>> inactive
>>>
'/dev/8b25f3be-7574-4f7a-8851-363129704e52/cbb10cf0-9600-465e-aed9-412f7157706b'
>>> [40,00 GiB] inherit
>>>
>>>
>>>
>>> and this
>>> [root@noveria ~]# vdsClient -s 0 prepareImage
>>> "00000000-0000-0000-0000-000000000000"
>>> "8b25f3be-7574-4f7a-8851-363129704e52" "
>>> bf3bdae1-7318-4443-a19b-7371de30b982"
>>> "cbb10cf0-9600-465e-aed9-412f7157706b"
>>> {'domainID': '8b25f3be-7574-4f7a-8851-363129704e52',
>>> 'imageID': 'bf3bdae1-7318-4443-a19b-7371de30b982',
>>> 'leaseOffset': 112197632,
>>> 'leasePath':
'/dev/8b25f3be-7574-4f7a-8851-363129704e52/leases',
>>> 'path':
>>>
'/rhev/data-center/mnt/blockSD/8b25f3be-7574-4f7a-8851-363129704e52/images/bf3bdae1-7318-4443-a19b-7371de30b982/cbb1
>>> 0cf0-9600-465e-aed9-412f7157706b',
>>> 'volType': 'path',
>>> 'volumeID': 'cbb10cf0-9600-465e-aed9-412f7157706b'}
>>>
>>>
>>> and
>>> [root@noveria ~]# vdsClient -s 0 getStorageDomainInfo
>>> 8b25f3be-7574-4f7a-8851-363129704e52
>>> uuid = 8b25f3be-7574-4f7a-8851-363129704e52
>>> vguuid = tJKiwH-Cn7v-QCxd-YQrg-MUxA-fbdC-kdga8m
>>> state = OK
>>> version = 3
>>> role = Regular
>>> type = ISCSI
>>> class = Data
>>> pool = []
>>> name = hosted_storage
>>>
>>>
>>> [root@noveria ~]# lvs
>>> File descriptor 9 (/dev/dri/card0) leaked on lvs invocation. Parent PID
>>> 3105: bash
>>> LV VG
>>> Attr LSize Pool Origin Data%
>>> Meta% Move
>>> Log Cpy%Sync Convert
>>> 0f260ab0-3631-4c71-b332-c6c7f67f7342
>>> 8b25f3be-7574-4f7a-8851-363129704e52 -wi-ao---- 128,00m
>>>
>>>
>>> ae352fab-7477-4376-aa27-04c321b4fbd1
>>> 8b25f3be-7574-4f7a-8851-363129704e52 -wi------- 1,00g
>>>
>>>
>>> cbb10cf0-9600-465e-aed9-412f7157706b
>>> 8b25f3be-7574-4f7a-8851-363129704e52 -wi-a----- 40,00g
>>>
>>>
>>> f05762e5-e8cd-45e7-ac19-303c1ade79d1
>>> 8b25f3be-7574-4f7a-8851-363129704e52 -wi------- 128,00m
>>>
>>>
>>> ids
>>> 8b25f3be-7574-4f7a-8851-363129704e52
>>> -wi-a----- 128,00m
>>>
>>> inbox
>>> 8b25f3be-7574-4f7a-8851-363129704e52
>>> -wi-a----- 128,00m
>>>
>>> leases
>>> 8b25f3be-7574-4f7a-8851-363129704e52
>>> -wi-a----- 2,00g
>>>
>>> master
>>> 8b25f3be-7574-4f7a-8851-363129704e52
>>> -wi-a----- 1,00g
>>>
>>> metadata
>>> 8b25f3be-7574-4f7a-8851-363129704e52 -wi-a-----
>>> 512,00m
>>>
>>> outbox
>>> 8b25f3be-7574-4f7a-8851-363129704e52
>>> -wi-a----- 128,00m
>>>
>>>
>>>
>>> VDSM logs doesn't show me anything
>>> MainThread::INFO::2015-09-01 23:34:49,551::vdsm::166::vds::(run)
>>> <WorkerThread(Thread-4, started daemon 139990108333824)>
>>> MainThread::INFO::2015-09-01 23:34:49,552::vdsm::166::vds::(run)
>>> <WorkerThread(Thread-3, started daemon 139990116726528)>
>>> MainThread::INFO::2015-09-02 16:07:49,510::vdsm::156::vds::(run) (PID:
>>> 1554) I am the actual vdsm 4.17.3-12.git7288ef7.fc22 noveria.wodel.wd
>>> (4.1.6-200.fc22.x86_64)
>>> MainThread::DEBUG::2015-09-02
>>>
16:07:49,524::resourceManager::421::Storage.ResourceManager::(registerNamespace)
>>> Registering namespace 'Storage'
>>> MainThread::DEBUG::2015-09-02
>>> 16:07:49,524::threadPool::29::Storage.ThreadPool::(__init__) Enter -
>>> numThreads: 10, waitTimeout: 3, maxTasks: 500
>>> MainThread::DEBUG::2015-09-02
>>> 16:07:49,526::fileUtils::143::Storage.fileUtils::(createdir) Creating
>>> directory: /rhev/data-center/mnt mode: None
>>> MainThread::WARNING::2015-09-02
>>> 16:07:49,526::fileUtils::152::Storage.fileUtils::(createdir) Dir
>>> /rhev/data-center/mnt already exists
>>> MainThread::DEBUG::2015-09-02
>>> 16:07:49,564::hsm::403::Storage.Misc.excCmd::(__validateLvmLockingType)
>>> /usr/bin/sudo -n /usr/sbin/lvm dumpconfig global/locking_type (cwd None)
>>> MainThread::DEBUG::2015-09-02
>>> 16:07:49,611::hsm::403::Storage.Misc.excCmd::(__validateLvmLockingType)
>>> SUCCESS: <err> = ''; <rc> = 0
>>> MainThread::DEBUG::2015-09-02
>>> 16:07:49,611::hsm::427::Storage.HSM::(__cleanStorageRepository) Started
>>> cleaning storage repository at '/rhev/data-center'
>>> MainThread::DEBUG::2015-09-02
>>> 16:07:49,614::hsm::459::Storage.HSM::(__cleanStorageRepository) White list:
>>> ['/rhev/data-center/hsm-tasks',
'/rhev/data-center/hsm-tasks/*',
>>> '/rhev/data-center/mnt']
>>> MainThread::DEBUG::2015-09-02
>>> 16:07:49,614::hsm::460::Storage.HSM::(__cleanStorageRepository) Mount list:
>>> []
>>> MainThread::DEBUG::2015-09-02
>>> 16:07:49,614::hsm::462::Storage.HSM::(__cleanStorageRepository) Cleaning
>>> leftovers
>>> MainThread::DEBUG::2015-09-02
>>> 16:07:49,615::hsm::505::Storage.HSM::(__cleanStorageRepository) Finished
>>> cleaning storage repository at '/rhev/data-center'
>>> storageRefresh::DEBUG::2015-09-02
>>> 16:07:49,616::misc::733::Storage.SamplingMethod::(__call__) Trying to enter
>>> sampling method (storage.sdc.refreshStorage)
>>> MainThread::INFO::2015-09-02
>>> 16:07:49,617::dispatcher::46::Storage.Dispatcher::(__init__) Starting
>>> StorageDispatcher...
>>> storageRefresh::DEBUG::2015-09-02
>>> 16:07:49,620::misc::736::Storage.SamplingMethod::(__call__) Got in to
>>> sampling method
>>> storageRefresh::DEBUG::2015-09-02
>>> 16:07:49,792::misc::733::Storage.SamplingMethod::(__call__) Trying to enter
>>> sampling method (storage.iscsi.rescan)
>>> storageRefresh::DEBUG::2015-09-02
>>> 16:07:49,793::misc::736::Storage.SamplingMethod::(__call__) Got in to
>>> sampling method
>>> storageRefresh::DEBUG::2015-09-02
>>> 16:07:49,793::iscsi::431::Storage.ISCSI::(rescan) Performing SCSI scan,
>>> this will take up to 30 seconds
>>> storageRefresh::DEBUG::2015-09-02
>>> 16:07:49,924::iscsiadm::97::Storage.Misc.excCmd::(_runCmd) /usr/bin/sudo -n
>>> /sbin/iscsiadm -m session -R (cwd None)
>>> MainThread::DEBUG::2015-09-02
>>> 16:07:49,924::task::595::Storage.TaskManager.Task::(_updateState)
>>> Task=`68d01d7d-b426-4465-829e-174e2cb47e9e`::moving from state init ->
>>> state preparing
>>> MainThread::INFO::2015-09-02
>>> 16:07:49,924::logUtils::48::dispatcher::(wrapper) Run and protect:
>>> registerDomainStateChangeCallback(callbackFunc=<functools.partial object
at
>>> 0x7fc2f03fa6d8>)
>>> MainThread::INFO::2015-09-02
>>> 16:07:49,924::logUtils::51::dispatcher::(wrapper) Run and protect:
>>> registerDomainStateChangeCallback, Return response: None
>>> MainThread::DEBUG::2015-09-02
>>> 16:07:49,927::task::1191::Storage.TaskManager.Task::(prepare)
>>> Task=`68d01d7d-b426-4465-829e-174e2cb47e9e`::finished: None
>>> MainThread::DEBUG::2015-09-02
>>> 16:07:49,927::task::595::Storage.TaskManager.Task::(_updateState)
>>> Task=`68d01d7d-b426-4465-829e-174e2cb47e9e`::moving from state preparing
->
>>> state finished
>>> MainThread::DEBUG::2015-09-02
>>>
16:07:49,927::resourceManager::940::Storage.ResourceManager.Owner::(releaseAll)
>>> Owner.releaseAll requests {} resources {}
>>> MainThread::DEBUG::2015-09-02
>>>
16:07:49,927::resourceManager::977::Storage.ResourceManager.Owner::(cancelAll)
>>> Owner.cancelAll requests {}
>>> MainThread::DEBUG::2015-09-02
>>> 16:07:49,928::task::993::Storage.TaskManager.Task::(_decref)
>>> Task=`68d01d7d-b426-4465-829e-174e2cb47e9e`::ref 0 aborting False
>>> MainThread::INFO::2015-09-02 16:07:49,928::momIF::46::MOM::(__init__)
>>> Preparing MOM interface
>>> MainThread::INFO::2015-09-02 16:07:49,929::momIF::55::MOM::(__init__)
>>> Using named unix socket /var/run/vdsm/mom-vdsm.sock
>>> MainThread::INFO::2015-09-02 16:07:49,929::secret::90::root::(clear)
>>> Unregistering all secrests
>>> MainThread::DEBUG::2015-09-02
>>> 16:07:49,929::libvirtconnection::160::root::(get) trying to connect libvirt
>>> MainThread::INFO::2015-09-02
>>> 16:07:49,933::vmchannels::196::vds::(settimeout) Setting channels'
timeout
>>> to 30 seconds.
>>> VM Channels Listener::DEBUG::2015-09-02
>>> 16:07:49,934::vmchannels::178::vds::(run) Starting VM channels listener
>>> thread.
>>> MainThread::INFO::2015-09-02
>>> 16:07:49,935::protocoldetector::172::vds.MultiProtocolAcceptor::(__init__)
>>> Listening at 0.0.0.0:54321
>>> MainThread::DEBUG::2015-09-02
>>>
16:07:50,063::protocoldetector::199::vds.MultiProtocolAcceptor::(add_detector)
>>> Adding detector <rpc.bindingxmlrpc.XmlDetector instance at
0x7fc2f00dc440>
>>> storageRefresh::DEBUG::2015-09-02
>>> 16:07:50,063::misc::743::Storage.SamplingMethod::(__call__) Returning last
>>> result
>>> storageRefresh::DEBUG::2015-09-02
>>> 16:07:50,080::misc::733::Storage.SamplingMethod::(__call__) Trying to enter
>>> sampling method (storage.hba.rescan)
>>> storageRefresh::DEBUG::2015-09-02
>>> 16:07:50,081::misc::736::Storage.SamplingMethod::(__call__) Got in to
>>> sampling method
>>> storageRefresh::DEBUG::2015-09-02
>>> 16:07:50,081::hba::56::Storage.HBA::(rescan) Starting scan
>>> storageRefresh::DEBUG::2015-09-02
>>> 16:07:50,081::supervdsm::76::SuperVdsmProxy::(_connect) Trying to connect
>>> to Super Vdsm
>>> MainThread::DEBUG::2015-09-02
>>>
16:07:50,157::protocoldetector::199::vds.MultiProtocolAcceptor::(add_detector)
>>> Adding detector <yajsonrpc.stompreactor.StompDetector instance at
>>> 0x7fc2e01332d8>
>>> BindingXMLRPC::INFO::2015-09-02
>>> 16:07:50,158::bindingxmlrpc::62::vds::(threaded_start) XMLRPC server running
>>> MainThread::DEBUG::2015-09-02
>>> 16:07:50,158::schedule::98::Scheduler::(start) Starting scheduler
>>> periodic-sched
>>> periodic-sched::DEBUG::2015-09-02
>>> 16:07:50,159::schedule::142::Scheduler::(_run) started
>>> MainThread::DEBUG::2015-09-02
>>> 16:07:50,159::executor::69::Executor::(start) Starting executor
>>> MainThread::DEBUG::2015-09-02
>>> 16:07:50,159::executor::157::Executor::(__init__) Starting worker periodic/0
>>> periodic/0::DEBUG::2015-09-02
>>> 16:07:50,159::executor::171::Executor::(_run) Worker started
>>> MainThread::DEBUG::2015-09-02
>>> 16:07:50,159::executor::157::Executor::(__init__) Starting worker periodic/1
>>> periodic/1::DEBUG::2015-09-02
>>> 16:07:50,160::executor::171::Executor::(_run) Worker started
>>> MainThread::DEBUG::2015-09-02
>>> 16:07:50,160::executor::157::Executor::(__init__) Starting worker periodic/2
>>> periodic/2::DEBUG::2015-09-02
>>> 16:07:50,160::executor::171::Executor::(_run) Worker started
>>> MainThread::DEBUG::2015-09-02
>>> 16:07:50,160::executor::157::Executor::(__init__) Starting worker periodic/3
>>> periodic/3::DEBUG::2015-09-02
>>> 16:07:50,160::executor::171::Executor::(_run) Worker started
>>> MainThread::DEBUG::2015-09-02
>>> 16:07:50,160::libvirtconnection::160::root::(get) trying to connect libvirt
>>> MainThread::DEBUG::2015-09-02
>>> 16:07:50,163::periodic::157::virt.periodic.Operation::(start) starting
>>> operation VmDispatcher(<class 'virt.periodic.UpdateVolumes'>)
>>> MainThread::DEBUG::2015-09-02
>>> 16:07:50,164::periodic::157::virt.periodic.Operation::(start) starting
>>> operation VmDispatcher(<class
'virt.periodic.NumaInfoMonitor'>)
>>> MainThread::DEBUG::2015-09-02
>>> 16:07:50,164::periodic::157::virt.periodic.Operation::(start) starting
>>> operation VmDispatcher(<class
'virt.periodic.BlockjobMonitor'>)
>>> MainThread::DEBUG::2015-09-02
>>> 16:07:50,164::periodic::157::virt.periodic.Operation::(start) starting
>>> operation <virt.sampling.VMBulkSampler object at 0x7fc2e0151d10>
>>> MainThread::DEBUG::2015-09-02
>>> 16:07:50,164::periodic::157::virt.periodic.Operation::(start) starting
>>> operation VmDispatcher(<class
'virt.periodic.DriveWatermarkMonitor'>)
>>> storageRefresh::DEBUG::2015-09-02
>>> 16:07:50,167::hba::62::Storage.HBA::(rescan) Scan finished
>>> storageRefresh::DEBUG::2015-09-02
>>> 16:07:50,167::misc::743::Storage.SamplingMethod::(__call__) Returning last
>>> result
>>> storageRefresh::DEBUG::2015-09-02
>>> 16:07:50,167::multipath::77::Storage.Misc.excCmd::(rescan) /usr/bin/sudo -n
>>> /usr/sbin/multipath (cwd None)
>>> storageRefresh::DEBUG::2015-09-02
>>> 16:07:50,513::multipath::77::Storage.Misc.excCmd::(rescan) SUCCESS:
<err> =
>>> ''; <rc> = 0
>>> storageRefresh::DEBUG::2015-09-02
>>> 16:07:50,513::utils::661::root::(execCmd) /sbin/udevadm settle --timeout=5
>>> (cwd None)
>>> Reactor thread::INFO::2015-09-02
>>>
16:07:50,590::protocoldetector::72::ProtocolDetector.AcceptorImpl::(handle_accept)
>>> Accepting connection from 127.0.0.1:56311
>>> Reactor thread::DEBUG::2015-09-02
>>> 16:07:50,596::protocoldetector::82::ProtocolDetector.Detector::(__init__)
>>> Using required_size=11
>>> Reactor thread::INFO::2015-09-02
>>>
16:07:50,596::protocoldetector::118::ProtocolDetector.Detector::(handle_read)
>>> Detected protocol xml from 127.0.0.1:56311
>>> Reactor thread::DEBUG::2015-09-02
>>> 16:07:50,596::bindingxmlrpc::1296::XmlDetector::(handle_socket) xml over
>>> http detected from ('127.0.0.1', 56311)
>>> BindingXMLRPC::INFO::2015-09-02
>>> 16:07:50,596::xmlrpc::73::vds.XMLRPCServer::(handle_request) Starting
>>> request handler for 127.0.0.1:56311
>>> Thread-13::INFO::2015-09-02
>>> 16:07:50,597::xmlrpc::84::vds.XMLRPCServer::(_process_requests) Request
>>> handler for 127.0.0.1:56311 started
>>> Thread-13::DEBUG::2015-09-02
>>> 16:07:50,597::bindingxmlrpc::1256::vds::(wrapper) client [127.0.0.1]::call
>>> getHardwareInfo with () {}
>>> Thread-13::DEBUG::2015-09-02
>>> 16:07:50,597::bindingxmlrpc::1263::vds::(wrapper) return getHardwareInfo
>>> with {'status': {'message': 'Recovering from crash or
Initializing',
>>> 'code': 99}}
>>> Thread-13::INFO::2015-09-02
>>> 16:07:50,599::xmlrpc::92::vds.XMLRPCServer::(_process_requests) Request
>>> handler for 127.0.0.1:56311 stopped
>>> Reactor thread::INFO::2015-09-02
>>>
16:07:51,607::protocoldetector::72::ProtocolDetector.AcceptorImpl::(handle_accept)
>>> Accepting connection from 127.0.0.1:56312
>>> Reactor thread::DEBUG::2015-09-02
>>> 16:07:51,613::protocoldetector::82::ProtocolDetector.Detector::(__init__)
>>> Using required_size=11
>>> Reactor thread::INFO::2015-09-02
>>>
16:07:51,613::protocoldetector::118::ProtocolDetector.Detector::(handle_read)
>>> Detected protocol xml from 127.0.0.1:56312
>>> Reactor thread::DEBUG::2015-09-02
>>> 16:07:51,613::bindingxmlrpc::1296::XmlDetector::(handle_socket) xml over
>>> http detected from ('127.0.0.1', 56312)
>>> BindingXMLRPC::INFO::2015-09-02
>>> 16:07:51,613::xmlrpc::73::vds.XMLRPCServer::(handle_request) Starting
>>> request handler for 127.0.0.1:56312
>>> Thread-14::INFO::2015-09-02
>>> 16:07:51,613::xmlrpc::84::vds.XMLRPCServer::(_process_requests) Request
>>> handler for 127.0.0.1:56312 started
>>> Thread-14::DEBUG::2015-09-02
>>> 16:07:51,614::bindingxmlrpc::1256::vds::(wrapper) client [127.0.0.1]::call
>>> getHardwareInfo with () {}
>>> Thread-14::DEBUG::2015-09-02
>>> 16:07:51,614::bindingxmlrpc::1263::vds::(wrapper) return getHardwareInfo
>>> with {'status': {'message': 'Recovering from crash or
Initializing',
>>> 'code': 99}}
>>> Thread-14::INFO::2015-09-02
>>> 16:07:51,615::xmlrpc::92::vds.XMLRPCServer::(_process_requests) Request
>>> handler for 127.0.0.1:56312 stopped
>>> storageRefresh::DEBUG::2015-09-02
>>> 16:07:51,924::utils::679::root::(execCmd) SUCCESS: <err> = '';
<rc> = 0
>>> storageRefresh::DEBUG::2015-09-02
>>> 16:07:51,926::lvm::498::Storage.OperationMutex::(_invalidateAllPvs)
>>> Operation 'lvm invalidate operation' got the operation mutex
>>> storageRefresh::DEBUG::2015-09-02
>>> 16:07:51,926::lvm::500::Storage.OperationMutex::(_invalidateAllPvs)
>>> Operation 'lvm invalidate operation' released the operation mutex
>>> storageRefresh::DEBUG::2015-09-02
>>> 16:07:51,926::lvm::509::Storage.OperationMutex::(_invalidateAllVgs)
>>> Operation 'lvm invalidate operation' got the operation mutex
>>> storageRefresh::DEBUG::2015-09-02
>>> 16:07:51,926::lvm::511::Storage.OperationMutex::(_invalidateAllVgs)
>>> Operation 'lvm invalidate operation' released the operation mutex
>>> storageRefresh::DEBUG::2015-09-02
>>> 16:07:51,926::lvm::529::Storage.OperationMutex::(_invalidateAllLvs)
>>> Operation 'lvm invalidate operation' got the operation mutex
>>> storageRefresh::DEBUG::2015-09-02
>>> 16:07:51,926::lvm::531::Storage.OperationMutex::(_invalidateAllLvs)
>>> Operation 'lvm invalidate operation' released the operation mutex
>>> storageRefresh::DEBUG::2015-09-02
>>> 16:07:51,926::misc::743::Storage.SamplingMethod::(__call__) Returning last
>>> result
>>> storageRefresh::DEBUG::2015-09-02
>>> 16:07:51,926::lvm::320::Storage.OperationMutex::(_reloadpvs) Operation
'lvm
>>> reload operation' got the operation mutex
>>> storageRefresh::DEBUG::2015-09-02
>>> 16:07:51,927::lvm::291::Storage.Misc.excCmd::(cmd) /usr/bin/sudo -n
>>> /usr/sbin/lvm pvs --config ' devices { preferred_names =
["^/dev/mapper/"]
>>> ignore_suspended_devices=1 write_cache_state=0 disable_after_error_count=3
>>> obtain_device_list_from_udev=0 filter = [
>>>
'\''a|/dev/mapper/Hitachi_HDS721010DLE630_MSK523Y209VK0B|'\'',
>>> '\''r|.*|'\'' ] } global { locking_type=1
prioritise_write_locks=1
>>> wait_for_locks=1 use_lvmetad=0 } backup { retain_min = 50 retain_days =
>>> 0 } ' --noheadings --units b --nosuffix --separator '|'
>>> --ignoreskippedcluster -o
>>>
uuid,name,size,vg_name,vg_uuid,pe_start,pe_count,pe_alloc_count,mda_count,dev_size
>>> (cwd None)
>>> storageRefresh::DEBUG::2015-09-02
>>> 16:07:52,341::lvm::291::Storage.Misc.excCmd::(cmd) SUCCESS: <err> =
'
>>> WARNING: lvmetad is running but disabled. Restart lvmetad before enabling
>>> it!\n'; <rc> = 0
>>> storageRefresh::DEBUG::2015-09-02
>>> 16:07:52,341::lvm::348::Storage.OperationMutex::(_reloadpvs) Operation
'lvm
>>> reload operation' released the operation mutex
>>> storageRefresh::DEBUG::2015-09-02
>>> 16:07:52,341::lvm::371::Storage.OperationMutex::(_reloadvgs) Operation
'lvm
>>> reload operation' got the operation mutex
>>> storageRefresh::DEBUG::2015-09-02
>>> 16:07:52,341::lvm::291::Storage.Misc.excCmd::(cmd) /usr/bin/sudo -n
>>> /usr/sbin/lvm vgs --config ' devices { preferred_names =
["^/dev/mapper/"]
>>> ignore_suspended_devices=1 write_cache_state=0 disable_after_error_count=3
>>> obtain_device_list_from_udev=0 filter = [
>>>
'\''a|/dev/mapper/Hitachi_HDS721010DLE630_MSK523Y209VK0B|'\'',
>>> '\''r|.*|'\'' ] } global { locking_type=1
prioritise_write_locks=1
>>> wait_for_locks=1 use_lvmetad=0 } backup { retain_min = 50 retain_days =
>>> 0 } ' --noheadings --units b --nosuffix --separator '|'
>>> --ignoreskippedcluster -o
>>>
uuid,name,attr,size,free,extent_size,extent_count,free_count,tags,vg_mda_size,vg_mda_free,lv_count,pv_count,pv_name
>>> (cwd None)
>>> storageRefresh::DEBUG::2015-09-02
>>> 16:07:52,405::lvm::291::Storage.Misc.excCmd::(cmd) SUCCESS: <err> =
'
>>> WARNING: lvmetad is running but disabled. Restart lvmetad before enabling
>>> it!\n'; <rc> = 0
>>> storageRefresh::DEBUG::2015-09-02
>>> 16:07:52,405::lvm::416::Storage.OperationMutex::(_reloadvgs) Operation
'lvm
>>> reload operation' released the operation mutex
>>> storageRefresh::DEBUG::2015-09-02
>>> 16:07:52,406::lvm::291::Storage.Misc.excCmd::(cmd) /usr/bin/sudo -n
>>> /usr/sbin/lvm lvs --config ' devices { preferred_names =
["^/dev/mapper/"]
>>> ignore_suspended_devices=1 write_cache_state=0 disable_after_error_count=3
>>> obtain_device_list_from_udev=0 filter = [
>>>
'\''a|/dev/mapper/Hitachi_HDS721010DLE630_MSK523Y209VK0B|'\'',
>>> '\''r|.*|'\'' ] } global { locking_type=1
prioritise_write_locks=1
>>> wait_for_locks=1 use_lvmetad=0 } backup { retain_min = 50 retain_days =
>>> 0 } ' --noheadings --units b --nosuffix --separator '|'
>>> --ignoreskippedcluster -o
>>> uuid,name,vg_name,attr,size,seg_start_pe,devices,tags (cwd None)
>>> storageRefresh::DEBUG::2015-09-02
>>> 16:07:52,458::lvm::291::Storage.Misc.excCmd::(cmd) SUCCESS: <err> =
'
>>> WARNING: lvmetad is running but disabled. Restart lvmetad before enabling
>>> it!\n'; <rc> = 0
>>> storageRefresh::DEBUG::2015-09-02
>>> 16:07:52,458::lvm::371::Storage.OperationMutex::(_reloadvgs) Operation
'lvm
>>> reload operation' got the operation mutex
>>> storageRefresh::DEBUG::2015-09-02
>>> 16:07:52,459::lvm::291::Storage.Misc.excCmd::(cmd) /usr/bin/sudo -n
>>> /usr/sbin/lvm vgs --config ' devices { preferred_names =
["^/dev/mapper/"]
>>> ignore_suspended_devices=1 write_cache_state=0 disable_after_error_count=3
>>> obtain_device_list_from_udev=0 filter = [
>>>
'\''a|/dev/mapper/Hitachi_HDS721010DLE630_MSK523Y209VK0B|'\'',
>>> '\''r|.*|'\'' ] } global { locking_type=1
prioritise_write_locks=1
>>> wait_for_locks=1 use_lvmetad=0 } backup { retain_min = 50 retain_days =
>>> 0 } ' --noheadings --units b --nosuffix --separator '|'
>>> --ignoreskippedcluster -o
>>>
uuid,name,attr,size,free,extent_size,extent_count,free_count,tags,vg_mda_size,vg_mda_free,lv_count,pv_count,pv_name
>>> (cwd None)
>>> storageRefresh::DEBUG::2015-09-02
>>> 16:07:52,491::lvm::291::Storage.Misc.excCmd::(cmd) SUCCESS: <err> =
'
>>> WARNING: lvmetad is running but disabled. Restart lvmetad before enabling
>>> it!\n'; <rc> = 0
>>> storageRefresh::DEBUG::2015-09-02
>>> 16:07:52,491::lvm::416::Storage.OperationMutex::(_reloadvgs) Operation
'lvm
>>> reload operation' released the operation mutex
>>> storageRefresh::DEBUG::2015-09-02
>>> 16:07:52,491::hsm::373::Storage.HSM::(storageRefresh) HSM is ready
>>> Reactor thread::INFO::2015-09-02
>>>
16:07:52,624::protocoldetector::72::ProtocolDetector.AcceptorImpl::(handle_accept)
>>> Accepting connection from 127.0.0.1:56313
>>> Reactor thread::DEBUG::2015-09-02
>>> 16:07:52,629::protocoldetector::82::ProtocolDetector.Detector::(__init__)
>>> Using required_size=11
>>> Reactor thread::INFO::2015-09-02
>>>
16:07:52,629::protocoldetector::118::ProtocolDetector.Detector::(handle_read)
>>> Detected protocol xml from 127.0.0.1:56313
>>> Reactor thread::DEBUG::2015-09-02
>>> 16:07:52,629::bindingxmlrpc::1296::XmlDetector::(handle_socket) xml over
>>> http detected from ('127.0.0.1', 56313)
>>> BindingXMLRPC::INFO::2015-09-02
>>> 16:07:52,629::xmlrpc::73::vds.XMLRPCServer::(handle_request) Starting
>>> request handler for 127.0.0.1:56313
>>> Thread-15::INFO::2015-09-02
>>> 16:07:52,630::xmlrpc::84::vds.XMLRPCServer::(_process_requests) Request
>>> handler for 127.0.0.1:56313 started
>>> Thread-15::DEBUG::2015-09-02
>>> 16:07:52,630::bindingxmlrpc::1256::vds::(wrapper) client [127.0.0.1]::call
>>> getHardwareInfo with () {}
>>> Thread-15::DEBUG::2015-09-02
>>> 16:07:52,719::bindingxmlrpc::1263::vds::(wrapper) return getHardwareInfo
>>> with {'status': {'message': 'Done', 'code':
0}, 'info':
>>> {'systemProductName': 'System Product Name',
'systemSerialNumber': 'System
>>> Serial Number', 'systemFamily': 'To be filled by O.E.M.',
'systemVersion':
>>> 'System Version', 'systemUUID':
'267A6B80-D7DA-11DD-81CF-C860009B3CD9',
>>> 'systemManufacturer': 'System manufacturer'}}
>>> Thread-15::INFO::2015-09-02
>>> 16:07:52,721::xmlrpc::92::vds.XMLRPCServer::(_process_requests) Request
>>> handler for 127.0.0.1:56313 stopped
>>> Reactor thread::INFO::2015-09-02
>>>
16:07:52,730::protocoldetector::72::ProtocolDetector.AcceptorImpl::(handle_accept)
>>> Accepting connection from 127.0.0.1:56314
>>> Reactor thread::DEBUG::2015-09-02
>>> 16:07:52,735::protocoldetector::82::ProtocolDetector.Detector::(__init__)
>>> Using required_size=11
>>> Reactor thread::INFO::2015-09-02
>>>
16:07:52,735::protocoldetector::118::ProtocolDetector.Detector::(handle_read)
>>> Detected protocol xml from 127.0.0.1:56314
>>> Reactor thread::DEBUG::2015-09-02
>>> 16:07:52,735::bindingxmlrpc::1296::XmlDetector::(handle_socket) xml over
>>> http detected from ('127.0.0.1', 56314)
>>> BindingXMLRPC::INFO::2015-09-02
>>> 16:07:52,735::xmlrpc::73::vds.XMLRPCServer::(handle_request) Starting
>>> request handler for 127.0.0.1:56314
>>> Thread-16::INFO::2015-09-02
>>> 16:07:52,735::xmlrpc::84::vds.XMLRPCServer::(_process_requests) Request
>>> handler for 127.0.0.1:56314 started
>>> Thread-16::DEBUG::2015-09-02
>>> 16:07:52,736::bindingxmlrpc::325::vds::(wrapper) client [127.0.0.1]
>>> Thread-16::DEBUG::2015-09-02
>>> 16:07:52,736::task::595::Storage.TaskManager.Task::(_updateState)
>>> Task=`c4a18001-912b-47dc-9713-7d50e5133b59`::moving from state init ->
>>> state preparing
>>> Thread-16::INFO::2015-09-02
>>> 16:07:52,736::logUtils::48::dispatcher::(wrapper) Run and protect:
>>> connectStorageServer(domType=3,
>>> spUUID='00000000-0000-0000-0000-000000000000',
conList=[{'id':
>>> '57bc98c0-560f-4e61-9d86-df92ad468d3b', 'connection':
'192.168.1.50',
>>> 'iqn': 'iqn.2015-08.openSuse.wodel:target00',
'portal': '1', 'user':
>>> 'iscsiuser', 'password': '********', 'port':
'3260'}], options=None)
>>> Thread-16::DEBUG::2015-09-02
>>> 16:07:52,737::iscsiadm::97::Storage.Misc.excCmd::(_runCmd) /usr/bin/sudo -n
>>> /sbin/iscsiadm -m node -T iqn.2015-08.openSuse.wodel:target00 -I default -p
>>> 192.168.1.50:3260,1 --op=new (cwd None)
>>> Thread-16::DEBUG::2015-09-02
>>> 16:07:52,789::iscsiadm::97::Storage.Misc.excCmd::(_runCmd) SUCCESS:
<err> =
>>> ''; <rc> = 0
>>> Thread-16::DEBUG::2015-09-02
>>> 16:07:52,789::iscsiadm::97::Storage.Misc.excCmd::(_runCmd) /sbin/iscsiadm
>>> -m node -T iqn.2015-08.openSuse.wodel:target00 -I default -p
>>> 192.168.1.50:3260,1 -n node.session.auth.authmethod -v '****'
>>> --op=update (cwd None)
>>> Thread-16::DEBUG::2015-09-02
>>> 16:07:52,811::iscsiadm::97::Storage.Misc.excCmd::(_runCmd) SUCCESS:
<err> =
>>> ''; <rc> = 0
>>> Thread-16::DEBUG::2015-09-02
>>> 16:07:52,812::iscsiadm::97::Storage.Misc.excCmd::(_runCmd) /sbin/iscsiadm
>>> -m node -T iqn.2015-08.openSuse.wodel:target00 -I default -p
>>> 192.168.1.50:3260,1 -n node.session.auth.username -v '****'
--op=update
>>> (cwd None)
>>> Thread-16::DEBUG::2015-09-02
>>> 16:07:52,846::iscsiadm::97::Storage.Misc.excCmd::(_runCmd) SUCCESS:
<err> =
>>> ''; <rc> = 0
>>> Thread-16::DEBUG::2015-09-02
>>> 16:07:52,847::iscsiadm::97::Storage.Misc.excCmd::(_runCmd) /sbin/iscsiadm
>>> -m node -T iqn.2015-08.openSuse.wodel:target00 -I default -p
>>> 192.168.1.50:3260,1 -n node.session.auth.password -v '****'
--op=update
>>> (cwd None)
>>> Thread-16::DEBUG::2015-09-02
>>> 16:07:52,868::iscsiadm::97::Storage.Misc.excCmd::(_runCmd) SUCCESS:
<err> =
>>> ''; <rc> = 0
>>> Thread-16::DEBUG::2015-09-02
>>> 16:07:52,868::iscsiadm::97::Storage.Misc.excCmd::(_runCmd) /usr/bin/sudo -n
>>> /sbin/iscsiadm -m iface -I default (cwd None)
>>> Thread-16::DEBUG::2015-09-02
>>> 16:07:52,905::iscsiadm::97::Storage.Misc.excCmd::(_runCmd) SUCCESS:
<err> =
>>> ''; <rc> = 0
>>> Thread-16::INFO::2015-09-02
>>> 16:07:52,905::iscsi::564::Storage.ISCSI::(setRpFilterIfNeeded) iSCSI
>>> iface.net_ifacename not provided. Skipping.
>>> Thread-16::DEBUG::2015-09-02
>>> 16:07:52,906::iscsiadm::97::Storage.Misc.excCmd::(_runCmd) /usr/bin/sudo -n
>>> /sbin/iscsiadm -m node -T iqn.2015-08.openSuse.wodel:target00 -I default -p
>>> 192.168.1.50:3260,1 -l (cwd None)
>>> Thread-16::DEBUG::2015-09-02
>>> 16:07:53,027::iscsiadm::97::Storage.Misc.excCmd::(_runCmd) SUCCESS:
<err> =
>>> ''; <rc> = 0
>>> Thread-16::DEBUG::2015-09-02
>>> 16:07:53,028::iscsiadm::97::Storage.Misc.excCmd::(_runCmd) /usr/bin/sudo -n
>>> /sbin/iscsiadm -m node -T iqn.2015-08.openSuse.wodel:target00 -I default -p
>>> 192.168.1.50:3260,1 -n node.startup -v manual --op=update (cwd None)
>>> Thread-16::DEBUG::2015-09-02
>>> 16:07:53,088::iscsiadm::97::Storage.Misc.excCmd::(_runCmd) SUCCESS:
<err> =
>>> ''; <rc> = 0
>>> Thread-16::DEBUG::2015-09-02 16:07:53,088::utils::661::root::(execCmd)
>>> /sbin/udevadm settle --timeout=5 (cwd None)
>>> Thread-16::DEBUG::2015-09-02 16:07:53,182::utils::679::root::(execCmd)
>>> SUCCESS: <err> = ''; <rc> = 0
>>> Thread-16::DEBUG::2015-09-02
>>> 16:07:53,182::misc::733::Storage.SamplingMethod::(__call__) Trying to enter
>>> sampling method (storage.sdc.refreshStorage)
>>> Thread-16::DEBUG::2015-09-02
>>> 16:07:53,182::misc::736::Storage.SamplingMethod::(__call__) Got in to
>>> sampling method
>>> Thread-16::DEBUG::2015-09-02
>>> 16:07:53,182::misc::733::Storage.SamplingMethod::(__call__) Trying to enter
>>> sampling method (storage.iscsi.rescan)
>>> Thread-16::DEBUG::2015-09-02
>>> 16:07:53,182::misc::736::Storage.SamplingMethod::(__call__) Got in to
>>> sampling method
>>> Thread-16::DEBUG::2015-09-02
>>> 16:07:53,182::iscsi::431::Storage.ISCSI::(rescan) Performing SCSI scan,
>>> this will take up to 30 seconds
>>> Thread-16::DEBUG::2015-09-02
>>> 16:07:53,182::iscsiadm::97::Storage.Misc.excCmd::(_runCmd) /usr/bin/sudo -n
>>> /sbin/iscsiadm -m session -R (cwd None)
>>> Thread-16::DEBUG::2015-09-02
>>> 16:07:53,229::misc::743::Storage.SamplingMethod::(__call__) Returning last
>>> result
>>> Thread-16::DEBUG::2015-09-02
>>> 16:07:53,229::misc::733::Storage.SamplingMethod::(__call__) Trying to enter
>>> sampling method (storage.hba.rescan)
>>> Thread-16::DEBUG::2015-09-02
>>> 16:07:53,229::misc::736::Storage.SamplingMethod::(__call__) Got in to
>>> sampling method
>>> Thread-16::DEBUG::2015-09-02
>>> 16:07:53,229::hba::56::Storage.HBA::(rescan) Starting scan
>>> Thread-16::DEBUG::2015-09-02
>>> 16:07:53,300::hba::62::Storage.HBA::(rescan) Scan finished
>>> Thread-16::DEBUG::2015-09-02
>>> 16:07:53,300::misc::743::Storage.SamplingMethod::(__call__) Returning last
>>> result
>>> Thread-16::DEBUG::2015-09-02
>>> 16:07:53,300::multipath::77::Storage.Misc.excCmd::(rescan) /usr/bin/sudo -n
>>> /usr/sbin/multipath (cwd None)
>>> Thread-16::DEBUG::2015-09-02
>>> 16:07:53,435::multipath::77::Storage.Misc.excCmd::(rescan) SUCCESS:
<err> =
>>> ''; <rc> = 0
>>> Thread-16::DEBUG::2015-09-02 16:07:53,435::utils::661::root::(execCmd)
>>> /sbin/udevadm settle --timeout=5 (cwd None)
>>> Thread-16::DEBUG::2015-09-02 16:07:53,919::utils::679::root::(execCmd)
>>> SUCCESS: <err> = ''; <rc> = 0
>>> Thread-16::DEBUG::2015-09-02
>>> 16:07:53,921::lvm::498::Storage.OperationMutex::(_invalidateAllPvs)
>>> Operation 'lvm invalidate operation' got the operation mutex
>>> Thread-16::DEBUG::2015-09-02
>>> 16:07:53,921::lvm::500::Storage.OperationMutex::(_invalidateAllPvs)
>>> Operation 'lvm invalidate operation' released the operation mutex
>>> Thread-16::DEBUG::2015-09-02
>>> 16:07:53,922::lvm::509::Storage.OperationMutex::(_invalidateAllVgs)
>>> Operation 'lvm invalidate operation' got the operation mutex
>>> Thread-16::DEBUG::2015-09-02
>>> 16:07:53,922::lvm::511::Storage.OperationMutex::(_invalidateAllVgs)
>>> Operation 'lvm invalidate operation' released the operation mutex
>>> Thread-16::DEBUG::2015-09-02
>>> 16:07:53,922::lvm::529::Storage.OperationMutex::(_invalidateAllLvs)
>>> Operation 'lvm invalidate operation' got the operation mutex
>>> Thread-16::DEBUG::2015-09-02
>>> 16:07:53,922::lvm::531::Storage.OperationMutex::(_invalidateAllLvs)
>>> Operation 'lvm invalidate operation' released the operation mutex
>>> Thread-16::DEBUG::2015-09-02
>>> 16:07:53,922::misc::743::Storage.SamplingMethod::(__call__) Returning last
>>> result
>>> Thread-16::DEBUG::2015-09-02
>>> 16:07:53,922::lvm::371::Storage.OperationMutex::(_reloadvgs) Operation
'lvm
>>> reload operation' got the operation mutex
>>> Thread-16::DEBUG::2015-09-02
>>> 16:07:53,923::lvm::291::Storage.Misc.excCmd::(cmd) /usr/bin/sudo -n
>>> /usr/sbin/lvm vgs --config ' devices { preferred_names =
["^/dev/mapper/"]
>>> ignore_suspended_devices=1 write_cache_state=0 disable_after_error_count=3
>>> obtain_device_list_from_udev=0 filter = [
>>>
'\''a|/dev/mapper/33000000100000001|/dev/mapper/Hitachi_HDS721010DLE630_MSK523Y209VK0B|'\'',
>>> '\''r|.*|'\'' ] } global { locking_type=1
prioritise_write_locks=1
>>> wait_for_locks=1 use_lvmetad=0 } backup { retain_min = 50 retain_days =
>>> 0 } ' --noheadings --units b --nosuffix --separator '|'
>>> --ignoreskippedcluster -o
>>>
uuid,name,attr,size,free,extent_size,extent_count,free_count,tags,vg_mda_size,vg_mda_free,lv_count,pv_count,pv_name
>>> (cwd None)
>>> Thread-16::DEBUG::2015-09-02
>>> 16:07:54,058::lvm::291::Storage.Misc.excCmd::(cmd) SUCCESS: <err> =
'
>>> WARNING: lvmetad is running but disabled. Restart lvmetad before enabling
>>> it!\n'; <rc> = 0
>>> Thread-16::DEBUG::2015-09-02
>>> 16:07:54,059::lvm::416::Storage.OperationMutex::(_reloadvgs) Operation
'lvm
>>> reload operation' released the operation mutex
>>> Thread-16::DEBUG::2015-09-02
>>> 16:07:54,059::hsm::2418::Storage.HSM::(__prefetchDomains) Found SD uuids:
>>> ('8b25f3be-7574-4f7a-8851-363129704e52',)
>>> Thread-16::DEBUG::2015-09-02
>>> 16:07:54,059::hsm::2478::Storage.HSM::(connectStorageServer) knownSDs:
>>> {8b25f3be-7574-4f7a-8851-363129704e52: storage.blockSD.findDomain}
>>> Thread-16::INFO::2015-09-02
>>> 16:07:54,059::logUtils::51::dispatcher::(wrapper) Run and protect:
>>> connectStorageServer, Return response: {'statuslist':
[{'status': 0, 'id':
>>> '57bc98c0-560f-4e61-9d86-df92ad468d3b'}]}
>>> Thread-16::DEBUG::2015-09-02
>>> 16:07:54,059::task::1191::Storage.TaskManager.Task::(prepare)
>>> Task=`c4a18001-912b-47dc-9713-7d50e5133b59`::finished:
{'statuslist':
>>> [{'status': 0, 'id':
'57bc98c0-560f-4e61-9d86-df92ad468d3b'}]}
>>> Thread-16::DEBUG::2015-09-02
>>> 16:07:54,059::task::595::Storage.TaskManager.Task::(_updateState)
>>> Task=`c4a18001-912b-47dc-9713-7d50e5133b59`::moving from state preparing
->
>>> state finished
>>> Thread-16::DEBUG::2015-09-02
>>>
16:07:54,060::resourceManager::940::Storage.ResourceManager.Owner::(releaseAll)
>>> Owner.releaseAll requests {} resources {}
>>> Thread-16::DEBUG::2015-09-02
>>>
16:07:54,060::resourceManager::977::Storage.ResourceManager.Owner::(cancelAll)
>>> Owner.cancelAll requests {}
>>> Thread-16::DEBUG::2015-09-02
>>> 16:07:54,060::task::993::Storage.TaskManager.Task::(_decref)
>>> Task=`c4a18001-912b-47dc-9713-7d50e5133b59`::ref 0 aborting False
>>> Thread-16::INFO::2015-09-02
>>> 16:07:54,062::xmlrpc::92::vds.XMLRPCServer::(_process_requests) Request
>>> handler for 127.0.0.1:56314 stopped
>>> Reactor thread::INFO::2015-09-02
>>>
16:07:54,070::protocoldetector::72::ProtocolDetector.AcceptorImpl::(handle_accept)
>>> Accepting connection from 127.0.0.1:56316
>>> Reactor thread::DEBUG::2015-09-02
>>> 16:07:54,075::protocoldetector::82::ProtocolDetector.Detector::(__init__)
>>> Using required_size=11
>>> Reactor thread::INFO::2015-09-02
>>>
16:07:54,076::protocoldetector::118::ProtocolDetector.Detector::(handle_read)
>>> Detected protocol xml from 127.0.0.1:56316
>>> Reactor thread::DEBUG::2015-09-02
>>> 16:07:54,076::bindingxmlrpc::1296::XmlDetector::(handle_socket) xml over
>>> http detected from ('127.0.0.1', 56316)
>>> BindingXMLRPC::INFO::2015-09-02
>>> 16:07:54,076::xmlrpc::73::vds.XMLRPCServer::(handle_request) Starting
>>> request handler for 127.0.0.1:56316
>>> Thread-17::INFO::2015-09-02
>>> 16:07:54,076::xmlrpc::84::vds.XMLRPCServer::(_process_requests) Request
>>> handler for 127.0.0.1:56316 started
>>> Thread-17::DEBUG::2015-09-02
>>> 16:07:54,077::bindingxmlrpc::325::vds::(wrapper) client [127.0.0.1]
>>> Thread-17::DEBUG::2015-09-02
>>> 16:07:54,077::task::595::Storage.TaskManager.Task::(_updateState)
>>> Task=`7936300e-8a1a-47f5-83c4-16ed19853e36`::moving from state init ->
>>> state preparing
>>> Thread-17::INFO::2015-09-02
>>> 16:07:54,077::logUtils::48::dispatcher::(wrapper) Run and protect:
>>> prepareImage(sdUUID='8b25f3be-7574-4f7a-8851-363129704e52',
>>> spUUID='00000000-0000-0000-0000-000000000000',
>>> imgUUID='bf3bdae1-7318-4443-a19b-7371de30b982',
>>> leafUUID='cbb10cf0-9600-465e-aed9-412f7157706b')
>>> Thread-17::DEBUG::2015-09-02
>>>
16:07:54,077::resourceManager::198::Storage.ResourceManager.Request::(__init__)
>>>
ResName=`Storage.8b25f3be-7574-4f7a-8851-363129704e52`ReqID=`fc59b8b4-51c5-4a15-9716-aedbb6de62e6`::Request
>>> was made in '/usr/share/vdsm/storage/hsm.py' line '3194' at
'prepareImage'
>>> Thread-17::DEBUG::2015-09-02
>>>
16:07:54,078::resourceManager::542::Storage.ResourceManager::(registerResource)
>>> Trying to register resource
'Storage.8b25f3be-7574-4f7a-8851-363129704e52'
>>> for lock type 'shared'
>>> Thread-17::DEBUG::2015-09-02
>>>
16:07:54,078::resourceManager::601::Storage.ResourceManager::(registerResource)
>>> Resource 'Storage.8b25f3be-7574-4f7a-8851-363129704e52' is free. Now
>>> locking as 'shared' (1 active user)
>>> Thread-17::DEBUG::2015-09-02
>>> 16:07:54,078::resourceManager::238::Storage.ResourceManager.Request::(grant)
>>>
ResName=`Storage.8b25f3be-7574-4f7a-8851-363129704e52`ReqID=`fc59b8b4-51c5-4a15-9716-aedbb6de62e6`::Granted
>>> request
>>> Thread-17::DEBUG::2015-09-02
>>> 16:07:54,078::task::827::Storage.TaskManager.Task::(resourceAcquired)
>>> Task=`7936300e-8a1a-47f5-83c4-16ed19853e36`::_resourcesAcquired:
>>> Storage.8b25f3be-7574-4f7a-8851-363129704e52 (shared)
>>> Thread-17::DEBUG::2015-09-02
>>> 16:07:54,078::task::993::Storage.TaskManager.Task::(_decref)
>>> Task=`7936300e-8a1a-47f5-83c4-16ed19853e36`::ref 1 aborting False
>>> Thread-17::DEBUG::2015-09-02
>>> 16:07:54,078::misc::733::Storage.SamplingMethod::(__call__) Trying to enter
>>> sampling method (storage.sdc.refreshStorage)
>>> Thread-17::DEBUG::2015-09-02
>>> 16:07:54,078::misc::736::Storage.SamplingMethod::(__call__) Got in to
>>> sampling method
>>> Thread-17::DEBUG::2015-09-02
>>> 16:07:54,078::misc::733::Storage.SamplingMethod::(__call__) Trying to enter
>>> sampling method (storage.iscsi.rescan)
>>> Thread-17::DEBUG::2015-09-02
>>> 16:07:54,078::misc::736::Storage.SamplingMethod::(__call__) Got in to
>>> sampling method
>>> Thread-17::DEBUG::2015-09-02
>>> 16:07:54,078::iscsi::431::Storage.ISCSI::(rescan) Performing SCSI scan,
>>> this will take up to 30 seconds
>>> Thread-17::DEBUG::2015-09-02
>>> 16:07:54,078::iscsiadm::97::Storage.Misc.excCmd::(_runCmd) /usr/bin/sudo -n
>>> /sbin/iscsiadm -m session -R (cwd None)
>>> Thread-17::DEBUG::2015-09-02
>>> 16:07:54,130::misc::743::Storage.SamplingMethod::(__call__) Returning last
>>> result
>>> Thread-17::DEBUG::2015-09-02
>>> 16:07:54,130::misc::733::Storage.SamplingMethod::(__call__) Trying to enter
>>> sampling method (storage.hba.rescan)
>>> Thread-17::DEBUG::2015-09-02
>>> 16:07:54,130::misc::736::Storage.SamplingMethod::(__call__) Got in to
>>> sampling method
>>> Thread-17::DEBUG::2015-09-02
>>> 16:07:54,130::hba::56::Storage.HBA::(rescan) Starting scan
>>> Thread-17::DEBUG::2015-09-02
>>> 16:07:54,197::hba::62::Storage.HBA::(rescan) Scan finished
>>> Thread-17::DEBUG::2015-09-02
>>> 16:07:54,197::misc::743::Storage.SamplingMethod::(__call__) Returning last
>>> result
>>> Thread-17::DEBUG::2015-09-02
>>> 16:07:54,197::multipath::77::Storage.Misc.excCmd::(rescan) /usr/bin/sudo -n
>>> /usr/sbin/multipath (cwd None)
>>> Thread-17::DEBUG::2015-09-02
>>> 16:07:54,298::multipath::77::Storage.Misc.excCmd::(rescan) SUCCESS:
<err> =
>>> ''; <rc> = 0
>>> Thread-17::DEBUG::2015-09-02 16:07:54,299::utils::661::root::(execCmd)
>>> /sbin/udevadm settle --timeout=5 (cwd None)
>>> Thread-17::DEBUG::2015-09-02 16:07:54,307::utils::679::root::(execCmd)
>>> SUCCESS: <err> = ''; <rc> = 0
>>> Thread-17::DEBUG::2015-09-02
>>> 16:07:54,309::lvm::498::Storage.OperationMutex::(_invalidateAllPvs)
>>> Operation 'lvm invalidate operation' got the operation mutex
>>> Thread-17::DEBUG::2015-09-02
>>> 16:07:54,310::lvm::500::Storage.OperationMutex::(_invalidateAllPvs)
>>> Operation 'lvm invalidate operation' released the operation mutex
>>> Thread-17::DEBUG::2015-09-02
>>> 16:07:54,310::lvm::509::Storage.OperationMutex::(_invalidateAllVgs)
>>> Operation 'lvm invalidate operation' got the operation mutex
>>> Thread-17::DEBUG::2015-09-02
>>> 16:07:54,310::lvm::511::Storage.OperationMutex::(_invalidateAllVgs)
>>> Operation 'lvm invalidate operation' released the operation mutex
>>> Thread-17::DEBUG::2015-09-02
>>> 16:07:54,310::lvm::529::Storage.OperationMutex::(_invalidateAllLvs)
>>> Operation 'lvm invalidate operation' got the operation mutex
>>> Thread-17::DEBUG::2015-09-02
>>> 16:07:54,310::lvm::531::Storage.OperationMutex::(_invalidateAllLvs)
>>> Operation 'lvm invalidate operation' released the operation mutex
>>> Thread-17::DEBUG::2015-09-02
>>> 16:07:54,310::misc::743::Storage.SamplingMethod::(__call__) Returning last
>>> result
>>> Thread-17::DEBUG::2015-09-02
>>> 16:07:54,310::lvm::371::Storage.OperationMutex::(_reloadvgs) Operation
'lvm
>>> reload operation' got the operation mutex
>>> Thread-17::DEBUG::2015-09-02
>>> 16:07:54,312::lvm::291::Storage.Misc.excCmd::(cmd) /usr/bin/sudo -n
>>> /usr/sbin/lvm vgs --config ' devices { preferred_names =
["^/dev/mapper/"]
>>> ignore_suspended_devices=1 write_cache_state=0 disable_after_error_count=3
>>> obtain_device_list_from_udev=0 filter = [
>>>
'\''a|/dev/mapper/33000000100000001|/dev/mapper/Hitachi_HDS721010DLE630_MSK523Y209VK0B|'\'',
>>> '\''r|.*|'\'' ] } global { locking_type=1
prioritise_write_locks=1
>>> wait_for_locks=1 use_lvmetad=0 } backup { retain_min = 50 retain_days =
>>> 0 } ' --noheadings --units b --nosuffix --separator '|'
>>> --ignoreskippedcluster -o
>>>
uuid,name,attr,size,free,extent_size,extent_count,free_count,tags,vg_mda_size,vg_mda_free,lv_count,pv_count,pv_name
>>> 8b25f3be-7574-4f7a-8851-363129704e52 (cwd None)
>>> Thread-17::DEBUG::2015-09-02
>>> 16:07:54,478::lvm::291::Storage.Misc.excCmd::(cmd) SUCCESS: <err> =
'
>>> WARNING: lvmetad is running but disabled. Restart lvmetad before enabling
>>> it!\n'; <rc> = 0
>>> Thread-17::DEBUG::2015-09-02
>>> 16:07:54,478::lvm::416::Storage.OperationMutex::(_reloadvgs) Operation
'lvm
>>> reload operation' released the operation mutex
>>> Thread-17::DEBUG::2015-09-02
>>> 16:07:54,479::persistentDict::192::Storage.PersistentDict::(__init__)
>>> Created a persistent dict with LvMetadataRW backend
>>> Thread-17::DEBUG::2015-09-02
>>> 16:07:54,479::blockSD::337::Storage.Misc.excCmd::(readlines) /usr/bin/dd
>>> iflag=direct skip=0 bs=2048
>>> if=/dev/8b25f3be-7574-4f7a-8851-363129704e52/metadata count=1 (cwd None)
>>> Thread-17::DEBUG::2015-09-02
>>> 16:07:54,553::blockSD::337::Storage.Misc.excCmd::(readlines) SUCCESS:
<err>
>>> = '1+0 records in\n1+0 records out\n2048 bytes (2.0 kB) copied,
0.00107202
>>> s, 1.9 MB/s\n'; <rc> = 0
>>> Thread-17::DEBUG::2015-09-02
>>> 16:07:54,553::misc::260::Storage.Misc::(validateDDBytes) err: ['1+0
records
>>> in', '1+0 records out', '2048 bytes (2.0 kB) copied,
0.00107202 s, 1.9
>>> MB/s'], size: 2048
>>> Thread-17::DEBUG::2015-09-02
>>> 16:07:54,553::persistentDict::234::Storage.PersistentDict::(refresh) read
>>> lines (LvMetadataRW)=[]
>>> Thread-17::DEBUG::2015-09-02
>>> 16:07:54,553::persistentDict::252::Storage.PersistentDict::(refresh) Empty
>>> metadata
>>> Thread-17::DEBUG::2015-09-02
>>> 16:07:54,553::persistentDict::192::Storage.PersistentDict::(__init__)
>>> Created a persistent dict with VGTagMetadataRW backend
>>> Thread-17::DEBUG::2015-09-02
>>> 16:07:54,554::lvm::504::Storage.OperationMutex::(_invalidatevgs) Operation
>>> 'lvm invalidate operation' got the operation mutex
>>> Thread-17::DEBUG::2015-09-02
>>> 16:07:54,554::lvm::506::Storage.OperationMutex::(_invalidatevgs) Operation
>>> 'lvm invalidate operation' released the operation mutex
>>> Thread-17::DEBUG::2015-09-02
>>> 16:07:54,554::lvm::514::Storage.OperationMutex::(_invalidatelvs) Operation
>>> 'lvm invalidate operation' got the operation mutex
>>> Thread-17::DEBUG::2015-09-02
>>> 16:07:54,554::lvm::526::Storage.OperationMutex::(_invalidatelvs) Operation
>>> 'lvm invalidate operation' released the operation mutex
>>> Thread-17::DEBUG::2015-09-02
>>> 16:07:54,554::lvm::371::Storage.OperationMutex::(_reloadvgs) Operation
'lvm
>>> reload operation' got the operation mutex
>>> Thread-17::DEBUG::2015-09-02
>>> 16:07:54,554::lvm::291::Storage.Misc.excCmd::(cmd) /usr/bin/sudo -n
>>> /usr/sbin/lvm vgs --config ' devices { preferred_names =
["^/dev/mapper/"]
>>> ignore_suspended_devices=1 write_cache_state=0 disable_after_error_count=3
>>> obtain_device_list_from_udev=0 filter = [
>>>
'\''a|/dev/mapper/33000000100000001|/dev/mapper/Hitachi_HDS721010DLE630_MSK523Y209VK0B|'\'',
>>> '\''r|.*|'\'' ] } global { locking_type=1
prioritise_write_locks=1
>>> wait_for_locks=1 use_lvmetad=0 } backup { retain_min = 50 retain_days =
>>> 0 } ' --noheadings --units b --nosuffix --separator '|'
>>> --ignoreskippedcluster -o
>>>
uuid,name,attr,size,free,extent_size,extent_count,free_count,tags,vg_mda_size,vg_mda_free,lv_count,pv_count,pv_name
>>> 8b25f3be-7574-4f7a-8851-363129704e52 (cwd None)
>>> Thread-17::DEBUG::2015-09-02
>>> 16:07:54,685::lvm::291::Storage.Misc.excCmd::(cmd) SUCCESS: <err> =
'
>>> WARNING: lvmetad is running but disabled. Restart lvmetad before enabling
>>> it!\n'; <rc> = 0
>>> Thread-17::DEBUG::2015-09-02
>>> 16:07:54,686::lvm::416::Storage.OperationMutex::(_reloadvgs) Operation
'lvm
>>> reload operation' released the operation mutex
>>> Thread-17::DEBUG::2015-09-02
>>> 16:07:54,686::persistentDict::234::Storage.PersistentDict::(refresh) read
>>> lines (VGTagMetadataRW)=['CLASS=Data',
'DESCRIPTION=hosted_storage',
>>> 'IOOPTIMEOUTSEC=10', 'LEASERETRIES=3',
'LEASETIMESEC=60', 'LOCKPOLICY=',
>>> 'LOCKRENEWALINTERVALSEC=5', 'LOGBLKSIZE=512',
'PHYBLKSIZE=4096',
>>> 'POOL_UUID=',
>>>
u'PV0=pv:33000000100000001,uuid:kTaQQh-4LCD-OghQ-cP5D-R7MM-aj6e-kTdQf0,pestart:0,pecount:397,mapoffset:0',
>>> 'ROLE=Regular',
'SDUUID=8b25f3be-7574-4f7a-8851-363129704e52',
>>> 'TYPE=ISCSI', 'VERSION=3',
'VGUUID=tJKiwH-Cn7v-QCxd-YQrg-MUxA-fbdC-kdga8m',
>>> '_SHA_CKSUM=4a100ce5195650f43971d849835a8b3d8c0343da']
>>> Thread-17::DEBUG::2015-09-02
>>>
16:07:54,687::resourceManager::421::Storage.ResourceManager::(registerNamespace)
>>> Registering namespace '8b25f3be-7574-4f7a-8851-363129704e52_imageNS'
>>> Thread-17::DEBUG::2015-09-02
>>>
16:07:54,687::resourceManager::421::Storage.ResourceManager::(registerNamespace)
>>> Registering namespace
'8b25f3be-7574-4f7a-8851-363129704e52_volumeNS'
>>> Thread-17::DEBUG::2015-09-02
>>>
16:07:54,687::resourceManager::421::Storage.ResourceManager::(registerNamespace)
>>> Registering namespace
'8b25f3be-7574-4f7a-8851-363129704e52_lvmActivationNS'
>>> Thread-17::DEBUG::2015-09-02
>>> 16:07:54,687::lvm::428::Storage.OperationMutex::(_reloadlvs) Operation
'lvm
>>> reload operation' got the operation mutex
>>>
>>> What should I do to bring the VM engine back?
>>>
>>>
>>> 2015-09-02 16:24 GMT+01:00 Simone Tiraboschi <stirabos(a)redhat.com>:
>>>
>>>>
>>>>
>>>> On Wed, Sep 2, 2015 at 10:49 AM, wodel youchi
<wodel.youchi(a)gmail.com>
>>>> wrote:
>>>>
>>>>> I will try this afternoon to do this, but just to clarify something.
>>>>>
>>>>> the hosted_engine setup creates it's own DC the hosted_DC, which
>>>>> contains the hosted engine storage domain, I am correct?
>>>>>
>>>>
>>>> No, ovirt-hosted-engine-setup doesn't create a special datacenter.
The
>>>> default is to add the host to the Default datacenter in the default
cluster.
>>>> You could choose a different one from ovirt-hosted-engine-setup, simply
>>>> import the hosted-engine storage domain in the datacenter of the cluster
>>>> you selected.
>>>>
>>>> In setup there is a question like this:
>>>> Local storage datacenter name is an internal name
>>>> and currently will not be shown in engine's admin UI.
>>>> Please enter local datacenter name
>>>> which ask about 'Local storage datacenter' which is basically
the
>>>> description we were using for the storage pool.
>>>>
>>>>
>>>>
>>>>> if yes, where will I import the hostedengine storage domain, into
the
>>>>> default DC?
>>>>>
>>>>> 2015-09-02 8:47 GMT+01:00 Roy Golan <rgolan(a)redhat.com>:
>>>>>
>>>>>>
>>>>>>
>>>>>> On Wed, Sep 2, 2015 at 12:51 AM, wodel youchi
<wodel.youchi(a)gmail.com
>>>>>> > wrote:
>>>>>>
>>>>>>> I could finally terminate the installation, but still no vm
engine
>>>>>>> on webui
>>>>>>>
>>>>>>> I added a data domain, the default DC is up, but no engine
VM.
>>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>> Good now you need to import the HostedEngine storage domain. Try
to
>>>>>> go to
>>>>>>
>>>>>> *Storage -> Import Domain and put the path to the domain which
you
>>>>>> used in the hosted-engine setup.*
>>>>>>
>>>>>>
>>>>>> *After the domain is imported, the engine will be imported
>>>>>> automatically. *
>>>>>>
>>>>>> *This whole process will become automatic eventually. (patch is
>>>>>> written currently)*
>>>>>>
>>>>>>
>>>>>>>
>>>>>>> 2015-09-01 21:22 GMT+01:00 wodel youchi
<wodel.youchi(a)gmail.com>:
>>>>>>>
>>>>>>>> Something mounted on /rhev/data-center/mnt I'm not
sure.
>>>>>>>>
>>>>>>>> there were directories, and under these directories there
were
>>>>>>>> other directories (dom_md, ha_agent, images), and under
them there were
>>>>>>>> symbolic links to devices under /dev
>>>>>>>> (ids, inbox, leases, etc...) the devices pointed to the
lvm
>>>>>>>> partitions created by the setup.
>>>>>>>>
>>>>>>>> but the mount command didn't show anything, unlike
nfs, when I used
>>>>>>>> it the mount and df commands showed up the engine's
VM mount point.
>>>>>>>>
>>>>>>>>
>>>>>>>> 2015-09-01 20:16 GMT+01:00 Simone Tiraboschi
<stirabos(a)redhat.com>:
>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>> On Tue, Sep 1, 2015 at 7:29 PM, wodel youchi <
>>>>>>>>> wodel.youchi(a)gmail.com> wrote:
>>>>>>>>>
>>>>>>>>>> Hi,
>>>>>>>>>>
>>>>>>>>>> After removing the -x from the sql files, the
installation
>>>>>>>>>> terminated successfully, but ...
>>>>>>>>>>
>>>>>>>>>> I had a problem with vdsm, and error about
permission denied with
>>>>>>>>>> KVM module, so I restarted my machine.
>>>>>>>>>> After the reboot the ovirt-ha-agent stops
complaining about the
>>>>>>>>>> vm.conf file not present in
/var/rum/ovirt-hosted-engine-ha
>>>>>>>>>>
>>>>>>>>>> And the mount command doesn't show any iscsi
mount, the disk is
>>>>>>>>>> detected via fdisk -l
>>>>>>>>>> the lvs command returns all logical volumes
created.
>>>>>>>>>>
>>>>>>>>>> I think it's a mount problem, but since there
are many lv, I
>>>>>>>>>> don't how to mount them manually.
>>>>>>>>>>
>>>>>>>>>
>>>>>>>>> Do you have something mounted under
/rhev/data-center/mnt ?
>>>>>>>>> If not you probably hit this bug:
>>>>>>>>>
https://bugzilla.redhat.com/1258465
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>
>>>>>>>>>> LV VG
>>>>>>>>>> Attr
LSize Pool Origin Data%
>>>>>>>>>> Meta% Move
>>>>>>>>>> Log Cpy%Sync Convert
>>>>>>>>>>
>>>>>>>>>> 3b894e23-429d-43bf-b6cd-6427a387799a
>>>>>>>>>> 5445bbee-bb3a-4e6d-9614-a0c9378fe078 -wi-ao----
128,00m
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> be78c0fd-52bf-445a-9555-64061029c2d9
>>>>>>>>>> 5445bbee-bb3a-4e6d-9614-a0c9378fe078 -wi-a-----
1,00g
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> c9f74ffc-2eba-40a9-9c1c-f3b6d8e12657
>>>>>>>>>> 5445bbee-bb3a-4e6d-9614-a0c9378fe078 -wi-a-----
40,00g
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> feede664-5754-4ca2-aeb3-af7aff32ed42
>>>>>>>>>> 5445bbee-bb3a-4e6d-9614-a0c9378fe078 -wi-a-----
128,00m
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> ids
>>>>>>>>>>
5445bbee-bb3a-4e6d-9614-a0c9378fe078
>>>>>>>>>> -wi-ao---- 128,00m
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> inbox
>>>>>>>>>>
5445bbee-bb3a-4e6d-9614-a0c9378fe078
>>>>>>>>>> -wi-a----- 128,00m
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> leases
>>>>>>>>>>
5445bbee-bb3a-4e6d-9614-a0c9378fe078
>>>>>>>>>> -wi-a----- 2,00g
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> master
>>>>>>>>>>
5445bbee-bb3a-4e6d-9614-a0c9378fe078
>>>>>>>>>> -wi-a----- 1,00g
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> metadata
>>>>>>>>>>
5445bbee-bb3a-4e6d-9614-a0c9378fe078 -wi-a-----
>>>>>>>>>> 512,00m
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> outbox
>>>>>>>>>>
5445bbee-bb3a-4e6d-9614-a0c9378fe078
>>>>>>>>>> -wi-a----- 128,00m
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>> 2015-09-01 16:57 GMT+01:00 Simone Tiraboschi
<stirabos(a)redhat.com
>>>>>>>>>> >:
>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> On Tue, Sep 1, 2015 at 5:08 PM, wodel youchi
<
>>>>>>>>>>> wodel.youchi(a)gmail.com> wrote:
>>>>>>>>>>>
>>>>>>>>>>>> Hi again,
>>>>>>>>>>>>
>>>>>>>>>>>> I tried with the snapshot repository, but
I am having this
>>>>>>>>>>>> error while executing engine-setup
>>>>>>>>>>>>
>>>>>>>>>>>> [ INFO ] Creating/refreshing Engine
database schema
>>>>>>>>>>>> [ ERROR ] Failed to execute stage
'Misc configuration':
>>>>>>>>>>>> Command
'/usr/share/ovirt-engine/dbscripts/schema.sh' failed to execu
>>>>>>>>>>>> te
>>>>>>>>>>>> [ INFO ] DNF Performing DNF transaction
rollback
>>>>>>>>>>>> [ INFO ] Rolling back database schema
>>>>>>>>>>>> [ INFO ] Clearing Engine database
engine
>>>>>>>>>>>> [ ERROR ] Engine database rollback
failed: must be owner of
>>>>>>>>>>>> schema pg_catalog
>>>>>>>>>>>> [ INFO ] Stage: Clean up
>>>>>>>>>>>> Log file is located at
>>>>>>>>>>>>
/var/log/ovirt-engine/setup/ovirt-engine-setup-20150901153202-w0ds25.log
>>>>>>>>>>>> [ INFO ] Generating answer file
>>>>>>>>>>>>
'/var/lib/ovirt-engine/setup/answers/20150901153939-setup.conf'
>>>>>>>>>>>> [ INFO ] Stage: Pre-termination
>>>>>>>>>>>> [ INFO ] Stage: Termination
>>>>>>>>>>>> [ ERROR ] Execution of setup failed
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> and in the deployement log I have these
errors
>>>>>>>>>>>>
>>>>>>>>>>>> Saving custom users permissions on
database objects...
>>>>>>>>>>>> upgrade script detected a change in
Config, View or Stored
>>>>>>>>>>>> Procedure...
>>>>>>>>>>>> Running upgrade shell script
>>>>>>>>>>>>
'/usr/share/ovirt-engine/dbscripts/upgrade/pre_upgrade/0000_config.sql'...
>>>>>>>>>>>> Running upgrade shell script
>>>>>>>>>>>>
'/usr/share/ovirt-engine/dbscripts/upgrade/pre_upgrade/0010_custom.sql'...
>>>>>>>>>>>> Running upgrade shell script
>>>>>>>>>>>>
'/usr/share/ovirt-engine/dbscripts/upgrade/pre_upgrade/0020_add_materialized_views_table.sql'...
>>>>>>>>>>>> Running upgrade shell script
>>>>>>>>>>>>
'/usr/share/ovirt-engine/dbscripts/upgrade/pre_upgrade/0030_materialized_views_extensions.sql'...
>>>>>>>>>>>> Running upgrade shell script
>>>>>>>>>>>>
'/usr/share/ovirt-engine/dbscripts/pre_upgrade/0040_extend_installed_by_column.sql'...
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> 2015-09-01 15:39:35 DEBUG
>>>>>>>>>>>>
otopi.plugins.ovirt_engine_setup.ovirt_engine.db.schema plugin.execute:941
>>>>>>>>>>>> execute-output:
['/usr/share/ovirt-engine/dbscripts/schema.sh', '-s',
>>>>>>>>>>>> 'localhost', '-p',
'5432', '-u', 'engine', '-d', 'engine',
'-l', '/var/log
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>>
*/ovirt-engine/setup/ovirt-engine-setup-20150901153202-w0ds25.log',
>>>>>>>>>>>> '-c', 'apply']
>>>>>>>>>>>>
stderr:/usr/share/ovirt-engine/dbscripts/upgrade/pre_upgrade/0000_config.sql:
>>>>>>>>>>>> ligne 1: /bin : is a
>>>>>>>>>>>>
directory/usr/share/ovirt-engine/dbscripts/upgrade/pre_upgrade/0000_config.sql:
>>>>>>>>>>>> ligne 2: DATABASE : command not
>>>>>>>>>>>>
found/usr/share/ovirt-engine/dbscripts/upgrade/pre_upgrade/0000_config.sql:
>>>>>>>>>>>> ligne 4: This : command not
>>>>>>>>>>>>
found/usr/share/ovirt-engine/dbscripts/upgrade/pre_upgrade/0000_config.sql:
>>>>>>>>>>>> ligne 5: The : command not
>>>>>>>>>>>>
found/usr/share/ovirt-engine/dbscripts/upgrade/pre_upgrade/0000_config.sql:
>>>>>>>>>>>> ligne 6: Add : command not
>>>>>>>>>>>>
found/usr/share/ovirt-engine/dbscripts/upgrade/pre_upgrade/0000_config.sql:
>>>>>>>>>>>> ligne 7: syntax error near the symbole
unexpected « (
>>>>>>>>>>>>
»/usr/share/ovirt-engine/dbscripts/upgrade/pre_upgrade/0000_config.sql:
>>>>>>>>>>>> ligne 7: ` Update section (w/o
overriding current
>>>>>>>>>>>>
value)'/usr/share/ovirt-engine/dbscripts/upgrade/pre_upgrade/0010_custom.sql:
>>>>>>>>>>>> ligne 1: /bin : is a
>>>>>>>>>>>>
directory/usr/share/ovirt-engine/dbscripts/upgrade/pre_upgrade/0010_custom.sql:
>>>>>>>>>>>> ligne 2: Currently : command not
>>>>>>>>>>>>
found/usr/share/ovirt-engine/dbscripts/upgrade/pre_upgrade/0010_custom.sql:
>>>>>>>>>>>> ligne 3: This : command not
>>>>>>>>>>>>
found/usr/share/ovirt-engine/dbscripts/upgrade/pre_upgrade/0010_custom.sql:
>>>>>>>>>>>> ligne 4: This : command not
>>>>>>>>>>>>
found/usr/share/ovirt-engine/dbscripts/upgrade/pre_upgrade/0010_custom.sql:
>>>>>>>>>>>> ligne 5: So, : command not
>>>>>>>>>>>>
found/usr/share/ovirt-engine/dbscripts/upgrade/pre_upgrade/0010_custom.sql:
>>>>>>>>>>>> ligne 6: Since : command not
>>>>>>>>>>>>
found/usr/share/ovirt-engine/dbscripts/upgrade/pre_upgrade/0010_custom.sql:
>>>>>>>>>>>> ligne 7: bin/ : is a
>>>>>>>>>>>>
directory/usr/share/ovirt-engine/dbscripts/upgrade/pre_upgrade/0010_custom.sql:
>>>>>>>>>>>> ligne 9: update : command not
>>>>>>>>>>>>
found/usr/share/ovirt-engine/dbscripts/upgrade/pre_upgrade/0010_custom.sql:
>>>>>>>>>>>> ligne 10: *
>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> *syntax error near the symbole unexpected
« (
>>>>>>>>>>>>
»/usr/share/ovirt-engine/dbscripts/upgrade/pre_upgrade/0010_custom.sql:
>>>>>>>>>>>> ligne 10: `and exists(select 1 from
schema_version where version =
>>>>>>>>>>>> '03010250' and current =
>>>>>>>>>>>>
true);'/usr/share/ovirt-engine/dbscripts/upgrade/pre_upgrade/0020_add_materialized_views_table.sql:
>>>>>>>>>>>> ligne 1: -- : command not
>>>>>>>>>>>>
found/usr/share/ovirt-engine/dbscripts/upgrade/pre_upgrade/0020_add_materialized_views_table.sql:
>>>>>>>>>>>> ligne 2: **syntax error near the symbole
unexpected « ( »*
>>>>>>>>>>>>
/usr/share/ovirt-engine/dbscripts/upgrade/pre_upgrade/0020_add_materialized_views_table.sql:
>>>>>>>>>>>> ligne 2: `CREATE FUNCTION
__temp__0030_add_materialized_views_table()'
>>>>>>>>>>>>
/usr/share/ovirt-engine/dbscripts/upgrade/pre_upgrade/0030_materialized_views_extensions.sql:
>>>>>>>>>>>> ligne 1: -- : commande introuvable
>>>>>>>>>>>>
/usr/share/ovirt-engine/dbscripts/upgrade/pre_upgrade/0030_materialized_views_extensions.sql:
>>>>>>>>>>>> ligne 2: erreur de syntaxe près du
symbole inattendu « ( »
>>>>>>>>>>>>
/usr/share/ovirt-engine/dbscripts/upgrade/pre_upgrade/0030_materialized_views_extensions.sql:
>>>>>>>>>>>> ligne 2: `select
fn_db_add_column('materialized_views',
>>>>>>>>>>>> 'min_refresh_rate_in_sec',
'int default 0');'
>>>>>>>>>>>>
/usr/share/ovirt-engine/dbscripts/upgrade/pre_upgrade/0040_extend_installed_by_column.sql:
>>>>>>>>>>>> ligne 1: erreur de syntaxe près du
symbole inattendu « ( »
>>>>>>>>>>>>
/usr/share/ovirt-engine/dbscripts/upgrade/pre_upgrade/0040_extend_installed_by_column.sql:
>>>>>>>>>>>> ligne 1: `ALTER TABLE schema_version
ALTER COLUMN installed_by TYPE
>>>>>>>>>>>> varchar(63);'
>>>>>>>>>>>> 2015-09-01 15:39:35 DEBUG otopi.context
>>>>>>>>>>>> context._executeMethod:156 method
exception
>>>>>>>>>>>> Traceback (most recent call last):
>>>>>>>>>>>> File
"/usr/lib/python2.7/site-packages/otopi/context.py",
>>>>>>>>>>>> line 146, in _executeMethod
>>>>>>>>>>>> method['method']()
>>>>>>>>>>>> File
>>>>>>>>>>>>
"/usr/share/ovirt-engine/setup/bin/../plugins/ovirt-engine-setup/ovirt-engine/db/schema.py",
>>>>>>>>>>>> line 291, in _misc
>>>>>>>>>>>> oenginecons.EngineDBEnv.PGPASS_FILE
>>>>>>>>>>>> File
"/usr/lib/python2.7/site-packages/otopi/plugin.py", line
>>>>>>>>>>>> 946, in execute
>>>>>>>>>>>> command=args[0],
>>>>>>>>>>>> RuntimeError: Command
>>>>>>>>>>>>
'/usr/share/ovirt-engine/dbscripts/schema.sh' failed to execute
>>>>>>>>>>>> 2015-09-01 15:39:35 ERROR otopi.context
>>>>>>>>>>>> context._executeMethod:165 Failed to
execute stage 'Misc configuration':
>>>>>>>>>>>> Command
'/usr/share/ovirt-engine/dbscripts/schema.sh' failed to execute
>>>>>>>>>>>> 2015-09-01 15:39:35 DEBUG
otopi.transaction
>>>>>>>>>>>> transaction.abort:134 aborting 'DNF
Transaction'
>>>>>>>>>>>> 2015-09-01 15:39:35 DEBUG
>>>>>>>>>>>> otopi.plugins.otopi.packagers.dnfpackager
dnfpackager.verbose:90 DNF
>>>>>>>>>>>> Closing transaction with rollback
>>>>>>>>>>>> 2015-09-01 15:39:35 INFO
>>>>>>>>>>>> otopi.plugins.otopi.packagers.dnfpackager
dnfpackager.info:94
>>>>>>>>>>>> DNF Performing DNF transaction rollback
>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>> It was an issue with package building: all
the sql files where
>>>>>>>>>>> executable and so the issue.
>>>>>>>>>>> We fixed it and tomorrow build should be OK.
If you prefer to
>>>>>>>>>>> continue right now simply recursively remove
the x attribute on each sql
>>>>>>>>>>> file under /usr/share/ovirt-engine/dbscripts
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>> 2015-09-01 13:04 GMT+01:00 Simone
Tiraboschi <
>>>>>>>>>>>> stirabos(a)redhat.com>:
>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>> On Tue, Sep 1, 2015 at 12:40 PM,
Yedidyah Bar David <
>>>>>>>>>>>>> didi(a)redhat.com> wrote:
>>>>>>>>>>>>>
>>>>>>>>>>>>>> On Tue, Sep 1, 2015 at 1:25 PM,
wodel youchi <
>>>>>>>>>>>>>> wodel.youchi(a)gmail.com>
wrote:
>>>>>>>>>>>>>> > Hi,
>>>>>>>>>>>>>> >
>>>>>>>>>>>>>> > I am using the repo of the
3.6 version
>>>>>>>>>>>>>> > (
>>>>>>>>>>>>>>
http://plain.resources.ovirt.org/pub/ovirt-3.6-pre/rpm/fc22/)
>>>>>>>>>>>>>> >
>>>>>>>>>>>>>> > I installed the
ovirt-hosted-engine-setup with it's
>>>>>>>>>>>>>> dependencies,and the
>>>>>>>>>>>>>> > ovirt-hosted-engine-ha
package is one of them.
>>>>>>>>>>>>>> >
>>>>>>>>>>>>>> > Correction: The problem with
this version
>>>>>>>>>>>>>> >
>>>>>>>>>>>>>>
ovirt-hosted-engine-ha-1.3.0-0.0.master.20150819082341.20150819082338.git183a4ff.fc22.noarch.rpm,
>>>>>>>>>>>>>> > is that after the
installation is done, the service
>>>>>>>>>>>>>> ovirt-ha-agent crashes
>>>>>>>>>>>>>> > after being started, see the
bug :
>>>>>>>>>>>>>> >
>>>>>>>>>>>>>> >
https://bugzilla.redhat.com/show_bug.cgi?id=1254745
>>>>>>>>>>>>>> >
>>>>>>>>>>>>>> > A new version was produced,
I downloaded it manually a few
>>>>>>>>>>>>>> days ago, this is
>>>>>>>>>>>>>> > it :
>>>>>>>>>>>>>> >
>>>>>>>>>>>>>>
ovirt-hosted-engine-ha-1.3.0-0.0.master.20150820064645.20150820064642.git02529e0.fc22.noarch.rpm
>>>>>>>>>>>>>> >
>>>>>>>>>>>>>> > This one did correct the
problem, but it's not present
>>>>>>>>>>>>>> anymore on the
>>>>>>>>>>>>>> > repository.
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> Was this on ovirt-3.6-pre?
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> ovirt-3.6-snapshot has a newer
version.
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> >
>>>>>>>>>>>>>> > For Simone: yes I did added
an NFS4 data domain, but no
>>>>>>>>>>>>>> success so far, no
>>>>>>>>>>>>>> > VM engine present.
>>>>>>>>>>>>>> >
>>>>>>>>>>>>>> > Regards.
>>>>>>>>>>>>>> >
>>>>>>>>>>>>>> > 2015-09-01 11:15 GMT+01:00
Simone Tiraboschi <
>>>>>>>>>>>>>> stirabos(a)redhat.com>:
>>>>>>>>>>>>>> >>
>>>>>>>>>>>>>> >>
>>>>>>>>>>>>>> >>
>>>>>>>>>>>>>> >> On Tue, Sep 1, 2015 at
11:46 AM, Yedidyah Bar David <
>>>>>>>>>>>>>> didi(a)redhat.com>
>>>>>>>>>>>>>> >> wrote:
>>>>>>>>>>>>>> >>>
>>>>>>>>>>>>>> >>> On Tue, Sep 1, 2015
at 11:25 AM, wodel youchi <
>>>>>>>>>>>>>> wodel.youchi(a)gmail.com>
>>>>>>>>>>>>>> >>> wrote:
>>>>>>>>>>>>>> >>> > Hi,
>>>>>>>>>>>>>> >>> >
>>>>>>>>>>>>>> >>> > Another test of
ovirt hosted-engine on FC22 using ovirt
>>>>>>>>>>>>>> 3.6 Beta3.
>>>>>>>>>>>>>> >>> > VM engine is
also a FC22
>>>>>>>>>>>>>> >>> >
>>>>>>>>>>>>>> >>> > Problem:
>>>>>>>>>>>>>> >>> > - No VM engine
on webui
>>>>>>>>>>>>>> >>>
>>>>>>>>>>>>>> >>> This is still not
supported, see/follow [1].
>>>>>>>>>>>>>> >>>
>>>>>>>>>>>>>> >>> [1]
https://bugzilla.redhat.com/show_bug.cgi?id=1224889
>>>>>>>>>>>>>> >>
>>>>>>>>>>>>>> >>
>>>>>>>>>>>>>> >> ? :-)
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> Sorry :-(
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
https://bugzilla.redhat.com/show_bug.cgi?id=1160094
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>> This is just about editing the VM
from the web GUI but in
>>>>>>>>>>>>> order to be able to edit the engine
VM you should be able at least to find
>>>>>>>>>>>>> the engine VM in the engine as it was
also in 3.5
>>>>>>>>>>>>>
>>>>>>>>>>>>> I'll try to reproduce verifying
another patch
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>> I see that all patches there are
merged, but bug is in POST.
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> >>
>>>>>>>>>>>>>> >> Did you also try adding
an additional storage domain for
>>>>>>>>>>>>>> regular VMs?
>>>>>>>>>>>>>> >> engine-VM will be shown
on the engine only when you add at
>>>>>>>>>>>>>> least one
>>>>>>>>>>>>>> >> additional storage
domain for regulars VM and the whole
>>>>>>>>>>>>>> datacenter goes up:
>>>>>>>>>>>>>> >>
https://bugzilla.redhat.com/show_bug.cgi?id=1222010#c1
>>>>>>>>>>>>>> >>
>>>>>>>>>>>>>> >>
>>>>>>>>>>>>>> >>>
>>>>>>>>>>>>>> >>> > Test
Environment
>>>>>>>>>>>>>> >>> > Just two
machines:
>>>>>>>>>>>>>> >>> >
>>>>>>>>>>>>>> >>> > 1 - Machine 1
used as storage :
>>>>>>>>>>>>>> >>> > - iscsi
target with a raw file for the VM engine
>>>>>>>>>>>>>> storage
>>>>>>>>>>>>>> >>> > - NFS4 for
other data domains
>>>>>>>>>>>>>> >>> >
>>>>>>>>>>>>>> >>> > 2 - Machine 2
used as hypervisor
>>>>>>>>>>>>>> >>> >
>>>>>>>>>>>>>> >>> >
>>>>>>>>>>>>>> >>> > The
installation went without problem, but as always,
>>>>>>>>>>>>>> the VM engine is
>>>>>>>>>>>>>> >>> > not
>>>>>>>>>>>>>> >>> > present on the
webui.
>>>>>>>>>>>>>> >>> >
>>>>>>>>>>>>>> >>> >
>>>>>>>>>>>>>> >>> > PS:
>>>>>>>>>>>>>> >>> > 1- I gave the
VM engine just 2Gb of memory since I
>>>>>>>>>>>>>> don't have too much
>>>>>>>>>>>>>> >>> > RAM
>>>>>>>>>>>>>> >>> > on hypervisor,
could that be the cause of the problem?
>>>>>>>>>>>>>> >>>
>>>>>>>>>>>>>> >>> Shouldn't be
related
>>>>>>>>>>>>>> >>>
>>>>>>>>>>>>>> >>> > 2- This version
of
>>>>>>>>>>>>>> >>> >
>>>>>>>>>>>>>> >>> >
>>>>>>>>>>>>>>
ovirt-hosted-engine-ha-1.3.0-0.0.master.20150424113926.20150424113923.git7c14f4c.fc22.noarch.rpm
>>>>>>>>>>>>>> >>> > package is
causing the ovirt-ha-agent to crash, it was
>>>>>>>>>>>>>> replaced with
>>>>>>>>>>>>>> >>> > another
>>>>>>>>>>>>>> >>> > which I still
have
>>>>>>>>>>>>>> >>> >
>>>>>>>>>>>>>> >>> >
>>>>>>>>>>>>>>
ovirt-hosted-engine-ha-1.3.0-0.0.master.20150820064645.20150820064642.git02529e0.fc22.noarch.rpm,
>>>>>>>>>>>>>> >>> > but it's
not present on the repository, I had to update
>>>>>>>>>>>>>> the package
>>>>>>>>>>>>>> >>> > manually
>>>>>>>>>>>>>> >>> > at the end of
ovirt-hosted-engine-setup installation.
>>>>>>>>>>>>>> >>>
>>>>>>>>>>>>>> >>> Not sure I follow.
>>>>>>>>>>>>>> >>>
>>>>>>>>>>>>>> >>> What exact repo was
used?
>>>>>>>>>>>>>> >>>
>>>>>>>>>>>>>> >>> hosted-engine
--deploy does not update/install packages
>>>>>>>>>>>>>> for you (as
>>>>>>>>>>>>>> >>> does engine-setup),
>>>>>>>>>>>>>> >>> it's up to you
to make sure what you want/need is
>>>>>>>>>>>>>> installed prior to
>>>>>>>>>>>>>> >>> running it.
>>>>>>>>>>>>>> >>>
>>>>>>>>>>>>>> >>> Best,
>>>>>>>>>>>>>> >>> --
>>>>>>>>>>>>>> >>> Didi
>>>>>>>>>>>>>> >>>
_______________________________________________
>>>>>>>>>>>>>> >>> Users mailing list
>>>>>>>>>>>>>> >>> Users(a)ovirt.org
>>>>>>>>>>>>>> >>>
http://lists.ovirt.org/mailman/listinfo/users
>>>>>>>>>>>>>> >>
>>>>>>>>>>>>>> >>
>>>>>>>>>>>>>> >
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>>
>>>>>>>>>>>>>> --
>>>>>>>>>>>>>> Didi
>>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>>
>>>>>>>>>>>>
>>>>>>>>>>>
>>>>>>>>>>
>>>>>>>>>
>>>>>>>>
>>>>>>>
>>>>>>
>>>>>
>>>>
>>>
>