Migrate between 3.3.4 and 3.4.1

I am trying to migrate my VMs from my old host running ovirt 3.3.4 to a new setup running 3.4.1. My basic set up is: OLD vmhost1 - ovirt 3.3.4 - NFS storages NEW ovirtmgr - ovirt 3.4.1 (virt only setup) - gluster storage domains vmhost2 - cluster1 vmhost3 - cluster2 vmhost4 - cluster2 vmhost5 - cluster3 vmhost6 - cluster3 My gluster volumes are created via gluster command line and I have All hosts are running scientific Linux 6.5, and the intention is to migrate vmhost1 to new environment cluster1. I have an NFS export storage domain which I am using to migrate VMs from vmhost1. Volume Name: vol-vminf Type: Distributed-Replicate Volume ID: b0b456bb-76e9-42e7-bb95-3415db79d631 Status: Started Number of Bricks: 2 x 2 = 4 Transport-type: tcp Bricks: Brick1: vmhost3:/storage/inf/br-inf Brick2: vmhost4:/storage/inf/br-inf Brick3: vmhost5:/storage/inf/br-inf Brick4: vmhost6:/storage/inf/br-inf Options Reconfigured: storage.owner-gid: 36 storage.owner-uid: 36 server.allow-insecure: on Volume Name: vol-vmimages Type: Distribute Volume ID: 91e2cf8b-2662-4c26-b937-84b8f5b62e2b Status: Started Number of Bricks: 4 Transport-type: tcp Bricks: Brick1: vmhost3:/storage/vmimages/br-vmimages Brick2: vmhost3:/storage/vmimages/br-vmimages Brick3: vmhost3:/storage/vmimages/br-vmimages Brick4: vmhost3:/storage/vmimages/br-vmimages Options Reconfigured: storage.owner-gid: 36 storage.owner-uid: 36 server.allow-insecure: on I have had many varying fail results which I have tried to match up with threads here, and I am now a bit stuck and would appreciate any help. I quite regularly cannot import vms, and I now cannot create a new disk for a new VM (no import). The error always seems to boild down to the following error on engine.log (the following specifically for the create of a new disk image on a gluster storage domain): 2014-05-20 08:51:21,136 INFO [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] (ajp--127.0.0.1-8702-9) [4637af09] Correlation ID: 2b0b55ab, Job ID: 1a583643-e28a-4f09-a39d-46e4fc6d20b8, Call Stack: null, Custom Event ID: -1, Message: Add-Disk operation of rhel-7_Disk1 was initiated on VM rhel-7 by peter.harris. 2014-05-20 08:51:21,137 INFO [org.ovirt.engine.core.bll.SPMAsyncTask] (ajp--127.0.0.1-8702-9) [4637af09] BaseAsyncTask::startPollingTask: Starting to poll task 720b4d92-1425-478c-8351-4ff827b8f728. 2014-05-20 08:51:28,077 INFO [org.ovirt.engine.core.bll.AsyncTaskManager] (DefaultQuartzScheduler_Worker-19) Polling and updating Async Tasks: 1 tasks, 1 tasks to poll now 2014-05-20 08:51:28,084 ERROR [org.ovirt.engine.core.vdsbroker.vdsbroker.HSMGetAllTasksStatusesVDSCommand] (DefaultQuartzScheduler_Worker-19) Failed in HSMGetAllTasksStatusesVDS method 2014-05-20 08:51:28,085 INFO [org.ovirt.engine.core.bll.SPMAsyncTask] (DefaultQuartzScheduler_Worker-19) SPMAsyncTask::PollTask: Polling task 720b4d92-1425-478c-8351-4ff827b8f728 (Parent Command AddDisk, Parameters Type org.ovirt.engine.core.common.asynctasks.AsyncTaskParameters) returned status finished, result 'cleanSuccess'. 2014-05-20 08:51:28,104 ERROR [org.ovirt.engine.core.bll.SPMAsyncTask] (DefaultQuartzScheduler_Worker-19) BaseAsyncTask::LogEndTaskFailure: Task 720b4d92-1425-478c-8351-4ff827b8f728 (Parent Command AddDisk, Parameters Type org.ovirt.engine.core.common.asynctasks.AsyncTaskParameters) ended with failure:^M -- Result: cleanSuccess^M -- Message: VDSGenericException: VDSErrorException: Failed to HSMGetAllTasksStatusesVDS, error = [Errno 2] No such file or directory: '/rhev/data-center/06930787-a091-49a3-8217-1418c5a9881e/967aec77-46d5-418b-8979-d0a86389a77b/images/7726b997-7e58-45f8-a5a6-9cb9a689a45a', code = 100,^M -- Exception: VDSGenericException: VDSErrorException: Failed to HSMGetAllTasksStatusesVDS, error = [Errno 2] No such file or directory: '/rhev/data-center/06930787-a091-49a3-8217-1418c5a9881e/967aec77-46d5-418b-8979-d0a86389a77b/images/7726b997-7e58-45f8-a5a6-9cb9a689a45a', code = 100 Certainly, if I check /rhev/data-center/06930787-a091-49a3-8217-1418c5a9881e/ on the SPM server, there is no 967aec77-46d5-418b-8979-d0a86389a77b subdirectory. The only elements I have are NFS mounts. There appear to be no errors in the SPM vdsm.log for this disk ============= When I tried to import the vm (the one that I then tried to create from scratch above), I had the following errors in the SPM vdsm log: Thread-2220::DEBUG::2014-05-20 08:35:05,255::task::595::TaskManager.Task::(_updateState) Task=`15bc07b5-201f-4bba-bf5f-f79eb92c6a61`::moving from state init -> state preparing Thread-2220::INFO::2014-05-20 08:35:05,255::logUtils::44::dispatcher::(wrapper) Run and protect: deleteImage(sdUUID='615647e2-1f60-47e1-8e55-be9f7ead6f15', spUUID='06930787-a091-49a3-8217-1418c5a9881e', imgUUID='80ed133c-fd72-4d35-aae5-e1313be3cf23', postZero='false', force='false') Thread-2220::DEBUG::2014-05-20 08:35:05,255::resourceManager::198::ResourceManager.Request::(__init__) ResName=`Storage.80ed133c-fd72-4d35-aae5-e1313be3cf23`ReqID=`499de454-c563-4156-a3ed-13b7eb9defa6`::Request was made in '/usr/share/vdsm/storage/hsm.py' line '1496' at 'deleteImage' Thread-2220::DEBUG::2014-05-20 08:35:05,255::resourceManager::542::ResourceManager::(registerResource) Trying to register resource 'Storage.80ed133c-fd72-4d35-aae5-e1313be3cf23' for lock type 'exclusive' Thread-2220::DEBUG::2014-05-20 08:35:05,255::resourceManager::601::ResourceManager::(registerResource) Resource 'Storage.80ed133c-fd72-4d35-aae5-e1313be3cf23' is free. Now locking as 'exclusive' (1 active user) Thread-2220::DEBUG::2014-05-20 08:35:05,256::resourceManager::238::ResourceManager.Request::(grant) ResName=`Storage.80ed133c-fd72-4d35-aae5-e1313be3cf23`ReqID=`499de454-c563-4156-a3ed-13b7eb9defa6`::Granted request Thread-2220::DEBUG::2014-05-20 08:35:05,256::task::827::TaskManager.Task::(resourceAcquired) Task=`15bc07b5-201f-4bba-bf5f-f79eb92c6a61`::_resourcesAcquired: Storage.80ed133c-fd72-4d35-aae5-e1313be3cf23 (exclusive) Thread-2220::DEBUG::2014-05-20 08:35:05,256::task::990::TaskManager.Task::(_decref) Task=`15bc07b5-201f-4bba-bf5f-f79eb92c6a61`::ref 1 aborting False Thread-2220::DEBUG::2014-05-20 08:35:05,256::resourceManager::198::ResourceManager.Request::(__init__) ResName=`Storage.615647e2-1f60-47e1-8e55-be9f7ead6f15`ReqID=`73f79517-f13a-4e5b-999a-6f1994d2818a`::Request was made in '/usr/share/vdsm/storage/hsm.py' line '1497' at 'deleteImage' Thread-2220::DEBUG::2014-05-20 08:35:05,256::resourceManager::542::ResourceManager::(registerResource) Trying to register resource 'Storage.615647e2-1f60-47e1-8e55-be9f7ead6f15' for lock type 'shared' Thread-2220::DEBUG::2014-05-20 08:35:05,257::resourceManager::601::ResourceManager::(registerResource) Resource 'Storage.615647e2-1f60-47e1-8e55-be9f7ead6f15' is free. Now locking as 'shared' (1 active user) Thread-2220::DEBUG::2014-05-20 08:35:05,257::resourceManager::238::ResourceManager.Request::(grant) ResName=`Storage.615647e2-1f60-47e1-8e55-be9f7ead6f15`ReqID=`73f79517-f13a-4e5b-999a-6f1994d2818a`::Granted request Thread-2220::DEBUG::2014-05-20 08:35:05,257::task::827::TaskManager.Task::(resourceAcquired) Task=`15bc07b5-201f-4bba-bf5f-f79eb92c6a61`::_resourcesAcquired: Storage.615647e2-1f60-47e1-8e55-be9f7ead6f15 (shared) Thread-2220::DEBUG::2014-05-20 08:35:05,257::task::990::TaskManager.Task::(_decref) Task=`15bc07b5-201f-4bba-bf5f-f79eb92c6a61`::ref 1 aborting False Thread-2220::ERROR::2014-05-20 08:35:05,266::hsm::1502::Storage.HSM::(deleteImage) Empty or not found image 80ed133c-fd72-4d35-aae5-e1313be3cf23 in SD 615647e2-1f60-47e1-8e55-be9f7ead6f15. {'1f41529a-e02e-4cd8-987c-b1ea4fcba2be': ImgsPar(imgs=('290f5cdf-b5d7-462b-958d-d41458a26bf6',), parent=None), '1748a8f0-8668-4f21-9b26-d2e3b180e35b': ImgsPar(imgs=('67fd552b-8b3d-4117-82d2-e801bb600992',), parent=None)} Thread-2220::ERROR::2014-05-20 08:35:05,266::task::866::TaskManager.Task::(_setError) Task=`15bc07b5-201f-4bba-bf5f-f79eb92c6a61`::Unexpected error When I installed/setup ovirt-engine, I did chose NFS as the file system. I am clearly doing something weird

A slight update to this. I did find that the gluster mount points re appear under SPM /rhev/data-center if I detach and then deactivate the gluster storages under the data storages. As this means shutting down any VMs that are running from that storage, is there any known way of getting the gluster points to reattach without dectivcate/reactivate a storage in the gui? Cheers Peter On 20 May 2014 09:37, Peter Harris <doilooksensible@gmail.com> wrote:
I am trying to migrate my VMs from my old host running ovirt 3.3.4 to a new setup running 3.4.1. My basic set up is:
OLD vmhost1 - ovirt 3.3.4 - NFS storages
NEW ovirtmgr - ovirt 3.4.1 (virt only setup) - gluster storage domains vmhost2 - cluster1 vmhost3 - cluster2 vmhost4 - cluster2 vmhost5 - cluster3 vmhost6 - cluster3
My gluster volumes are created via gluster command line and I have
All hosts are running scientific Linux 6.5, and the intention is to migrate vmhost1 to new environment cluster1.
I have an NFS export storage domain which I am using to migrate VMs from vmhost1. Volume Name: vol-vminf Type: Distributed-Replicate Volume ID: b0b456bb-76e9-42e7-bb95-3415db79d631 Status: Started Number of Bricks: 2 x 2 = 4 Transport-type: tcp Bricks: Brick1: vmhost3:/storage/inf/br-inf Brick2: vmhost4:/storage/inf/br-inf Brick3: vmhost5:/storage/inf/br-inf Brick4: vmhost6:/storage/inf/br-inf Options Reconfigured: storage.owner-gid: 36 storage.owner-uid: 36 server.allow-insecure: on
Volume Name: vol-vmimages Type: Distribute Volume ID: 91e2cf8b-2662-4c26-b937-84b8f5b62e2b Status: Started Number of Bricks: 4 Transport-type: tcp Bricks: Brick1: vmhost3:/storage/vmimages/br-vmimages Brick2: vmhost3:/storage/vmimages/br-vmimages Brick3: vmhost3:/storage/vmimages/br-vmimages Brick4: vmhost3:/storage/vmimages/br-vmimages Options Reconfigured: storage.owner-gid: 36 storage.owner-uid: 36 server.allow-insecure: on
I have had many varying fail results which I have tried to match up with threads here, and I am now a bit stuck and would appreciate any help.
I quite regularly cannot import vms, and I now cannot create a new disk for a new VM (no import). The error always seems to boild down to the following error on engine.log (the following specifically for the create of a new disk image on a gluster storage domain):
2014-05-20 08:51:21,136 INFO [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] (ajp--127.0.0.1-8702-9) [4637af09] Correlation ID: 2b0b55ab, Job ID: 1a583643-e28a-4f09-a39d-46e4fc6d20b8, Call Stack: null, Custom Event ID: -1, Message: Add-Disk operation of rhel-7_Disk1 was initiated on VM rhel-7 by peter.harris. 2014-05-20 08:51:21,137 INFO [org.ovirt.engine.core.bll.SPMAsyncTask] (ajp--127.0.0.1-8702-9) [4637af09] BaseAsyncTask::startPollingTask: Starting to poll task 720b4d92-1425-478c-8351-4ff827b8f728. 2014-05-20 08:51:28,077 INFO [org.ovirt.engine.core.bll.AsyncTaskManager] (DefaultQuartzScheduler_Worker-19) Polling and updating Async Tasks: 1 tasks, 1 tasks to poll now 2014-05-20 08:51:28,084 ERROR [org.ovirt.engine.core.vdsbroker.vdsbroker.HSMGetAllTasksStatusesVDSCommand] (DefaultQuartzScheduler_Worker-19) Failed in HSMGetAllTasksStatusesVDS method 2014-05-20 08:51:28,085 INFO [org.ovirt.engine.core.bll.SPMAsyncTask] (DefaultQuartzScheduler_Worker-19) SPMAsyncTask::PollTask: Polling task 720b4d92-1425-478c-8351-4ff827b8f728 (Parent Command AddDisk, Parameters Type org.ovirt.engine.core.common.asynctasks.AsyncTaskParameters) returned status finished, result 'cleanSuccess'. 2014-05-20 08:51:28,104 ERROR [org.ovirt.engine.core.bll.SPMAsyncTask] (DefaultQuartzScheduler_Worker-19) BaseAsyncTask::LogEndTaskFailure: Task 720b4d92-1425-478c-8351-4ff827b8f728 (Parent Command AddDisk, Parameters Type org.ovirt.engine.core.common.asynctasks.AsyncTaskParameters) ended with failure:^M -- Result: cleanSuccess^M -- Message: VDSGenericException: VDSErrorException: Failed to HSMGetAllTasksStatusesVDS, error = [Errno 2] No such file or directory: '/rhev/data-center/06930787-a091-49a3-8217-1418c5a9881e/967aec77-46d5-418b-8979-d0a86389a77b/images/7726b997-7e58-45f8-a5a6-9cb9a689a45a', code = 100,^M -- Exception: VDSGenericException: VDSErrorException: Failed to HSMGetAllTasksStatusesVDS, error = [Errno 2] No such file or directory: '/rhev/data-center/06930787-a091-49a3-8217-1418c5a9881e/967aec77-46d5-418b-8979-d0a86389a77b/images/7726b997-7e58-45f8-a5a6-9cb9a689a45a', code = 100
Certainly, if I check /rhev/data-center/06930787-a091-49a3-8217-1418c5a9881e/ on the SPM server, there is no 967aec77-46d5-418b-8979-d0a86389a77b subdirectory. The only elements I have are NFS mounts.
There appear to be no errors in the SPM vdsm.log for this disk
============= When I tried to import the vm (the one that I then tried to create from scratch above), I had the following errors in the SPM vdsm log: Thread-2220::DEBUG::2014-05-20 08:35:05,255::task::595::TaskManager.Task::(_updateState) Task=`15bc07b5-201f-4bba-bf5f-f79eb92c6a61`::moving from state init -> state preparing Thread-2220::INFO::2014-05-20 08:35:05,255::logUtils::44::dispatcher::(wrapper) Run and protect: deleteImage(sdUUID='615647e2-1f60-47e1-8e55-be9f7ead6f15', spUUID='06930787-a091-49a3-8217-1418c5a9881e', imgUUID='80ed133c-fd72-4d35-aae5-e1313be3cf23', postZero='false', force='false') Thread-2220::DEBUG::2014-05-20 08:35:05,255::resourceManager::198::ResourceManager.Request::(__init__) ResName=`Storage.80ed133c-fd72-4d35-aae5-e1313be3cf23`ReqID=`499de454-c563-4156-a3ed-13b7eb9defa6`::Request was made in '/usr/share/vdsm/storage/hsm.py' line '1496' at 'deleteImage' Thread-2220::DEBUG::2014-05-20 08:35:05,255::resourceManager::542::ResourceManager::(registerResource) Trying to register resource 'Storage.80ed133c-fd72-4d35-aae5-e1313be3cf23' for lock type 'exclusive' Thread-2220::DEBUG::2014-05-20 08:35:05,255::resourceManager::601::ResourceManager::(registerResource) Resource 'Storage.80ed133c-fd72-4d35-aae5-e1313be3cf23' is free. Now locking as 'exclusive' (1 active user) Thread-2220::DEBUG::2014-05-20 08:35:05,256::resourceManager::238::ResourceManager.Request::(grant) ResName=`Storage.80ed133c-fd72-4d35-aae5-e1313be3cf23`ReqID=`499de454-c563-4156-a3ed-13b7eb9defa6`::Granted request Thread-2220::DEBUG::2014-05-20 08:35:05,256::task::827::TaskManager.Task::(resourceAcquired) Task=`15bc07b5-201f-4bba-bf5f-f79eb92c6a61`::_resourcesAcquired: Storage.80ed133c-fd72-4d35-aae5-e1313be3cf23 (exclusive) Thread-2220::DEBUG::2014-05-20 08:35:05,256::task::990::TaskManager.Task::(_decref) Task=`15bc07b5-201f-4bba-bf5f-f79eb92c6a61`::ref 1 aborting False Thread-2220::DEBUG::2014-05-20 08:35:05,256::resourceManager::198::ResourceManager.Request::(__init__) ResName=`Storage.615647e2-1f60-47e1-8e55-be9f7ead6f15`ReqID=`73f79517-f13a-4e5b-999a-6f1994d2818a`::Request was made in '/usr/share/vdsm/storage/hsm.py' line '1497' at 'deleteImage' Thread-2220::DEBUG::2014-05-20 08:35:05,256::resourceManager::542::ResourceManager::(registerResource) Trying to register resource 'Storage.615647e2-1f60-47e1-8e55-be9f7ead6f15' for lock type 'shared' Thread-2220::DEBUG::2014-05-20 08:35:05,257::resourceManager::601::ResourceManager::(registerResource) Resource 'Storage.615647e2-1f60-47e1-8e55-be9f7ead6f15' is free. Now locking as 'shared' (1 active user) Thread-2220::DEBUG::2014-05-20 08:35:05,257::resourceManager::238::ResourceManager.Request::(grant) ResName=`Storage.615647e2-1f60-47e1-8e55-be9f7ead6f15`ReqID=`73f79517-f13a-4e5b-999a-6f1994d2818a`::Granted request Thread-2220::DEBUG::2014-05-20 08:35:05,257::task::827::TaskManager.Task::(resourceAcquired) Task=`15bc07b5-201f-4bba-bf5f-f79eb92c6a61`::_resourcesAcquired: Storage.615647e2-1f60-47e1-8e55-be9f7ead6f15 (shared) Thread-2220::DEBUG::2014-05-20 08:35:05,257::task::990::TaskManager.Task::(_decref) Task=`15bc07b5-201f-4bba-bf5f-f79eb92c6a61`::ref 1 aborting False Thread-2220::ERROR::2014-05-20 08:35:05,266::hsm::1502::Storage.HSM::(deleteImage) Empty or not found image 80ed133c-fd72-4d35-aae5-e1313be3cf23 in SD 615647e2-1f60-47e1-8e55-be9f7ead6f15. {'1f41529a-e02e-4cd8-987c-b1ea4fcba2be': ImgsPar(imgs=('290f5cdf-b5d7-462b-958d-d41458a26bf6',), parent=None), '1748a8f0-8668-4f21-9b26-d2e3b180e35b': ImgsPar(imgs=('67fd552b-8b3d-4117-82d2-e801bb600992',), parent=None)} Thread-2220::ERROR::2014-05-20 08:35:05,266::task::866::TaskManager.Task::(_setError) Task=`15bc07b5-201f-4bba-bf5f-f79eb92c6a61`::Unexpected error
When I installed/setup ovirt-engine, I did chose NFS as the file system.
I am clearly doing something weird

------=_Part_8459191_1741827386.1400593098736 Content-Type: text/plain; charset=utf-8 Content-Transfer-Encoding: 7bit This sounds a lot like a https://bugzilla.redhat.com/show_bug.cgi?id=1069772 which we solved for 3.4.1. I wonder if we should backport it to 3.3.z ----- Original Message -----
A slight update to this.
I did find that the gluster mount points re appear under SPM /rhev/data-center if I detach and then deactivate the gluster storages under the data storages.
As this means shutting down any VMs that are running from that storage, is there any known way of getting the gluster points to reattach without dectivcate/reactivate a storage in the gui?
Cheers
Peter
On 20 May 2014 09:37, Peter Harris < doilooksensible@gmail.com > wrote:
I am trying to migrate my VMs from my old host running ovirt 3.3.4 to a new setup running 3.4.1. My basic set up is:
OLD
vmhost1 - ovirt 3.3.4 - NFS storages
NEW
ovirtmgr - ovirt 3.4.1 (virt only setup) - gluster storage domains
vmhost2 - cluster1
vmhost3 - cluster2
vmhost4 - cluster2
vmhost5 - cluster3
vmhost6 - cluster3
My gluster volumes are created via gluster command line and I have
All hosts are running scientific Linux 6.5, and the intention is to migrate vmhost1 to new environment cluster1.
I have an NFS export storage domain which I am using to migrate VMs from vmhost1.
Volume Name: vol-vminf
Type: Distributed-Replicate
Volume ID: b0b456bb-76e9-42e7-bb95-3415db79d631
Status: Started
Number of Bricks: 2 x 2 = 4
Transport-type: tcp
Bricks:
Brick1: vmhost3:/storage/inf/br-inf
Brick2: vmhost4:/storage/inf/br-inf
Brick3: vmhost5:/storage/inf/br-inf
Brick4: vmhost6:/storage/inf/br-inf
Options Reconfigured:
storage.owner-gid: 36
storage.owner-uid: 36
server.allow-insecure: on
Volume Name: vol-vmimages
Type: Distribute
Volume ID: 91e2cf8b-2662-4c26-b937-84b8f5b62e2b
Status: Started
Number of Bricks: 4
Transport-type: tcp
Bricks:
Brick1: vmhost3:/storage/vmimages/br-vmimages
Brick2: vmhost3:/storage/vmimages/br-vmimages
Brick3: vmhost3:/storage/vmimages/br-vmimages
Brick4: vmhost3:/storage/vmimages/br-vmimages
Options Reconfigured:
storage.owner-gid: 36
storage.owner-uid: 36
server.allow-insecure: on
I have had many varying fail results which I have tried to match up with threads here, and I am now a bit stuck and would appreciate any help.
I quite regularly cannot import vms, and I now cannot create a new disk for a new VM (no import). The error always seems to boild down to the following error on engine.log (the following specifically for the create of a new disk image on a gluster storage domain):
2014-05-20 08:51:21,136 INFO [org.ovirt.engine.core.dal.dbbroker.auditloghandling.AuditLogDirector] (ajp--127.0.0.1-8702-9) [4637af09] Correlation ID: 2b0b55ab, Job ID: 1a583643-e28a-4f09-a39d-46e4fc6d20b8, Call Stack: null, Custom Event ID: -1, Message: Add-Disk operation of rhel-7_Disk1 was initiated on VM rhel-7 by peter.harris.
2014-05-20 08:51:21,137 INFO [org.ovirt.engine.core.bll.SPMAsyncTask] (ajp--127.0.0.1-8702-9) [4637af09] BaseAsyncTask::startPollingTask: Starting to poll task 720b4d92-1425-478c-8351-4ff827b8f728.
2014-05-20 08:51:28,077 INFO [org.ovirt.engine.core.bll.AsyncTaskManager] (DefaultQuartzScheduler_Worker-19) Polling and updating Async Tasks: 1 tasks, 1 tasks to poll now
2014-05-20 08:51:28,084 ERROR [org.ovirt.engine.core.vdsbroker.vdsbroker.HSMGetAllTasksStatusesVDSCommand] (DefaultQuartzScheduler_Worker-19) Failed in HSMGetAllTasksStatusesVDS method
2014-05-20 08:51:28,085 INFO [org.ovirt.engine.core.bll.SPMAsyncTask] (DefaultQuartzScheduler_Worker-19) SPMAsyncTask::PollTask: Polling task 720b4d92-1425-478c-8351-4ff827b8f728 (Parent Command AddDisk, Parameters Type org.ovirt.engine.core.common.asynctasks.AsyncTaskParameters) returned status finished, result 'cleanSuccess'.
2014-05-20 08:51:28,104 ERROR [org.ovirt.engine.core.bll.SPMAsyncTask] (DefaultQuartzScheduler_Worker-19) BaseAsyncTask::LogEndTaskFailure: Task 720b4d92-1425-478c-8351-4ff827b8f728 (Parent Command AddDisk, Parameters Type org.ovirt.engine.core.common.asynctasks.AsyncTaskParameters) ended with failure:^M
-- Result: cleanSuccess^M
-- Message: VDSGenericException: VDSErrorException: Failed to HSMGetAllTasksStatusesVDS, error = [Errno 2] No such file or directory: '/rhev/data-center/06930787-a091-49a3-8217-1418c5a9881e/967aec77-46d5-418b-8979-d0a86389a77b/images/7726b997-7e58-45f8-a5a6-9cb9a689a45a', code = 100,^M
-- Exception: VDSGenericException: VDSErrorException: Failed to HSMGetAllTasksStatusesVDS, error = [Errno 2] No such file or directory: '/rhev/data-center/06930787-a091-49a3-8217-1418c5a9881e/967aec77-46d5-418b-8979-d0a86389a77b/images/7726b997-7e58-45f8-a5a6-9cb9a689a45a', code = 100
Certainly, if I check /rhev/data-center/06930787-a091-49a3-8217-1418c5a9881e/ on the SPM server, there is no 967aec77-46d5-418b-8979-d0a86389a77b subdirectory. The only elements I have are NFS mounts.
There appear to be no errors in the SPM vdsm.log for this disk
=============
When I tried to import the vm (the one that I then tried to create from scratch above), I had the following errors in the SPM vdsm log:
Thread-2220::DEBUG::2014-05-20 08:35:05,255::task::595::TaskManager.Task::(_updateState) Task=`15bc07b5-201f-4bba-bf5f-f79eb92c6a61`::moving from state init -> state preparing
Thread-2220::INFO::2014-05-20 08:35:05,255::logUtils::44::dispatcher::(wrapper) Run and protect: deleteImage(sdUUID='615647e2-1f60-47e1-8e55-be9f7ead6f15', spUUID='06930787-a091-49a3-8217-1418c5a9881e', imgUUID='80ed133c-fd72-4d35-aae5-e1313be3cf23', postZero='false', force='false')
Thread-2220::DEBUG::2014-05-20 08:35:05,255::resourceManager::198::ResourceManager.Request::(__init__) ResName=`Storage.80ed133c-fd72-4d35-aae5-e1313be3cf23`ReqID=`499de454-c563-4156-a3ed-13b7eb9defa6`::Request was made in '/usr/share/vdsm/storage/hsm.py' line '1496' at 'deleteImage'
Thread-2220::DEBUG::2014-05-20 08:35:05,255::resourceManager::542::ResourceManager::(registerResource) Trying to register resource 'Storage.80ed133c-fd72-4d35-aae5-e1313be3cf23' for lock type 'exclusive'
Thread-2220::DEBUG::2014-05-20 08:35:05,255::resourceManager::601::ResourceManager::(registerResource) Resource 'Storage.80ed133c-fd72-4d35-aae5-e1313be3cf23' is free. Now locking as 'exclusive' (1 active user)
Thread-2220::DEBUG::2014-05-20 08:35:05,256::resourceManager::238::ResourceManager.Request::(grant) ResName=`Storage.80ed133c-fd72-4d35-aae5-e1313be3cf23`ReqID=`499de454-c563-4156-a3ed-13b7eb9defa6`::Granted request
Thread-2220::DEBUG::2014-05-20 08:35:05,256::task::827::TaskManager.Task::(resourceAcquired) Task=`15bc07b5-201f-4bba-bf5f-f79eb92c6a61`::_resourcesAcquired: Storage.80ed133c-fd72-4d35-aae5-e1313be3cf23 (exclusive)
Thread-2220::DEBUG::2014-05-20 08:35:05,256::task::990::TaskManager.Task::(_decref) Task=`15bc07b5-201f-4bba-bf5f-f79eb92c6a61`::ref 1 aborting False
Thread-2220::DEBUG::2014-05-20 08:35:05,256::resourceManager::198::ResourceManager.Request::(__init__) ResName=`Storage.615647e2-1f60-47e1-8e55-be9f7ead6f15`ReqID=`73f79517-f13a-4e5b-999a-6f1994d2818a`::Request was made in '/usr/share/vdsm/storage/hsm.py' line '1497' at 'deleteImage'
Thread-2220::DEBUG::2014-05-20 08:35:05,256::resourceManager::542::ResourceManager::(registerResource) Trying to register resource 'Storage.615647e2-1f60-47e1-8e55-be9f7ead6f15' for lock type 'shared'
Thread-2220::DEBUG::2014-05-20 08:35:05,257::resourceManager::601::ResourceManager::(registerResource) Resource 'Storage.615647e2-1f60-47e1-8e55-be9f7ead6f15' is free. Now locking as 'shared' (1 active user)
Thread-2220::DEBUG::2014-05-20 08:35:05,257::resourceManager::238::ResourceManager.Request::(grant) ResName=`Storage.615647e2-1f60-47e1-8e55-be9f7ead6f15`ReqID=`73f79517-f13a-4e5b-999a-6f1994d2818a`::Granted request
Thread-2220::DEBUG::2014-05-20 08:35:05,257::task::827::TaskManager.Task::(resourceAcquired) Task=`15bc07b5-201f-4bba-bf5f-f79eb92c6a61`::_resourcesAcquired: Storage.615647e2-1f60-47e1-8e55-be9f7ead6f15 (shared)
Thread-2220::DEBUG::2014-05-20 08:35:05,257::task::990::TaskManager.Task::(_decref) Task=`15bc07b5-201f-4bba-bf5f-f79eb92c6a61`::ref 1 aborting False
Thread-2220::ERROR::2014-05-20 08:35:05,266::hsm::1502::Storage.HSM::(deleteImage) Empty or not found image 80ed133c-fd72-4d35-aae5-e1313be3cf23 in SD 615647e2-1f60-47e1-8e55-be9f7ead6f15. {'1f41529a-e02e-4cd8-987c-b1ea4fcba2be': ImgsPar(imgs=('290f5cdf-b5d7-462b-958d-d41458a26bf6',), parent=None), '1748a8f0-8668-4f21-9b26-d2e3b180e35b': ImgsPar(imgs=('67fd552b-8b3d-4117-82d2-e801bb600992',), parent=None)}
Thread-2220::ERROR::2014-05-20 08:35:05,266::task::866::TaskManager.Task::(_setError) Task=`15bc07b5-201f-4bba-bf5f-f79eb92c6a61`::Unexpected error
When I installed/setup ovirt-engine, I did chose NFS as the file system.
I am clearly doing something weird
_______________________________________________ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users
<<a href=3D"mailto:doilooksensible@gmail.com" target=3D"_blank">doilook= sensible@gmail.com</a>></span> wrote:<br> <blockquote class=3D"gmail_quote" style=3D"margin:0 0 0 .8ex;border-left:1p= x #ccc solid;padding-left:1ex"><div dir=3D"ltr"><div><div><div><div><div><d= iv><div><div><div><div><div><div><div><div>I am trying to migrate my VMs fr= om my old host running ovirt 3.3.4 to a new setup running 3.4.1. My basic s= et up is:<br> <br></div>OLD<br> vmhost1 - ovirt 3.3.4 - NFS storages<br><div><br></div></div>NEW<br></div><=
------=_Part_8459191_1741827386.1400593098736 Content-Type: text/html; charset=utf-8 Content-Transfer-Encoding: quoted-printable <html><body><div style=3D"font-family: times new roman,new york,times,serif= ; font-size: 12pt; color: #000000"><div>This sounds a lot like a <a hr= ef=3D"https://bugzilla.redhat.com/show_bug.cgi?id=3D1069772" data-mce-href= =3D"https://bugzilla.redhat.com/show_bug.cgi?id=3D1069772">https://bugzilla= .redhat.com/show_bug.cgi?id=3D1069772</a> which we solved for 3.4.1.</= div><div>I wonder if we should backport it to 3.3.z = </div><div><br></div><hr id=3D"zwchr"><blockquote style=3D"border-left:2px = solid #1010FF;margin-left:5px;padding-left:5px;color:#000;font-weight:norma= l;font-style:normal;text-decoration:none;font-family:Helvetica,Arial,sans-s= erif;font-size:12pt;"><div dir=3D"ltr"><div><div><div><div>A slight update = to this.<br><div><br></div></div>I did find that the gluster mount points r= e appear under SPM /rhev/data-center if I detach and then deactivate the gl= uster storages under the data storages.<br> <br></div>As this means shutting down any VMs that are running from that st= orage, is there any known way of getting the gluster points to reattach wit= hout dectivcate/reactivate a storage in the gui?<br><div><br></div></div>Ch= eers<br> <br></div>Peter<br></div><div class=3D"gmail_extra"><br><div><br></div><div= class=3D"gmail_quote">On 20 May 2014 09:37, Peter Harris <span dir=3D"ltr"= div>ovirtmgr - ovirt 3.4.1 (virt only setup) - gluster storage domains<br><= /div>vmhost2 - cluster1<br></div>vmhost3 - cluster2<br></div>vmhost4 - clus= ter2<br> </div>vmhost5 - cluster3<br></div>vmhost6 - cluster3<br><div><br></div></di= v><div>My gluster volumes are created via gluster command line and I have <= br></div><div><br></div>All hosts are running scientific Linux 6.5, and the= intention is to migrate vmhost1 to new environment cluster1.<br> <br></div>I have an NFS export storage domain which I am using to migrate V= Ms from vmhost1.<br>Volume Name: vol-vminf<br>Type: Distributed-Replicate<b= r>Volume ID: b0b456bb-76e9-42e7-bb95-3415db79d631<br>Status: Started<br> Number of Bricks: 2 x 2 =3D 4<br>Transport-type: tcp<br>Bricks:<br>Brick1: = vmhost3:/storage/inf/br-inf<br>Brick2: vmhost4:/storage/inf/br-inf<br>Brick= 3: vmhost5:/storage/inf/br-inf<br>Brick4: vmhost6:/storage/inf/br-inf<br> Options Reconfigured:<br>storage.owner-gid: 36<br>storage.owner-uid: 36<br>= server.allow-insecure: on<br><div><br></div>Volume Name: vol-vmimages<br>Ty= pe: Distribute<br>Volume ID: 91e2cf8b-2662-4c26-b937-84b8f5b62e2b<br>Status= : Started<br> Number of Bricks: 4<br>Transport-type: tcp<br>Bricks:<br>Brick1: vmhost3:/s= torage/vmimages/br-vmimages<br>Brick2: vmhost3:/storage/vmimages/br-vmimage= s<br>Brick3: vmhost3:/storage/vmimages/br-vmimages<br>Brick4: vmhost3:/stor= age/vmimages/br-vmimages<br> Options Reconfigured:<br>storage.owner-gid: 36<br>storage.owner-uid: 36<br>= server.allow-insecure: on<br><div><br></div></div>I have had many varying f= ail results which I have tried to match up with threads here, and I am now = a bit stuck and would appreciate any help.<br> <br></div>I quite regularly cannot import vms, and I now cannot create a ne= w disk for a new VM (no import). The error always seems to boild down to th= e following error on engine.log (the following specifically for the create = of a new disk image on a gluster storage domain):<br> <br>2014-05-20 08:51:21,136 INFO [org.ovirt.engine.core.dal.dbbroker.= auditloghandling.AuditLogDirector] (ajp--127.0.0.1-8702-9) [4637af09] Corre= lation ID: 2b0b55ab, Job ID: 1a583643-e28a-4f09-a39d-46e4fc6d20b8, Call Sta= ck: null, Custom Event ID: -1, Message: Add-Disk operation of rhel-7_Disk1 = was initiated on VM rhel-7 by peter.harris.<br> 2014-05-20 08:51:21,137 INFO [org.ovirt.engine.core.bll.SPMAsyncTask]= (ajp--127.0.0.1-8702-9) [4637af09] BaseAsyncTask::startPollingTask: Starti= ng to poll task 720b4d92-1425-478c-8351-4ff827b8f728.<br>2014-05-20 08:51:2= 8,077 INFO [org.ovirt.engine.core.bll.AsyncTaskManager] (DefaultQuart= zScheduler_Worker-19) Polling and updating Async Tasks: 1 tasks, 1 tasks to= poll now<br> 2014-05-20 08:51:28,084 ERROR [org.ovirt.engine.core.vdsbroker.vdsbroker.HS= MGetAllTasksStatusesVDSCommand] (DefaultQuartzScheduler_Worker-19) Failed i= n HSMGetAllTasksStatusesVDS method<br>2014-05-20 08:51:28,085 INFO [o= rg.ovirt.engine.core.bll.SPMAsyncTask] (DefaultQuartzScheduler_Worker-19) S= PMAsyncTask::PollTask: Polling task 720b4d92-1425-478c-8351-4ff827b8f728 (P= arent Command AddDisk, Parameters Type org.ovirt.engine.core.common.asyncta= sks.AsyncTaskParameters) returned status finished, result 'cleanSuccess'.<b= r> 2014-05-20 08:51:28,104 ERROR [org.ovirt.engine.core.bll.SPMAsyncTask] (Def= aultQuartzScheduler_Worker-19) BaseAsyncTask::LogEndTaskFailure: Task 720b4= d92-1425-478c-8351-4ff827b8f728 (Parent Command AddDisk, Parameters Type or= g.ovirt.engine.core.common.asynctasks.AsyncTaskParameters) ended with failu= re:^M<br> -- Result: cleanSuccess^M<br>-- Message: VDSGenericException: VDSErrorExcep= tion: Failed to HSMGetAllTasksStatusesVDS, error =3D [Errno 2] No such file= or directory: '/rhev/data-center/06930787-a091-49a3-8217-1418c5a9881e/967a= ec77-46d5-418b-8979-d0a86389a77b/images/7726b997-7e58-45f8-a5a6-9cb9a689a45= a', code =3D 100,^M<br> -- Exception: VDSGenericException: VDSErrorException: Failed to HSMGetAllTa= sksStatusesVDS, error =3D [Errno 2] No such file or directory: '/rhev/data-= center/06930787-a091-49a3-8217-1418c5a9881e/967aec77-46d5-418b-8979-d0a8638= 9a77b/images/7726b997-7e58-45f8-a5a6-9cb9a689a45a', code =3D 100<br> <br></div>Certainly, if I check /rhev/data-center/06930787-a091-49a3-8217-1= 418c5a9881e/ on the SPM server, there is no 967aec77-46d5-418b-8979-d0a8638= 9a77b subdirectory. The only elements I have are NFS mounts.<br><div><br></= div></div> <div>There appear to be no errors in the SPM vdsm.log for this disk<br></di= v><div><br>=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D=3D<br></div><div>When I tri= ed to import the vm (the one that I then tried to create from scratch above= ), I had the following errors in the SPM vdsm log:<br> Thread-2220::DEBUG::2014-05-20 08:35:05,255::task::595::TaskManager.Task::(= _updateState) Task=3D`15bc07b5-201f-4bba-bf5f-f79eb92c6a61`::moving from st= ate init -> state preparing<br>Thread-2220::INFO::2014-05-20 08:35:05,25= 5::logUtils::44::dispatcher::(wrapper) Run and protect: deleteImage(sdUUID= =3D'615647e2-1f60-47e1-8e55-be9f7ead6f15', spUUID=3D'06930787-a091-49a3-821= 7-1418c5a9881e', imgUUID=3D'80ed133c-fd72-4d35-aae5-e1313be3cf23', postZero= =3D'false', force=3D'false')<br> Thread-2220::DEBUG::2014-05-20 08:35:05,255::resourceManager::198::Resource= Manager.Request::(__init__) ResName=3D`Storage.80ed133c-fd72-4d35-aae5-e131= 3be3cf23`ReqID=3D`499de454-c563-4156-a3ed-13b7eb9defa6`::Request was made i= n '/usr/share/vdsm/storage/hsm.py' line '1496' at 'deleteImage'<br> Thread-2220::DEBUG::2014-05-20 08:35:05,255::resourceManager::542::Resource= Manager::(registerResource) Trying to register resource 'Storage.80ed133c-f= d72-4d35-aae5-e1313be3cf23' for lock type 'exclusive'<br> Thread-2220::DEBUG::2014-05-20 08:35:05,255::resourceManager::601::Resource= Manager::(registerResource) Resource 'Storage.80ed133c-fd72-4d35-aae5-e1313= be3cf23' is free. Now locking as 'exclusive' (1 active user)<br> Thread-2220::DEBUG::2014-05-20 08:35:05,256::resourceManager::238::Resource= Manager.Request::(grant) ResName=3D`Storage.80ed133c-fd72-4d35-aae5-e1313be= 3cf23`ReqID=3D`499de454-c563-4156-a3ed-13b7eb9defa6`::Granted request<br>Th= read-2220::DEBUG::2014-05-20 08:35:05,256::task::827::TaskManager.Task::(re= sourceAcquired) Task=3D`15bc07b5-201f-4bba-bf5f-f79eb92c6a61`::_resourcesAc= quired: Storage.80ed133c-fd72-4d35-aae5-e1313be3cf23 (exclusive)<br> Thread-2220::DEBUG::2014-05-20 08:35:05,256::task::990::TaskManager.Task::(= _decref) Task=3D`15bc07b5-201f-4bba-bf5f-f79eb92c6a61`::ref 1 aborting Fals= e<br>Thread-2220::DEBUG::2014-05-20 08:35:05,256::resourceManager::198::Res= ourceManager.Request::(__init__) ResName=3D`Storage.615647e2-1f60-47e1-8e55= -be9f7ead6f15`ReqID=3D`73f79517-f13a-4e5b-999a-6f1994d2818a`::Request was m= ade in '/usr/share/vdsm/storage/hsm.py' line '1497' at 'deleteImage'<br> Thread-2220::DEBUG::2014-05-20 08:35:05,256::resourceManager::542::Resource= Manager::(registerResource) Trying to register resource 'Storage.615647e2-1= f60-47e1-8e55-be9f7ead6f15' for lock type 'shared'<br> Thread-2220::DEBUG::2014-05-20 08:35:05,257::resourceManager::601::Resource= Manager::(registerResource) Resource 'Storage.615647e2-1f60-47e1-8e55-be9f7= ead6f15' is free. Now locking as 'shared' (1 active user)<br> Thread-2220::DEBUG::2014-05-20 08:35:05,257::resourceManager::238::Resource= Manager.Request::(grant) ResName=3D`Storage.615647e2-1f60-47e1-8e55-be9f7ea= d6f15`ReqID=3D`73f79517-f13a-4e5b-999a-6f1994d2818a`::Granted request<br>Th= read-2220::DEBUG::2014-05-20 08:35:05,257::task::827::TaskManager.Task::(re= sourceAcquired) Task=3D`15bc07b5-201f-4bba-bf5f-f79eb92c6a61`::_resourcesAc= quired: Storage.615647e2-1f60-47e1-8e55-be9f7ead6f15 (shared)<br> Thread-2220::DEBUG::2014-05-20 08:35:05,257::task::990::TaskManager.Task::(= _decref) Task=3D`15bc07b5-201f-4bba-bf5f-f79eb92c6a61`::ref 1 aborting Fals= e<br>Thread-2220::ERROR::2014-05-20 08:35:05,266::hsm::1502::Storage.HSM::(= deleteImage) Empty or not found image 80ed133c-fd72-4d35-aae5-e1313be3cf23 = in SD 615647e2-1f60-47e1-8e55-be9f7ead6f15. {'1f41529a-e02e-4cd8-987c-b1ea4= fcba2be': ImgsPar(imgs=3D('290f5cdf-b5d7-462b-958d-d41458a26bf6',), parent= =3DNone), '1748a8f0-8668-4f21-9b26-d2e3b180e35b': ImgsPar(imgs=3D('67fd552b= -8b3d-4117-82d2-e801bb600992',), parent=3DNone)}<br> Thread-2220::ERROR::2014-05-20 08:35:05,266::task::866::TaskManager.Task::(= _setError) Task=3D`15bc07b5-201f-4bba-bf5f-f79eb92c6a61`::Unexpected error<= br><div><br></div><br><div><br></div></div><div><br></div>When I installed/= setup ovirt-engine, I did chose NFS as the file system.<br> <br></div>I am clearly doing something weird<br><div><div><div><br><div><di= v><br><div><br><div><div><div><div><div><div><div><br><div><br></div></div>= </div></div></div></div></div></div></div></div></div></div></div></div></d= iv> </blockquote></div><br></div> <br>_______________________________________________<br>Users mailing list<b= r>Users@ovirt.org<br>http://lists.ovirt.org/mailman/listinfo/users<br></blo= ckquote><div><br></div></div></body></html> ------=_Part_8459191_1741827386.1400593098736--

afaik there will be no more 3.3.z (which would be z=6) 3.3.5 was the last 3.3. release imho. Am 20.05.2014 15:38, schrieb Allon Mureinik:
This sounds a lot like a https://bugzilla.redhat.com/show_bug.cgi?id=1069772 which we solved for 3.4.1. I wonder if we should backport it to 3.3.z
-- Mit freundlichen Grüßen / Regards Sven Kieske Systemadministrator Mittwald CM Service GmbH & Co. KG Königsberger Straße 6 32339 Espelkamp T: +49-5772-293-100 F: +49-5772-293-333 https://www.mittwald.de Geschäftsführer: Robert Meyer St.Nr.: 331/5721/1033, USt-IdNr.: DE814773217, HRA 6640, AG Bad Oeynhausen Komplementärin: Robert Meyer Verwaltungs GmbH, HRB 13260, AG Bad Oeynhausen
participants (3)
-
Allon Mureinik
-
Peter Harris
-
Sven Kieske