[ovirt-users] Manually copying VM disks from FC data domain

Adam Litke alitke at redhat.com
Fri Jun 2 13:09:50 UTC 2017


hmm, strange.  So all three of those missing volumes have associated LVs.
Can you try to activate them and see if you can read from them?

  lvchange --config 'global {use_lvmetad=0}' -ay 0fcd2921-8a55-4ff7-9cba-
dca6ffb7b67e/3b089aed-b3e1-4423-8585-e65752d19ffe
  dd
if=/dev/0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e/3b089aed-b3e1-4423-8585-e65752d19ffe
of=/dev/null bs=1M count=1

If this works, then one way to recover the data is to use the UI to create
new disks of the same size as the old ones.  Then, activate the LVs
associated with the old volumes and the new ones.  Then use dd (or qemu-img
convert) to copy from old to new.  Then attach the new disks to your VM.

On Thu, Jun 1, 2017 at 6:44 PM, Николаев Алексей <
alexeynikolaev.post at yandex.ru> wrote:

> Thx for your help!
>
> 01.06.2017, 16:46, "Adam Litke" <alitke at redhat.com>:
>
> When you say "not visible in oVirt" you mean that you do not see them in
> the UI?
>
>
> Yes, I can see some VM with prefix "external-" and without disks.
>
>
> Do you know the specific uuids for the missing volumes?  You could use lvm
> to check if the LVs are visible to the host.
>
> lvs --config 'global {use_lvmetad=0}' -o +tags
>
> For each LV, the tag beginning with IU_ indicates the image the volume
> belongs to.
>
>
>   LV                                   VG
> Attr       LSize    Pool Origin Data%  Meta%  Move Log Cpy%Sync Convert LV
> Tags
>
>   03917876-0e28-4457-bf44-53c7ea2b4d12 0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e
> -wi-------  117,00g
> IU_3edd1f60-fd43-4e12-9615-b12bcd1a17ab,MD_6,PU_00000000-
> 0000-0000-0000-000000000000
>   309a325a-6f13-4a24-b204-e825ddbf0e41 0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e
> -wi-ao----  300,00g
> IU_9cbde83c-7d1d-4b78-bc7a-6d540584c888,MD_17,PU_00000000-0000-0000-0000-
> 000000000000
>   3b089aed-b3e1-4423-8585-e65752d19ffe 0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e
> -wi-------   17,00g
> IU_986849a8-04ea-4b7d-a29f-f023da9020b3,MD_5,PU_00000000-
> 0000-0000-0000-000000000000
>   48b6f4f5-9eeb-4cf3-922a-c9c0c2239bee 0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e
> -wi-------  128,00m
> IU_caed7169-0c90-467f-86d3-a82148b1f0af,MD_10,PU_00000000-0000-0000-0000-
> 000000000000
>   7700e1e1-351b-49b5-8681-e121bbf67177 0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e
> -wi-------   83,00g
> IU_9cf3a067-2996-4ff2-b481-b13f7cc73c33,MD_7,PU_00000000-
> 0000-0000-0000-000000000000
>   91a7aa8c-4e13-477c-8703-d8224b85bc84 0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e
> -wi-ao---- 1000,00g
> IU_fd2ecb98-a8c5-4aff-a6d3-7ac3087ab994,MD_15,PU_00000000-0000-0000-0000-
> 000000000000
>   a3f29de7-c6b9-410e-b635-9b3016da7ba2 0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e
> -wi-------    7,50g
> IU_71f18680-1b93-4a2a-9bd1-95baeccf2d89,MD_11,PU_00000000-0000-0000-0000-
> 000000000000
>   ba26f8b7-b807-4a6c-a840-6a83e8ec526e 0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e
> -wi-------    8,00g
> IU_e204e228-fa35-4457-a31d-3b7964156538,MD_8,PU_00000000-
> 0000-0000-0000-000000000000
>   ba29b8fd-3618-42d6-a70f-17d883fde0ed 0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e
> -wi-------   50,00g
> IU_c46d5cbb-1802-4086-a31a-5b3f0b874454,MD_16,PU_00000000-0000-0000-0000-
> 000000000000
>   ba92b801-3619-4753-9e06-3e2028a408cb 0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e
> -wi-------  128,00m
> IU_0d0983f0-60d7-4ce0-bf1b-12cfc456acd8,MD_9,PU_00000000-
> 0000-0000-0000-000000000000
>   c0bcf836-c648-4421-9340-febcc0e0abfe 0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e
> -wi-------   50,00g
> IU_4fe1b047-60b2-4bc3-b73b-2fad8a81cc02,MD_13,PU_00000000-0000-0000-0000-
> 000000000000
>   c1d08688-aad1-478e-82a3-e5b5fde85706 0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e
> -wi-ao---- 1000,00g
> IU_f58cd721-d8b8-4794-b059-8789a9fecf62,MD_18,PU_00000000-0000-0000-0000-
> 000000000000
>   c7a4782f-cde6-40db-a625-810fd2856dfa 0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e
> -wi-------  300,00g
> IU_8917cb1a-8bd4-4386-a95a-0189a04866ad,MD_19,PU_00000000-0000-0000-0000-
> 000000000000
>   cb36e25a-b9b2-4f54-9b03-a10837bc26ab 0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e
> -wi-------   50,00g
> IU_e8720901-0383-4266-98a7-fb5b9fb27e52,MD_14,PU_00000000-0000-0000-0000-
> 000000000000
>   f60ccaa4-663b-4a39-8ad0-2ed3fb208cb0 0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e
> -wi-------   50,00g
> IU_f0d441bb-2bd2-4523-ab59-beef544727b5,MD_12,PU_00000000-0000-0000-0000-
> 000000000000
>   fd8822ee-4fc9-49ba-9760-87a85d56bf91 0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e
> -wi-------  413,00g
> IU_7be49698-f3a5-4995-b411-f0490a819950,MD_4,PU_00000000-
> 0000-0000-0000-000000000000
>   ids                                  0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e
> -wi-ao----  128,00m
>
>
>   inbox                                0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e
> -wi-a-----  128,00m
>
>
>   leases                               0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e
> -wi-a-----    2,00g
>
>
>   master                               0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e
> -wi-a-----    1,00g
>
>
>   metadata                             0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e
> -wi-a-----  512,00m
>
>
>   outbox                               0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e
> -wi-a-----  128,00m
>
>
>   root                                 centos_node13-05
> -wi-ao----  101,49g
>
>
>   swap                                 centos_node13-05
> -wi-ao----    9,77g
>
>
>   lv-filestore                         vg-filestore
> -wi-a----- 1000,00g
>
>
>   lv-opt                               vg-opt
> -wi-a-----  300,00g
>
>
>   lv-postgres                          vg-postgres
> -wi-a----- 1000,00g
>
>
>
> You could also try to use the vdsm commands:
>
>
> I can only use vdsClient at 3.5....
>
>
>
> # Get the storage pool id
> sudo vdsm-client Host getConnectedStoragePools
>
>
> dsClient -s 0 getConnectedStoragePoolsList
>
> 00000002-0002-0002-0002-00000000009b
>
>
>
> # Get a list of storagedomain IDs
> sudo vdsm-client Host getStorageDomains # gives you a list of
> storagedomainIDs
>
>
> vdsClient -s 0 getStorageDomainsList
>
> 0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e
>
>
>
> # Get a list of image ids on a domain
> sudo vdsm-client StorageDomain getImages sdUUID=<domain id>
>
>
> vdsClient -s 0 getImagesList 0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e
>
> c46d5cbb-1802-4086-a31a-5b3f0b874454
> f0d441bb-2bd2-4523-ab59-beef544727b5
> 4fe1b047-60b2-4bc3-b73b-2fad8a81cc02
> 0d0983f0-60d7-4ce0-bf1b-12cfc456acd8
> 71f18680-1b93-4a2a-9bd1-95baeccf2d89
> caed7169-0c90-467f-86d3-a82148b1f0af
> 986849a8-04ea-4b7d-a29f-f023da9020b3
> 9cf3a067-2996-4ff2-b481-b13f7cc73c33
> 3edd1f60-fd43-4e12-9615-b12bcd1a17ab
> e8720901-0383-4266-98a7-fb5b9fb27e52
> fd2ecb98-a8c5-4aff-a6d3-7ac3087ab994
> e204e228-fa35-4457-a31d-3b7964156538
> f58cd721-d8b8-4794-b059-8789a9fecf62
> 8917cb1a-8bd4-4386-a95a-0189a04866ad
> 9cbde83c-7d1d-4b78-bc7a-6d540584c888
> 7be49698-f3a5-4995-b411-f0490a819950
>
>
>
> # Get a list of volume ids in an image
> sudo vdsm-client Image getVolumes imageID=<image> storagepoolID=<pool>
> storagedomainID=<domain>
>
>
> vdsClient -s 0 getVolumesList 0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e
> 00000002-0002-0002-0002-00000000009b
>
> ERROR: ba29b8fd-3618-42d6-a70f-17d883fde0ed : {'status': {'message':
> "Image path does not exist or cannot be accessed/created:
> ('/rhev/data-center/mnt/blockSD/0fcd2921-8a55-4ff7-
> 9cba-dca6ffb7b67e/images/c46d5cbb-1802-4086-a31a-5b3f0b874454',)",
> 'code': 254}}
> ERROR: f60ccaa4-663b-4a39-8ad0-2ed3fb208cb0 : {'status': {'message':
> "Image path does not exist or cannot be accessed/created:
> ('/rhev/data-center/mnt/blockSD/0fcd2921-8a55-4ff7-
> 9cba-dca6ffb7b67e/images/f0d441bb-2bd2-4523-ab59-beef544727b5',)",
> 'code': 254}}
> ERROR: c0bcf836-c648-4421-9340-febcc0e0abfe : {'status': {'message':
> "Image path does not exist or cannot be accessed/created:
> ('/rhev/data-center/mnt/blockSD/0fcd2921-8a55-4ff7-
> 9cba-dca6ffb7b67e/images/4fe1b047-60b2-4bc3-b73b-2fad8a81cc02',)",
> 'code': 254}}
> ba92b801-3619-4753-9e06-3e2028a408cb : {"Updated":true,"Disk
> Description":"OVF_STORE","Storage Domains":[{"uuid":"0fcd2921-
> 8a55-4ff7-9cba-dca6ffb7b67e"}],"Last Updated":"Tue May 30 11:50:06 MSK
> 2017","Size":61440}.
> ERROR: a3f29de7-c6b9-410e-b635-9b3016da7ba2 : {'status': {'message':
> "Image path does not exist or cannot be accessed/created:
> ('/rhev/data-center/mnt/blockSD/0fcd2921-8a55-4ff7-
> 9cba-dca6ffb7b67e/images/71f18680-1b93-4a2a-9bd1-95baeccf2d89',)",
> 'code': 254}}
> 48b6f4f5-9eeb-4cf3-922a-c9c0c2239bee : {"Updated":true,"Disk
> Description":"OVF_STORE","Storage Domains":[{"uuid":"0fcd2921-
> 8a55-4ff7-9cba-dca6ffb7b67e"}],"Last Updated":"Tue May 30 11:50:06 MSK
> 2017","Size":61440}.
> 3b089aed-b3e1-4423-8585-e65752d19ffe : {"DiskAlias":"r13-sed-app_
> Disk1","DiskDescription":""}.
> 7700e1e1-351b-49b5-8681-e121bbf67177 : {"DiskAlias":"r13-sed-db_
> Disk2","DiskDescription":""}.
> 03917876-0e28-4457-bf44-53c7ea2b4d12 : {"DiskAlias":"r13-sed-app_
> Disk2","DiskDescription":""}.
> ERROR: cb36e25a-b9b2-4f54-9b03-a10837bc26ab : {'status': {'message':
> "Image path does not exist or cannot be accessed/created:
> ('/rhev/data-center/mnt/blockSD/0fcd2921-8a55-4ff7-
> 9cba-dca6ffb7b67e/images/e8720901-0383-4266-98a7-fb5b9fb27e52',)",
> 'code': 254}}
> ERROR: 91a7aa8c-4e13-477c-8703-d8224b85bc84 : {'status': {'message':
> "Image path does not exist or cannot be accessed/created:
> ('/rhev/data-center/mnt/blockSD/0fcd2921-8a55-4ff7-
> 9cba-dca6ffb7b67e/images/fd2ecb98-a8c5-4aff-a6d3-7ac3087ab994',)",
> 'code': 254}}
> ba26f8b7-b807-4a6c-a840-6a83e8ec526e : {"DiskAlias":"r13-sed-db_
> Disk1","DiskDescription":""}.
> ERROR: c1d08688-aad1-478e-82a3-e5b5fde85706 : {'status': {'message':
> "Image path does not exist or cannot be accessed/created:
> ('/rhev/data-center/mnt/blockSD/0fcd2921-8a55-4ff7-
> 9cba-dca6ffb7b67e/images/f58cd721-d8b8-4794-b059-8789a9fecf62',)",
> 'code': 254}}
> ERROR: c7a4782f-cde6-40db-a625-810fd2856dfa : {'status': {'message':
> "Image path does not exist or cannot be accessed/created:
> ('/rhev/data-center/mnt/blockSD/0fcd2921-8a55-4ff7-
> 9cba-dca6ffb7b67e/images/8917cb1a-8bd4-4386-a95a-0189a04866ad',)",
> 'code': 254}}
> ERROR: 309a325a-6f13-4a24-b204-e825ddbf0e41 : {'status': {'message':
> "Image path does not exist or cannot be accessed/created:
> ('/rhev/data-center/mnt/blockSD/0fcd2921-8a55-4ff7-
> 9cba-dca6ffb7b67e/images/9cbde83c-7d1d-4b78-bc7a-6d540584c888',)",
> 'code': 254}}
> fd8822ee-4fc9-49ba-9760-87a85d56bf91 : {"DiskAlias":"r13-sed-app_
> Disk3","DiskDescription":""}.
>
>
>
>
>
> If the wanted images are present on the host but not in engine then you
> had a problem with import.  If the images are not visible to the host then
> you probably have some storage connection issue.
>
>
> This disks are presented in oVirt and VM works OK.
>
> ba26f8b7-b807-4a6c-a840-6a83e8ec526e : {"DiskAlias":"r13-sed-db_
> Disk1","DiskDescription":""}.
> 7700e1e1-351b-49b5-8681-e121bbf67177 : {"DiskAlias":"r13-sed-db_
> Disk2","DiskDescription":""}.
>
> I can see another disks that NOT presented in oVirt UI.
>
> 3b089aed-b3e1-4423-8585-e65752d19ffe : {"DiskAlias":"r13-sed-app_
> Disk1","DiskDescription":""}.
> 03917876-0e28-4457-bf44-53c7ea2b4d12 : {"DiskAlias":"r13-sed-app_
> Disk2","DiskDescription":""}.
> fd8822ee-4fc9-49ba-9760-87a85d56bf91 : {"DiskAlias":"r13-sed-app_
> Disk3","DiskDescription":""}.
>
> Is any way to manage/copy this disks?
>
> Another disks with ERROR code are fully corrupted?
>
>
>
>
> On Wed, May 31, 2017 at 6:11 PM, Николаев Алексей <
> alexeynikolaev.post at yandex.ru> wrote:
>
> Please advise, where I can find some materials how oVirt (or maybe VDSM)
> works with block storage, multipath?
>
>
>
> 30.05.2017, 11:57, "Николаев Алексей" <alexeynikolaev.post at yandex.ru>:
>
> Hi, community!
>
> After import DATA DOMAIN some VM disks not visible in oVirt.
> Can I manually copy VM disks from data domain at FC block device?
> ,
>
> _______________________________________________
> Users mailing list
> Users at ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
>
> _______________________________________________
> Users mailing list
> Users at ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
>
>
>
> --
> Adam Litke
>
>


-- 
Adam Litke
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.ovirt.org/pipermail/users/attachments/20170602/3671ac2c/attachment-0001.html>


More information about the Users mailing list