[ovirt-users] Manually copying VM disks from FC data domain
Adam Litke
alitke at redhat.com
Mon Jun 5 13:59:56 UTC 2017
Hmm, try:
pvscan /dev/vdc
Use fdisk to see if partitions can be read. Also, when you copied the data
with dd, did you use the oflag=fsync option? If not, your data could be
cached in memory waiting to be flushed to disk even though the dd command
completed. Try the copy again with oflag=fsync.
On Fri, Jun 2, 2017 at 8:31 PM, Николаев Алексей <
alexeynikolaev.post at yandex.ru> wrote:
>
>
> 02.06.2017, 16:09, "Adam Litke" <alitke at redhat.com>:
>
> hmm, strange. So all three of those missing volumes have associated LVs.
> Can you try to activate them and see if you can read from them?
>
> lvchange --config 'global {use_lvmetad=0}' -ay 0fcd2921-8a55-4ff7-9cba-
> dca6ffb7b67e/3b089aed-b3e1-4423-8585-e65752d19ffe
> dd if=/dev/0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e/3b089aed-b3e1-4423-8585-e65752d19ffe
> of=/dev/null bs=1M count=1
>
>
> # lvchange --config 'global {use_lvmetad=0}' -ay 0fcd2921-8a55-4ff7-9cba-
> dca6ffb7b67e/3b089aed-b3e1-4423-8585-e65752d19ffe
> WARNING: lvmetad is running but disabled. Restart lvmetad before enabling
> it!
>
>
> # dd if=/dev/0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e/3b089aed-b3e1-4423-8585-e65752d19ffe
> of=/dev/null bs=1M count=1
>
>
> 1+0 записей получено
>
>
>
> 1+0 записей отправлено
>
>
>
> скопировано 1048576 байт (1,0 MB), 0,00918524 c, 114 MB/c
>
> # vchange --config 'global {use_lvmetad=0}' -ay 0fcd2921-8a55-4ff7-9cba-
> dca6ffb7b67e/03917876-0e28-4457-bf44-53c7ea2b4d12
> WARNING: lvmetad is running but disabled. Restart lvmetad before
> enabling it!
> # dd if=/dev/0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e/03917876-0e28-4457-bf44-53c7ea2b4d12
> of=/dev/null bs=1M count=1
> 1+0 записей получено
> 1+0 записей отправлено
> скопировано 1048576 байт (1,0 MB), 0,00440839 c, 238 MB/c
>
> # lvchange --config 'global {use_lvmetad=0}' -ay 0fcd2921-8a55-4ff7-9cba-
> dca6ffb7b67e/fd8822ee-4fc9-49ba-9760-87a85d56bf91
> WARNING: lvmetad is running but disabled. Restart lvmetad before
> enabling it!
> # dd if=/dev/0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e/fd8822ee-4fc9-49ba-9760-87a85d56bf91
> of=/dev/null bs=1M count=1
> 1+0 записей получено
> 1+0 записей отправлено
> скопировано 1048576 байт (1,0 MB), 0,00448938 c, 234 MB/c
>
> Well, read operation is OK.
>
>
>
>
> If this works, then one way to recover the data is to use the UI to create
> new disks of the same size as the old ones. Then, activate the LVs
> associated with the old volumes and the new ones. Then use dd (or qemu-img
> convert) to copy from old to new. Then attach the new disks to your VM.
>
>
> I have create a new disk in the UI and activate it.
>
> 6050885b-5dd5-476c-b907-4ce2b3f37b0a : {"DiskAlias":"r13-sed-app_
> Disk1-recovery","DiskDescription":""}.
> # lvchange --config 'global {use_lvmetad=0}' -ay 0fcd2921-8a55-4ff7-9cba-
> dca6ffb7b67e/6050885b-5dd5-476c-b907-4ce2b3f37b0a
>
> Copy by dd from old DiskAlias:r13-sed-app_Disk1 to new
> DiskAlias:r13-sed-app_Disk1-recovery
>
> # dd if=/dev/0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e/3b089aed-b3e1-4423-8585-e65752d19ffe
> of=/dev/0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e/6050885b-5dd5-476c-b907-4ce2b3f37b0a
> status=progress
> скопировано 18215707136 байт (18 GB), 496,661644 s, 36,7 MB/ss
> 35651584+0 записей получено
> 35651584+0 записей отправлено
> скопировано 18253611008 <(825)%20361-1008> байт (18 GB), 502,111 c, 36,4
> MB/c
>
> Add new disk to existing VM (vdc). But I can't see LVM volumes on this
> disk. Where can I be wrong?
>
> # lsblk
> NAME MAJ:MIN RM SIZE RO TYPE MOUNTPOINT
> sr0 11:0 1 1024M 0 rom
> vda 252:0 0 50G 0 disk
> ├─vda1 252:1 0 200M 0 part /boot
> └─vda2 252:2 0 49,8G 0 part
> ├─vg_r34seddb-LogVol03 (dm-0) 253:0 0 26,8G 0 lvm /
> ├─vg_r34seddb-LogVol02 (dm-1) 253:1 0 8G 0 lvm [SWAP]
> ├─vg_r34seddb-LogVol01 (dm-3) 253:3 0 5G 0 lvm /tmp
> └─vg_r34seddb-LogVol00 (dm-4) 253:4 0 10G 0 lvm /var
> vdb 252:16 0 1000G 0 disk
> └─vdb1 252:17 0 1000G 0 part
> └─vg_r34seddb00-LogVol00 (dm-2) 253:2 0 1000G 0 lvm /var/lib/pgsql
> vdc 252:32 0 50G 0 disk
>
>
>
> On Thu, Jun 1, 2017 at 6:44 PM, Николаев Алексей <
> alexeynikolaev.post at yandex.ru> wrote:
>
> Thx for your help!
>
> 01.06.2017, 16:46, "Adam Litke" <alitke at redhat.com>:
>
> When you say "not visible in oVirt" you mean that you do not see them in
> the UI?
>
>
> Yes, I can see some VM with prefix "external-" and without disks.
>
>
> Do you know the specific uuids for the missing volumes? You could use lvm
> to check if the LVs are visible to the host.
>
> lvs --config 'global {use_lvmetad=0}' -o +tags
>
> For each LV, the tag beginning with IU_ indicates the image the volume
> belongs to.
>
>
> LV VG
> Attr LSize Pool Origin Data% Meta% Move Log Cpy%Sync Convert LV
> Tags
>
> 03917876-0e28-4457-bf44-53c7ea2b4d12 0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e
> -wi------- 117,00g
> IU_3edd1f60-fd43-4e12-9615-b12bcd1a17ab,MD_6,PU_00000000-
> 0000-0000-0000-000000000000
> 309a325a-6f13-4a24-b204-e825ddbf0e41 0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e
> -wi-ao---- 300,00g
> IU_9cbde83c-7d1d-4b78-bc7a-6d540584c888,MD_17,PU_00000000-0000-0000-0000-
> 000000000000
> 3b089aed-b3e1-4423-8585-e65752d19ffe 0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e
> -wi------- 17,00g
> IU_986849a8-04ea-4b7d-a29f-f023da9020b3,MD_5,PU_00000000-
> 0000-0000-0000-000000000000
> 48b6f4f5-9eeb-4cf3-922a-c9c0c2239bee 0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e
> -wi------- 128,00m
> IU_caed7169-0c90-467f-86d3-a82148b1f0af,MD_10,PU_00000000-0000-0000-0000-
> 000000000000
> 7700e1e1-351b-49b5-8681-e121bbf67177 0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e
> -wi------- 83,00g
> IU_9cf3a067-2996-4ff2-b481-b13f7cc73c33,MD_7,PU_00000000-
> 0000-0000-0000-000000000000
> 91a7aa8c-4e13-477c-8703-d8224b85bc84 0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e
> -wi-ao---- 1000,00g
> IU_fd2ecb98-a8c5-4aff-a6d3-7ac3087ab994,MD_15,PU_00000000-0000-0000-0000-
> 000000000000
> a3f29de7-c6b9-410e-b635-9b3016da7ba2 0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e
> -wi------- 7,50g
> IU_71f18680-1b93-4a2a-9bd1-95baeccf2d89,MD_11,PU_00000000-0000-0000-0000-
> 000000000000
> ba26f8b7-b807-4a6c-a840-6a83e8ec526e 0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e
> -wi------- 8,00g
> IU_e204e228-fa35-4457-a31d-3b7964156538,MD_8,PU_00000000-000
> 0-0000-0000-000000000000
> ba29b8fd-3618-42d6-a70f-17d883fde0ed 0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e
> -wi------- 50,00g
> IU_c46d5cbb-1802-4086-a31a-5b3f0b874454,MD_16,PU_00000000-0000-0000-0000-
> 000000000000
> ba92b801-3619-4753-9e06-3e2028a408cb 0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e
> -wi------- 128,00m
> IU_0d0983f0-60d7-4ce0-bf1b-12cfc456acd8,MD_9,PU_00000000-
> 0000-0000-0000-000000000000
> c0bcf836-c648-4421-9340-febcc0e0abfe 0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e
> -wi------- 50,00g
> IU_4fe1b047-60b2-4bc3-b73b-2fad8a81cc02,MD_13,PU_00000000-0000-0000-0000-
> 000000000000
> c1d08688-aad1-478e-82a3-e5b5fde85706 0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e
> -wi-ao---- 1000,00g
> IU_f58cd721-d8b8-4794-b059-8789a9fecf62,MD_18,PU_00000000-0000-0000-0000-
> 000000000000
> c7a4782f-cde6-40db-a625-810fd2856dfa 0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e
> -wi------- 300,00g
> IU_8917cb1a-8bd4-4386-a95a-0189a04866ad,MD_19,PU_00000000-0000-0000-0000-
> 000000000000
> cb36e25a-b9b2-4f54-9b03-a10837bc26ab 0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e
> -wi------- 50,00g
> IU_e8720901-0383-4266-98a7-fb5b9fb27e52,MD_14,PU_00000000-0000-0000-0000-
> 000000000000
> f60ccaa4-663b-4a39-8ad0-2ed3fb208cb0 0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e
> -wi------- 50,00g
> IU_f0d441bb-2bd2-4523-ab59-beef544727b5,MD_12,PU_00000000-0000-0000-0000-
> 000000000000
> fd8822ee-4fc9-49ba-9760-87a85d56bf91 0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e
> -wi------- 413,00g
> IU_7be49698-f3a5-4995-b411-f0490a819950,MD_4,PU_00000000-
> 0000-0000-0000-000000000000
> ids 0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e
> -wi-ao---- 128,00m
>
>
> inbox 0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e
> -wi-a----- 128,00m
>
>
> leases 0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e
> -wi-a----- 2,00g
>
>
> master 0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e
> -wi-a----- 1,00g
>
>
> metadata 0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e
> -wi-a----- 512,00m
>
>
> outbox 0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e
> -wi-a----- 128,00m
>
>
> root centos_node13-05
> -wi-ao---- 101,49g
>
>
> swap centos_node13-05
> -wi-ao---- 9,77g
>
>
> lv-filestore vg-filestore
> -wi-a----- 1000,00g
>
>
> lv-opt vg-opt
> -wi-a----- 300,00g
>
>
> lv-postgres vg-postgres
> -wi-a----- 1000,00g
>
>
>
> You could also try to use the vdsm commands:
>
>
> I can only use vdsClient at 3.5....
>
>
>
> # Get the storage pool id
> sudo vdsm-client Host getConnectedStoragePools
>
>
> dsClient -s 0 getConnectedStoragePoolsList
>
> 00000002-0002-0002-0002-00000000009b
>
>
>
> # Get a list of storagedomain IDs
> sudo vdsm-client Host getStorageDomains # gives you a list of
> storagedomainIDs
>
>
> vdsClient -s 0 getStorageDomainsList
>
> 0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e
>
>
>
> # Get a list of image ids on a domain
> sudo vdsm-client StorageDomain getImages sdUUID=<domain id>
>
>
> vdsClient -s 0 getImagesList 0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e
>
> c46d5cbb-1802-4086-a31a-5b3f0b874454
> f0d441bb-2bd2-4523-ab59-beef544727b5
> 4fe1b047-60b2-4bc3-b73b-2fad8a81cc02
> 0d0983f0-60d7-4ce0-bf1b-12cfc456acd8
> 71f18680-1b93-4a2a-9bd1-95baeccf2d89
> caed7169-0c90-467f-86d3-a82148b1f0af
> 986849a8-04ea-4b7d-a29f-f023da9020b3
> 9cf3a067-2996-4ff2-b481-b13f7cc73c33
> 3edd1f60-fd43-4e12-9615-b12bcd1a17ab
> e8720901-0383-4266-98a7-fb5b9fb27e52
> fd2ecb98-a8c5-4aff-a6d3-7ac3087ab994
> e204e228-fa35-4457-a31d-3b7964156538
> f58cd721-d8b8-4794-b059-8789a9fecf62
> 8917cb1a-8bd4-4386-a95a-0189a04866ad
> 9cbde83c-7d1d-4b78-bc7a-6d540584c888
> 7be49698-f3a5-4995-b411-f0490a819950
>
>
>
> # Get a list of volume ids in an image
> sudo vdsm-client Image getVolumes imageID=<image> storagepoolID=<pool>
> storagedomainID=<domain>
>
>
> vdsClient -s 0 getVolumesList 0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e
> 00000002-0002-0002-0002-00000000009b
>
> ERROR: ba29b8fd-3618-42d6-a70f-17d883fde0ed : {'status': {'message':
> "Image path does not exist or cannot be accessed/created:
> ('/rhev/data-center/mnt/blockSD/0fcd2921-8a55-4ff7-
> 9cba-dca6ffb7b67e/images/c46d5cbb-1802-4086-a31a-5b3f0b874454',)",
> 'code': 254}}
> ERROR: f60ccaa4-663b-4a39-8ad0-2ed3fb208cb0 : {'status': {'message':
> "Image path does not exist or cannot be accessed/created:
> ('/rhev/data-center/mnt/blockSD/0fcd2921-8a55-4ff7-
> 9cba-dca6ffb7b67e/images/f0d441bb-2bd2-4523-ab59-beef544727b5',)",
> 'code': 254}}
> ERROR: c0bcf836-c648-4421-9340-febcc0e0abfe : {'status': {'message':
> "Image path does not exist or cannot be accessed/created:
> ('/rhev/data-center/mnt/blockSD/0fcd2921-8a55-4ff7-
> 9cba-dca6ffb7b67e/images/4fe1b047-60b2-4bc3-b73b-2fad8a81cc02',)",
> 'code': 254}}
> ba92b801-3619-4753-9e06-3e2028a408cb : {"Updated":true,"Disk
> Description":"OVF_STORE","Storage Domains":[{"uuid":"0fcd2921-
> 8a55-4ff7-9cba-dca6ffb7b67e"}],"Last Updated":"Tue May 30 11:50:06 MSK
> 2017","Size":61440}.
> ERROR: a3f29de7-c6b9-410e-b635-9b3016da7ba2 : {'status': {'message':
> "Image path does not exist or cannot be accessed/created:
> ('/rhev/data-center/mnt/blockSD/0fcd2921-8a55-4ff7-
> 9cba-dca6ffb7b67e/images/71f18680-1b93-4a2a-9bd1-95baeccf2d89',)",
> 'code': 254}}
> 48b6f4f5-9eeb-4cf3-922a-c9c0c2239bee : {"Updated":true,"Disk
> Description":"OVF_STORE","Storage Domains":[{"uuid":"0fcd2921-
> 8a55-4ff7-9cba-dca6ffb7b67e"}],"Last Updated":"Tue May 30 11:50:06 MSK
> 2017","Size":61440}.
> 3b089aed-b3e1-4423-8585-e65752d19ffe : {"DiskAlias":"r13-sed-app_
> Disk1","DiskDescription":""}.
> 7700e1e1-351b-49b5-8681-e121bbf67177 : {"DiskAlias":"r13-sed-db_
> Disk2","DiskDescription":""}.
> 03917876-0e28-4457-bf44-53c7ea2b4d12 : {"DiskAlias":"r13-sed-app_
> Disk2","DiskDescription":""}.
> ERROR: cb36e25a-b9b2-4f54-9b03-a10837bc26ab : {'status': {'message':
> "Image path does not exist or cannot be accessed/created:
> ('/rhev/data-center/mnt/blockSD/0fcd2921-8a55-4ff7-
> 9cba-dca6ffb7b67e/images/e8720901-0383-4266-98a7-fb5b9fb27e52',)",
> 'code': 254}}
> ERROR: 91a7aa8c-4e13-477c-8703-d8224b85bc84 : {'status': {'message':
> "Image path does not exist or cannot be accessed/created:
> ('/rhev/data-center/mnt/blockSD/0fcd2921-8a55-4ff7-
> 9cba-dca6ffb7b67e/images/fd2ecb98-a8c5-4aff-a6d3-7ac3087ab994',)",
> 'code': 254}}
> ba26f8b7-b807-4a6c-a840-6a83e8ec526e : {"DiskAlias":"r13-sed-db_
> Disk1","DiskDescription":""}.
> ERROR: c1d08688-aad1-478e-82a3-e5b5fde85706 : {'status': {'message':
> "Image path does not exist or cannot be accessed/created:
> ('/rhev/data-center/mnt/blockSD/0fcd2921-8a55-4ff7-
> 9cba-dca6ffb7b67e/images/f58cd721-d8b8-4794-b059-8789a9fecf62',)",
> 'code': 254}}
> ERROR: c7a4782f-cde6-40db-a625-810fd2856dfa : {'status': {'message':
> "Image path does not exist or cannot be accessed/created:
> ('/rhev/data-center/mnt/blockSD/0fcd2921-8a55-4ff7-
> 9cba-dca6ffb7b67e/images/8917cb1a-8bd4-4386-a95a-0189a04866ad',)",
> 'code': 254}}
> ERROR: 309a325a-6f13-4a24-b204-e825ddbf0e41 : {'status': {'message':
> "Image path does not exist or cannot be accessed/created:
> ('/rhev/data-center/mnt/blockSD/0fcd2921-8a55-4ff7-
> 9cba-dca6ffb7b67e/images/9cbde83c-7d1d-4b78-bc7a-6d540584c888',)",
> 'code': 254}}
> fd8822ee-4fc9-49ba-9760-87a85d56bf91 : {"DiskAlias":"r13-sed-app_
> Disk3","DiskDescription":""}.
>
>
>
>
>
> If the wanted images are present on the host but not in engine then you
> had a problem with import. If the images are not visible to the host then
> you probably have some storage connection issue.
>
>
> This disks are presented in oVirt and VM works OK.
>
> ba26f8b7-b807-4a6c-a840-6a83e8ec526e : {"DiskAlias":"r13-sed-db_
> Disk1","DiskDescription":""}.
> 7700e1e1-351b-49b5-8681-e121bbf67177 : {"DiskAlias":"r13-sed-db_
> Disk2","DiskDescription":""}.
>
> I can see another disks that NOT presented in oVirt UI.
>
> 3b089aed-b3e1-4423-8585-e65752d19ffe : {"DiskAlias":"r13-sed-app_
> Disk1","DiskDescription":""}.
> 03917876-0e28-4457-bf44-53c7ea2b4d12 : {"DiskAlias":"r13-sed-app_
> Disk2","DiskDescription":""}.
> fd8822ee-4fc9-49ba-9760-87a85d56bf91 : {"DiskAlias":"r13-sed-app_
> Disk3","DiskDescription":""}.
>
> Is any way to manage/copy this disks?
>
> Another disks with ERROR code are fully corrupted?
>
>
>
>
> On Wed, May 31, 2017 at 6:11 PM, Николаев Алексей <
> alexeynikolaev.post at yandex.ru> wrote:
>
> Please advise, where I can find some materials how oVirt (or maybe VDSM)
> works with block storage, multipath?
>
>
>
> 30.05.2017, 11:57, "Николаев Алексей" <alexeynikolaev.post at yandex.ru>:
>
> Hi, community!
>
> After import DATA DOMAIN some VM disks not visible in oVirt.
> Can I manually copy VM disks from data domain at FC block device?
> ,
>
> _______________________________________________
> Users mailing list
> Users at ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
>
> _______________________________________________
> Users mailing list
> Users at ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
>
>
>
>
> --
> Adam Litke
>
>
>
> --
> Adam Litke
>
>
--
Adam Litke
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.ovirt.org/pipermail/users/attachments/20170605/5a94f410/attachment-0001.html>
More information about the Users
mailing list