02.06.2017, 16:09, "Adam Litke" <alitke@redhat.com>:
hmm, strange.  So all three of those missing volumes have associated LVs.  Can you try to activate them and see if you can read from them?

  lvchange --config 'global {use_lvmetad=0}' -ay 0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e/3b089aed-b3e1-4423-8585-e65752d19ffe
  dd if=/dev/0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e/3b089aed-b3e1-4423-8585-e65752d19ffe of=/dev/null bs=1M count=1
 
# lvchange --config 'global {use_lvmetad=0}' -ay 0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e/3b089aed-b3e1-4423-8585-e65752d19ffe
WARNING: lvmetad is running but disabled. Restart lvmetad before enabling it!                                                                                                                                                               
# dd if=/dev/0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e/3b089aed-b3e1-4423-8585-e65752d19ffe of=/dev/null bs=1M count=1                                                                                                           
1+0 записей получено                                                                                                                                                                                                                          
1+0 записей отправлено                                                                                                                                                                                                                        
скопировано 1048576 байт (1,0 MB), 0,00918524 c, 114 MB/c
 
# vchange --config 'global {use_lvmetad=0}' -ay 0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e/03917876-0e28-4457-bf44-53c7ea2b4d12
  WARNING: lvmetad is running but disabled. Restart lvmetad before enabling it!
# dd if=/dev/0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e/03917876-0e28-4457-bf44-53c7ea2b4d12 of=/dev/null bs=1M count=1
1+0 записей получено
1+0 записей отправлено
 скопировано 1048576 байт (1,0 MB), 0,00440839 c, 238 MB/c
 
# lvchange --config 'global {use_lvmetad=0}' -ay 0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e/fd8822ee-4fc9-49ba-9760-87a85d56bf91
  WARNING: lvmetad is running but disabled. Restart lvmetad before enabling it!
# dd if=/dev/0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e/fd8822ee-4fc9-49ba-9760-87a85d56bf91 of=/dev/null bs=1M count=1
1+0 записей получено
1+0 записей отправлено
 скопировано 1048576 байт (1,0 MB), 0,00448938 c, 234 MB/c
 
Well, read operation is OK.
 
 
 
If this works, then one way to recover the data is to use the UI to create new disks of the same size as the old ones.  Then, activate the LVs associated with the old volumes and the new ones.  Then use dd (or qemu-img convert) to copy from old to new.  Then attach the new disks to your VM.
 
I have create a new disk in the UI and activate it.
 
6050885b-5dd5-476c-b907-4ce2b3f37b0a : {"DiskAlias":"r13-sed-app_Disk1-recovery","DiskDescription":""}.
# lvchange --config 'global {use_lvmetad=0}' -ay 0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e/6050885b-5dd5-476c-b907-4ce2b3f37b0a
 
Copy by dd from old DiskAlias:r13-sed-app_Disk1 to new DiskAlias:r13-sed-app_Disk1-recovery
 
# dd if=/dev/0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e/3b089aed-b3e1-4423-8585-e65752d19ffe of=/dev/0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e/6050885b-5dd5-476c-b907-4ce2b3f37b0a status=progress
 скопировано 18215707136 байт (18 GB), 496,661644 s, 36,7 MB/ss
35651584+0 записей получено
35651584+0 записей отправлено
 скопировано 18253611008 байт (18 GB), 502,111 c, 36,4 MB/c
 
Add new disk to existing VM (vdc). But I can't see LVM volumes on this disk. Where can I be wrong?
 
# lsblk
NAME                              MAJ:MIN RM  SIZE RO TYPE MOUNTPOINT
sr0                                11:0    1 1024M  0 rom  
vda                               252:0    0   50G  0 disk
├─vda1                            252:1    0  200M  0 part /boot
└─vda2                            252:2    0 49,8G  0 part
  ├─vg_r34seddb-LogVol03 (dm-0)   253:0    0 26,8G  0 lvm  /
  ├─vg_r34seddb-LogVol02 (dm-1)   253:1    0    8G  0 lvm  [SWAP]
  ├─vg_r34seddb-LogVol01 (dm-3)   253:3    0    5G  0 lvm  /tmp
  └─vg_r34seddb-LogVol00 (dm-4)   253:4    0   10G  0 lvm  /var
vdb                               252:16   0 1000G  0 disk
└─vdb1                            252:17   0 1000G  0 part
  └─vg_r34seddb00-LogVol00 (dm-2) 253:2    0 1000G  0 lvm  /var/lib/pgsql
vdc                               252:32   0   50G  0 disk
 
 
On Thu, Jun 1, 2017 at 6:44 PM, Николаев Алексей <alexeynikolaev.post@yandex.ru> wrote:
Thx for your help!
 
01.06.2017, 16:46, "Adam Litke" <alitke@redhat.com>:
When you say "not visible in oVirt" you mean that you do not see them in the UI? 

Yes, I can see some VM with prefix "external-" and without disks.
 
Do you know the specific uuids for the missing volumes?  You could use lvm to check if the LVs are visible to the host.
 
lvs --config 'global {use_lvmetad=0}' -o +tags
 
For each LV, the tag beginning with IU_ indicates the image the volume belongs to.
 
  LV                                   VG                                   Attr       LSize    Pool Origin Data%  Meta%  Move Log Cpy%Sync Convert LV Tags                                                                              
  03917876-0e28-4457-bf44-53c7ea2b4d12 0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e -wi-------  117,00g                                                     IU_3edd1f60-fd43-4e12-9615-b12bcd1a17ab,MD_6,PU_00000000-0000-0000-0000-000000000000
  309a325a-6f13-4a24-b204-e825ddbf0e41 0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e -wi-ao----  300,00g                                                     IU_9cbde83c-7d1d-4b78-bc7a-6d540584c888,MD_17,PU_00000000-0000-0000-0000-000000000000
  3b089aed-b3e1-4423-8585-e65752d19ffe 0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e -wi-------   17,00g                                                     IU_986849a8-04ea-4b7d-a29f-f023da9020b3,MD_5,PU_00000000-0000-0000-0000-000000000000
  48b6f4f5-9eeb-4cf3-922a-c9c0c2239bee 0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e -wi-------  128,00m                                                     IU_caed7169-0c90-467f-86d3-a82148b1f0af,MD_10,PU_00000000-0000-0000-0000-000000000000
  7700e1e1-351b-49b5-8681-e121bbf67177 0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e -wi-------   83,00g                                                     IU_9cf3a067-2996-4ff2-b481-b13f7cc73c33,MD_7,PU_00000000-0000-0000-0000-000000000000
  91a7aa8c-4e13-477c-8703-d8224b85bc84 0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e -wi-ao---- 1000,00g                                                     IU_fd2ecb98-a8c5-4aff-a6d3-7ac3087ab994,MD_15,PU_00000000-0000-0000-0000-000000000000
  a3f29de7-c6b9-410e-b635-9b3016da7ba2 0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e -wi-------    7,50g                                                     IU_71f18680-1b93-4a2a-9bd1-95baeccf2d89,MD_11,PU_00000000-0000-0000-0000-000000000000
  ba26f8b7-b807-4a6c-a840-6a83e8ec526e 0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e -wi-------    8,00g                                                     IU_e204e228-fa35-4457-a31d-3b7964156538,MD_8,PU_00000000-0000-0000-0000-000000000000
  ba29b8fd-3618-42d6-a70f-17d883fde0ed 0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e -wi-------   50,00g                                                     IU_c46d5cbb-1802-4086-a31a-5b3f0b874454,MD_16,PU_00000000-0000-0000-0000-000000000000
  ba92b801-3619-4753-9e06-3e2028a408cb 0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e -wi-------  128,00m                                                     IU_0d0983f0-60d7-4ce0-bf1b-12cfc456acd8,MD_9,PU_00000000-0000-0000-0000-000000000000
  c0bcf836-c648-4421-9340-febcc0e0abfe 0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e -wi-------   50,00g                                                     IU_4fe1b047-60b2-4bc3-b73b-2fad8a81cc02,MD_13,PU_00000000-0000-0000-0000-000000000000
  c1d08688-aad1-478e-82a3-e5b5fde85706 0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e -wi-ao---- 1000,00g                                                     IU_f58cd721-d8b8-4794-b059-8789a9fecf62,MD_18,PU_00000000-0000-0000-0000-000000000000
  c7a4782f-cde6-40db-a625-810fd2856dfa 0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e -wi-------  300,00g                                                     IU_8917cb1a-8bd4-4386-a95a-0189a04866ad,MD_19,PU_00000000-0000-0000-0000-000000000000
  cb36e25a-b9b2-4f54-9b03-a10837bc26ab 0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e -wi-------   50,00g                                                     IU_e8720901-0383-4266-98a7-fb5b9fb27e52,MD_14,PU_00000000-0000-0000-0000-000000000000
  f60ccaa4-663b-4a39-8ad0-2ed3fb208cb0 0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e -wi-------   50,00g                                                     IU_f0d441bb-2bd2-4523-ab59-beef544727b5,MD_12,PU_00000000-0000-0000-0000-000000000000
  fd8822ee-4fc9-49ba-9760-87a85d56bf91 0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e -wi-------  413,00g                                                     IU_7be49698-f3a5-4995-b411-f0490a819950,MD_4,PU_00000000-0000-0000-0000-000000000000
  ids                                  0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e -wi-ao----  128,00m                                                                                                                                          
  inbox                                0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e -wi-a-----  128,00m                                                                                                                                          
  leases                               0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e -wi-a-----    2,00g                                                                                                                                          
  master                               0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e -wi-a-----    1,00g                                                                                                                                          
  metadata                             0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e -wi-a-----  512,00m                                                                                                                                          
  outbox                               0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e -wi-a-----  128,00m                                                                                                                                          
  root                                 centos_node13-05                     -wi-ao----  101,49g                                                                                                                                          
  swap                                 centos_node13-05                     -wi-ao----    9,77g                                                                                                                                          
  lv-filestore                         vg-filestore                         -wi-a----- 1000,00g                                                                                                                                          
  lv-opt                               vg-opt                               -wi-a-----  300,00g                                                                                                                                          
  lv-postgres                          vg-postgres                          -wi-a----- 1000,00g
 
 
You could also try to use the vdsm commands:

I can only use vdsClient at 3.5....
 
 
# Get the storage pool id
sudo vdsm-client Host getConnectedStoragePools 
 
dsClient -s 0 getConnectedStoragePoolsList
 
00000002-0002-0002-0002-00000000009b
 
 
# Get a list of storagedomain IDs
sudo vdsm-client Host getStorageDomains # gives you a list of storagedomainIDs
 
vdsClient -s 0 getStorageDomainsList
 
0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e
 
 
# Get a list of image ids on a domain
sudo vdsm-client StorageDomain getImages sdUUID=<domain id>
 
vdsClient -s 0 getImagesList 0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e
 
c46d5cbb-1802-4086-a31a-5b3f0b874454
f0d441bb-2bd2-4523-ab59-beef544727b5
4fe1b047-60b2-4bc3-b73b-2fad8a81cc02
0d0983f0-60d7-4ce0-bf1b-12cfc456acd8
71f18680-1b93-4a2a-9bd1-95baeccf2d89
caed7169-0c90-467f-86d3-a82148b1f0af
986849a8-04ea-4b7d-a29f-f023da9020b3
9cf3a067-2996-4ff2-b481-b13f7cc73c33
3edd1f60-fd43-4e12-9615-b12bcd1a17ab
e8720901-0383-4266-98a7-fb5b9fb27e52
fd2ecb98-a8c5-4aff-a6d3-7ac3087ab994
e204e228-fa35-4457-a31d-3b7964156538
f58cd721-d8b8-4794-b059-8789a9fecf62
8917cb1a-8bd4-4386-a95a-0189a04866ad
9cbde83c-7d1d-4b78-bc7a-6d540584c888
7be49698-f3a5-4995-b411-f0490a819950
 
 
# Get a list of volume ids in an image
sudo vdsm-client Image getVolumes imageID=<image> storagepoolID=<pool> storagedomainID=<domain>
 
vdsClient -s 0 getVolumesList 0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e 00000002-0002-0002-0002-00000000009b
 
ERROR: ba29b8fd-3618-42d6-a70f-17d883fde0ed : {'status': {'message': "Image path does not exist or cannot be accessed/created: ('/rhev/data-center/mnt/blockSD/0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e/images/c46d5cbb-1802-4086-a31a-5b3f0b874454',)", 'code': 254}}
ERROR: f60ccaa4-663b-4a39-8ad0-2ed3fb208cb0 : {'status': {'message': "Image path does not exist or cannot be accessed/created: ('/rhev/data-center/mnt/blockSD/0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e/images/f0d441bb-2bd2-4523-ab59-beef544727b5',)", 'code': 254}}
ERROR: c0bcf836-c648-4421-9340-febcc0e0abfe : {'status': {'message': "Image path does not exist or cannot be accessed/created: ('/rhev/data-center/mnt/blockSD/0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e/images/4fe1b047-60b2-4bc3-b73b-2fad8a81cc02',)", 'code': 254}}
ba92b801-3619-4753-9e06-3e2028a408cb : {"Updated":true,"Disk Description":"OVF_STORE","Storage Domains":[{"uuid":"0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e"}],"Last Updated":"Tue May 30 11:50:06 MSK 2017","Size":61440}.
ERROR: a3f29de7-c6b9-410e-b635-9b3016da7ba2 : {'status': {'message': "Image path does not exist or cannot be accessed/created: ('/rhev/data-center/mnt/blockSD/0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e/images/71f18680-1b93-4a2a-9bd1-95baeccf2d89',)", 'code': 254}}
48b6f4f5-9eeb-4cf3-922a-c9c0c2239bee : {"Updated":true,"Disk Description":"OVF_STORE","Storage Domains":[{"uuid":"0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e"}],"Last Updated":"Tue May 30 11:50:06 MSK 2017","Size":61440}.
3b089aed-b3e1-4423-8585-e65752d19ffe : {"DiskAlias":"r13-sed-app_Disk1","DiskDescription":""}.
7700e1e1-351b-49b5-8681-e121bbf67177 : {"DiskAlias":"r13-sed-db_Disk2","DiskDescription":""}.
03917876-0e28-4457-bf44-53c7ea2b4d12 : {"DiskAlias":"r13-sed-app_Disk2","DiskDescription":""}.
ERROR: cb36e25a-b9b2-4f54-9b03-a10837bc26ab : {'status': {'message': "Image path does not exist or cannot be accessed/created: ('/rhev/data-center/mnt/blockSD/0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e/images/e8720901-0383-4266-98a7-fb5b9fb27e52',)", 'code': 254}}
ERROR: 91a7aa8c-4e13-477c-8703-d8224b85bc84 : {'status': {'message': "Image path does not exist or cannot be accessed/created: ('/rhev/data-center/mnt/blockSD/0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e/images/fd2ecb98-a8c5-4aff-a6d3-7ac3087ab994',)", 'code': 254}}
ba26f8b7-b807-4a6c-a840-6a83e8ec526e : {"DiskAlias":"r13-sed-db_Disk1","DiskDescription":""}.
ERROR: c1d08688-aad1-478e-82a3-e5b5fde85706 : {'status': {'message': "Image path does not exist or cannot be accessed/created: ('/rhev/data-center/mnt/blockSD/0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e/images/f58cd721-d8b8-4794-b059-8789a9fecf62',)", 'code': 254}}
ERROR: c7a4782f-cde6-40db-a625-810fd2856dfa : {'status': {'message': "Image path does not exist or cannot be accessed/created: ('/rhev/data-center/mnt/blockSD/0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e/images/8917cb1a-8bd4-4386-a95a-0189a04866ad',)", 'code': 254}}
ERROR: 309a325a-6f13-4a24-b204-e825ddbf0e41 : {'status': {'message': "Image path does not exist or cannot be accessed/created: ('/rhev/data-center/mnt/blockSD/0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e/images/9cbde83c-7d1d-4b78-bc7a-6d540584c888',)", 'code': 254}}
fd8822ee-4fc9-49ba-9760-87a85d56bf91 : {"DiskAlias":"r13-sed-app_Disk3","DiskDescription":""}.
 
 
 
 
If the wanted images are present on the host but not in engine then you had a problem with import.  If the images are not visible to the host then you probably have some storage connection issue.
 
This disks are presented in oVirt and VM works OK.
 
ba26f8b7-b807-4a6c-a840-6a83e8ec526e : {"DiskAlias":"r13-sed-db_Disk1","DiskDescription":""}.
7700e1e1-351b-49b5-8681-e121bbf67177 : {"DiskAlias":"r13-sed-db_Disk2","DiskDescription":""}.
 
I can see another disks that NOT presented in oVirt UI.
 
3b089aed-b3e1-4423-8585-e65752d19ffe : {"DiskAlias":"r13-sed-app_Disk1","DiskDescription":""}.
03917876-0e28-4457-bf44-53c7ea2b4d12 : {"DiskAlias":"r13-sed-app_Disk2","DiskDescription":""}.
fd8822ee-4fc9-49ba-9760-87a85d56bf91 : {"DiskAlias":"r13-sed-app_Disk3","DiskDescription":""}.
 
Is any way to manage/copy this disks?
 
Another disks with ERROR code are fully corrupted?
 
 
 
On Wed, May 31, 2017 at 6:11 PM, Николаев Алексей <alexeynikolaev.post@yandex.ru> wrote:
Please advise, where I can find some materials how oVirt (or maybe VDSM) works with block storage, multipath?
 
 
 
30.05.2017, 11:57, "Николаев Алексей" <alexeynikolaev.post@yandex.ru>:
Hi, community!
 
After import DATA DOMAIN some VM disks not visible in oVirt.
Can I manually copy VM disks from data domain at FC block device?
,

_______________________________________________
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users


_______________________________________________
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users
 
 
 
--
Adam Litke
 
 
--
Adam Litke