Hmm, try:
pvscan /dev/vdc
Use fdisk to see if partitions can be read. Also, when you copied the data
with dd, did you use the oflag=fsync option? If not, your data could be
cached in memory waiting to be flushed to disk even though the dd command
completed. Try the copy again with oflag=fsync.
On Fri, Jun 2, 2017 at 8:31 PM, Николаев Алексей <
alexeynikolaev.post(a)yandex.ru> wrote:
02.06.2017, 16:09, "Adam Litke" <alitke(a)redhat.com>:
hmm, strange. So all three of those missing volumes have associated LVs.
Can you try to activate them and see if you can read from them?
lvchange --config 'global {use_lvmetad=0}' -ay 0fcd2921-8a55-4ff7-9cba-
dca6ffb7b67e/3b089aed-b3e1-4423-8585-e65752d19ffe
dd if=/dev/0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e/3b089aed-b3e1-4423-8585-e65752d19ffe
of=/dev/null bs=1M count=1
# lvchange --config 'global {use_lvmetad=0}' -ay 0fcd2921-8a55-4ff7-9cba-
dca6ffb7b67e/3b089aed-b3e1-4423-8585-e65752d19ffe
WARNING: lvmetad is running but disabled. Restart lvmetad before enabling
it!
# dd if=/dev/0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e/3b089aed-b3e1-4423-8585-e65752d19ffe
of=/dev/null bs=1M count=1
1+0 записей получено
1+0 записей отправлено
скопировано 1048576 байт (1,0 MB), 0,00918524 c, 114 MB/c
# vchange --config 'global {use_lvmetad=0}' -ay 0fcd2921-8a55-4ff7-9cba-
dca6ffb7b67e/03917876-0e28-4457-bf44-53c7ea2b4d12
WARNING: lvmetad is running but disabled. Restart lvmetad before
enabling it!
# dd if=/dev/0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e/03917876-0e28-4457-bf44-53c7ea2b4d12
of=/dev/null bs=1M count=1
1+0 записей получено
1+0 записей отправлено
скопировано 1048576 байт (1,0 MB), 0,00440839 c, 238 MB/c
# lvchange --config 'global {use_lvmetad=0}' -ay 0fcd2921-8a55-4ff7-9cba-
dca6ffb7b67e/fd8822ee-4fc9-49ba-9760-87a85d56bf91
WARNING: lvmetad is running but disabled. Restart lvmetad before
enabling it!
# dd if=/dev/0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e/fd8822ee-4fc9-49ba-9760-87a85d56bf91
of=/dev/null bs=1M count=1
1+0 записей получено
1+0 записей отправлено
скопировано 1048576 байт (1,0 MB), 0,00448938 c, 234 MB/c
Well, read operation is OK.
If this works, then one way to recover the data is to use the UI to create
new disks of the same size as the old ones. Then, activate the LVs
associated with the old volumes and the new ones. Then use dd (or qemu-img
convert) to copy from old to new. Then attach the new disks to your VM.
I have create a new disk in the UI and activate it.
6050885b-5dd5-476c-b907-4ce2b3f37b0a : {"DiskAlias":"r13-sed-app_
Disk1-recovery","DiskDescription":""}.
# lvchange --config 'global {use_lvmetad=0}' -ay 0fcd2921-8a55-4ff7-9cba-
dca6ffb7b67e/6050885b-5dd5-476c-b907-4ce2b3f37b0a
Copy by dd from old DiskAlias:r13-sed-app_Disk1 to new
DiskAlias:r13-sed-app_Disk1-recovery
# dd if=/dev/0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e/3b089aed-b3e1-4423-8585-e65752d19ffe
of=/dev/0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e/6050885b-5dd5-476c-b907-4ce2b3f37b0a
status=progress
скопировано 18215707136 байт (18 GB), 496,661644 s, 36,7 MB/ss
35651584+0 записей получено
35651584+0 записей отправлено
скопировано 18253611008 <(825)%20361-1008> байт (18 GB), 502,111 c, 36,4
MB/c
Add new disk to existing VM (vdc). But I can't see LVM volumes on this
disk. Where can I be wrong?
# lsblk
NAME MAJ:MIN RM SIZE RO TYPE MOUNTPOINT
sr0 11:0 1 1024M 0 rom
vda 252:0 0 50G 0 disk
├─vda1 252:1 0 200M 0 part /boot
└─vda2 252:2 0 49,8G 0 part
├─vg_r34seddb-LogVol03 (dm-0) 253:0 0 26,8G 0 lvm /
├─vg_r34seddb-LogVol02 (dm-1) 253:1 0 8G 0 lvm [SWAP]
├─vg_r34seddb-LogVol01 (dm-3) 253:3 0 5G 0 lvm /tmp
└─vg_r34seddb-LogVol00 (dm-4) 253:4 0 10G 0 lvm /var
vdb 252:16 0 1000G 0 disk
└─vdb1 252:17 0 1000G 0 part
└─vg_r34seddb00-LogVol00 (dm-2) 253:2 0 1000G 0 lvm /var/lib/pgsql
vdc 252:32 0 50G 0 disk
On Thu, Jun 1, 2017 at 6:44 PM, Николаев Алексей <
alexeynikolaev.post(a)yandex.ru> wrote:
Thx for your help!
01.06.2017, 16:46, "Adam Litke" <alitke(a)redhat.com>:
When you say "not visible in oVirt" you mean that you do not see them in
the UI?
Yes, I can see some VM with prefix "external-" and without disks.
Do you know the specific uuids for the missing volumes? You could use lvm
to check if the LVs are visible to the host.
lvs --config 'global {use_lvmetad=0}' -o +tags
For each LV, the tag beginning with IU_ indicates the image the volume
belongs to.
LV VG
Attr LSize Pool Origin Data% Meta% Move Log Cpy%Sync Convert LV
Tags
03917876-0e28-4457-bf44-53c7ea2b4d12 0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e
-wi------- 117,00g
IU_3edd1f60-fd43-4e12-9615-b12bcd1a17ab,MD_6,PU_00000000-
0000-0000-0000-000000000000
309a325a-6f13-4a24-b204-e825ddbf0e41 0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e
-wi-ao---- 300,00g
IU_9cbde83c-7d1d-4b78-bc7a-6d540584c888,MD_17,PU_00000000-0000-0000-0000-
000000000000
3b089aed-b3e1-4423-8585-e65752d19ffe 0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e
-wi------- 17,00g
IU_986849a8-04ea-4b7d-a29f-f023da9020b3,MD_5,PU_00000000-
0000-0000-0000-000000000000
48b6f4f5-9eeb-4cf3-922a-c9c0c2239bee 0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e
-wi------- 128,00m
IU_caed7169-0c90-467f-86d3-a82148b1f0af,MD_10,PU_00000000-0000-0000-0000-
000000000000
7700e1e1-351b-49b5-8681-e121bbf67177 0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e
-wi------- 83,00g
IU_9cf3a067-2996-4ff2-b481-b13f7cc73c33,MD_7,PU_00000000-
0000-0000-0000-000000000000
91a7aa8c-4e13-477c-8703-d8224b85bc84 0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e
-wi-ao---- 1000,00g
IU_fd2ecb98-a8c5-4aff-a6d3-7ac3087ab994,MD_15,PU_00000000-0000-0000-0000-
000000000000
a3f29de7-c6b9-410e-b635-9b3016da7ba2 0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e
-wi------- 7,50g
IU_71f18680-1b93-4a2a-9bd1-95baeccf2d89,MD_11,PU_00000000-0000-0000-0000-
000000000000
ba26f8b7-b807-4a6c-a840-6a83e8ec526e 0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e
-wi------- 8,00g
IU_e204e228-fa35-4457-a31d-3b7964156538,MD_8,PU_00000000-000
0-0000-0000-000000000000
ba29b8fd-3618-42d6-a70f-17d883fde0ed 0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e
-wi------- 50,00g
IU_c46d5cbb-1802-4086-a31a-5b3f0b874454,MD_16,PU_00000000-0000-0000-0000-
000000000000
ba92b801-3619-4753-9e06-3e2028a408cb 0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e
-wi------- 128,00m
IU_0d0983f0-60d7-4ce0-bf1b-12cfc456acd8,MD_9,PU_00000000-
0000-0000-0000-000000000000
c0bcf836-c648-4421-9340-febcc0e0abfe 0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e
-wi------- 50,00g
IU_4fe1b047-60b2-4bc3-b73b-2fad8a81cc02,MD_13,PU_00000000-0000-0000-0000-
000000000000
c1d08688-aad1-478e-82a3-e5b5fde85706 0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e
-wi-ao---- 1000,00g
IU_f58cd721-d8b8-4794-b059-8789a9fecf62,MD_18,PU_00000000-0000-0000-0000-
000000000000
c7a4782f-cde6-40db-a625-810fd2856dfa 0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e
-wi------- 300,00g
IU_8917cb1a-8bd4-4386-a95a-0189a04866ad,MD_19,PU_00000000-0000-0000-0000-
000000000000
cb36e25a-b9b2-4f54-9b03-a10837bc26ab 0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e
-wi------- 50,00g
IU_e8720901-0383-4266-98a7-fb5b9fb27e52,MD_14,PU_00000000-0000-0000-0000-
000000000000
f60ccaa4-663b-4a39-8ad0-2ed3fb208cb0 0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e
-wi------- 50,00g
IU_f0d441bb-2bd2-4523-ab59-beef544727b5,MD_12,PU_00000000-0000-0000-0000-
000000000000
fd8822ee-4fc9-49ba-9760-87a85d56bf91 0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e
-wi------- 413,00g
IU_7be49698-f3a5-4995-b411-f0490a819950,MD_4,PU_00000000-
0000-0000-0000-000000000000
ids 0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e
-wi-ao---- 128,00m
inbox 0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e
-wi-a----- 128,00m
leases 0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e
-wi-a----- 2,00g
master 0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e
-wi-a----- 1,00g
metadata 0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e
-wi-a----- 512,00m
outbox 0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e
-wi-a----- 128,00m
root centos_node13-05
-wi-ao---- 101,49g
swap centos_node13-05
-wi-ao---- 9,77g
lv-filestore vg-filestore
-wi-a----- 1000,00g
lv-opt vg-opt
-wi-a----- 300,00g
lv-postgres vg-postgres
-wi-a----- 1000,00g
You could also try to use the vdsm commands:
I can only use vdsClient at 3.5....
# Get the storage pool id
sudo vdsm-client Host getConnectedStoragePools
dsClient -s 0 getConnectedStoragePoolsList
00000002-0002-0002-0002-00000000009b
# Get a list of storagedomain IDs
sudo vdsm-client Host getStorageDomains # gives you a list of
storagedomainIDs
vdsClient -s 0 getStorageDomainsList
0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e
# Get a list of image ids on a domain
sudo vdsm-client StorageDomain getImages sdUUID=<domain id>
vdsClient -s 0 getImagesList 0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e
c46d5cbb-1802-4086-a31a-5b3f0b874454
f0d441bb-2bd2-4523-ab59-beef544727b5
4fe1b047-60b2-4bc3-b73b-2fad8a81cc02
0d0983f0-60d7-4ce0-bf1b-12cfc456acd8
71f18680-1b93-4a2a-9bd1-95baeccf2d89
caed7169-0c90-467f-86d3-a82148b1f0af
986849a8-04ea-4b7d-a29f-f023da9020b3
9cf3a067-2996-4ff2-b481-b13f7cc73c33
3edd1f60-fd43-4e12-9615-b12bcd1a17ab
e8720901-0383-4266-98a7-fb5b9fb27e52
fd2ecb98-a8c5-4aff-a6d3-7ac3087ab994
e204e228-fa35-4457-a31d-3b7964156538
f58cd721-d8b8-4794-b059-8789a9fecf62
8917cb1a-8bd4-4386-a95a-0189a04866ad
9cbde83c-7d1d-4b78-bc7a-6d540584c888
7be49698-f3a5-4995-b411-f0490a819950
# Get a list of volume ids in an image
sudo vdsm-client Image getVolumes imageID=<image> storagepoolID=<pool>
storagedomainID=<domain>
vdsClient -s 0 getVolumesList 0fcd2921-8a55-4ff7-9cba-dca6ffb7b67e
00000002-0002-0002-0002-00000000009b
ERROR: ba29b8fd-3618-42d6-a70f-17d883fde0ed : {'status': {'message':
"Image path does not exist or cannot be accessed/created:
('/rhev/data-center/mnt/blockSD/0fcd2921-8a55-4ff7-
9cba-dca6ffb7b67e/images/c46d5cbb-1802-4086-a31a-5b3f0b874454',)",
'code': 254}}
ERROR: f60ccaa4-663b-4a39-8ad0-2ed3fb208cb0 : {'status': {'message':
"Image path does not exist or cannot be accessed/created:
('/rhev/data-center/mnt/blockSD/0fcd2921-8a55-4ff7-
9cba-dca6ffb7b67e/images/f0d441bb-2bd2-4523-ab59-beef544727b5',)",
'code': 254}}
ERROR: c0bcf836-c648-4421-9340-febcc0e0abfe : {'status': {'message':
"Image path does not exist or cannot be accessed/created:
('/rhev/data-center/mnt/blockSD/0fcd2921-8a55-4ff7-
9cba-dca6ffb7b67e/images/4fe1b047-60b2-4bc3-b73b-2fad8a81cc02',)",
'code': 254}}
ba92b801-3619-4753-9e06-3e2028a408cb : {"Updated":true,"Disk
Description":"OVF_STORE","Storage
Domains":[{"uuid":"0fcd2921-
8a55-4ff7-9cba-dca6ffb7b67e"}],"Last Updated":"Tue May 30 11:50:06
MSK
2017","Size":61440}.
ERROR: a3f29de7-c6b9-410e-b635-9b3016da7ba2 : {'status': {'message':
"Image path does not exist or cannot be accessed/created:
('/rhev/data-center/mnt/blockSD/0fcd2921-8a55-4ff7-
9cba-dca6ffb7b67e/images/71f18680-1b93-4a2a-9bd1-95baeccf2d89',)",
'code': 254}}
48b6f4f5-9eeb-4cf3-922a-c9c0c2239bee : {"Updated":true,"Disk
Description":"OVF_STORE","Storage
Domains":[{"uuid":"0fcd2921-
8a55-4ff7-9cba-dca6ffb7b67e"}],"Last Updated":"Tue May 30 11:50:06
MSK
2017","Size":61440}.
3b089aed-b3e1-4423-8585-e65752d19ffe : {"DiskAlias":"r13-sed-app_
Disk1","DiskDescription":""}.
7700e1e1-351b-49b5-8681-e121bbf67177 : {"DiskAlias":"r13-sed-db_
Disk2","DiskDescription":""}.
03917876-0e28-4457-bf44-53c7ea2b4d12 : {"DiskAlias":"r13-sed-app_
Disk2","DiskDescription":""}.
ERROR: cb36e25a-b9b2-4f54-9b03-a10837bc26ab : {'status': {'message':
"Image path does not exist or cannot be accessed/created:
('/rhev/data-center/mnt/blockSD/0fcd2921-8a55-4ff7-
9cba-dca6ffb7b67e/images/e8720901-0383-4266-98a7-fb5b9fb27e52',)",
'code': 254}}
ERROR: 91a7aa8c-4e13-477c-8703-d8224b85bc84 : {'status': {'message':
"Image path does not exist or cannot be accessed/created:
('/rhev/data-center/mnt/blockSD/0fcd2921-8a55-4ff7-
9cba-dca6ffb7b67e/images/fd2ecb98-a8c5-4aff-a6d3-7ac3087ab994',)",
'code': 254}}
ba26f8b7-b807-4a6c-a840-6a83e8ec526e : {"DiskAlias":"r13-sed-db_
Disk1","DiskDescription":""}.
ERROR: c1d08688-aad1-478e-82a3-e5b5fde85706 : {'status': {'message':
"Image path does not exist or cannot be accessed/created:
('/rhev/data-center/mnt/blockSD/0fcd2921-8a55-4ff7-
9cba-dca6ffb7b67e/images/f58cd721-d8b8-4794-b059-8789a9fecf62',)",
'code': 254}}
ERROR: c7a4782f-cde6-40db-a625-810fd2856dfa : {'status': {'message':
"Image path does not exist or cannot be accessed/created:
('/rhev/data-center/mnt/blockSD/0fcd2921-8a55-4ff7-
9cba-dca6ffb7b67e/images/8917cb1a-8bd4-4386-a95a-0189a04866ad',)",
'code': 254}}
ERROR: 309a325a-6f13-4a24-b204-e825ddbf0e41 : {'status': {'message':
"Image path does not exist or cannot be accessed/created:
('/rhev/data-center/mnt/blockSD/0fcd2921-8a55-4ff7-
9cba-dca6ffb7b67e/images/9cbde83c-7d1d-4b78-bc7a-6d540584c888',)",
'code': 254}}
fd8822ee-4fc9-49ba-9760-87a85d56bf91 : {"DiskAlias":"r13-sed-app_
Disk3","DiskDescription":""}.
If the wanted images are present on the host but not in engine then you
had a problem with import. If the images are not visible to the host then
you probably have some storage connection issue.
This disks are presented in oVirt and VM works OK.
ba26f8b7-b807-4a6c-a840-6a83e8ec526e : {"DiskAlias":"r13-sed-db_
Disk1","DiskDescription":""}.
7700e1e1-351b-49b5-8681-e121bbf67177 : {"DiskAlias":"r13-sed-db_
Disk2","DiskDescription":""}.
I can see another disks that NOT presented in oVirt UI.
3b089aed-b3e1-4423-8585-e65752d19ffe : {"DiskAlias":"r13-sed-app_
Disk1","DiskDescription":""}.
03917876-0e28-4457-bf44-53c7ea2b4d12 : {"DiskAlias":"r13-sed-app_
Disk2","DiskDescription":""}.
fd8822ee-4fc9-49ba-9760-87a85d56bf91 : {"DiskAlias":"r13-sed-app_
Disk3","DiskDescription":""}.
Is any way to manage/copy this disks?
Another disks with ERROR code are fully corrupted?
On Wed, May 31, 2017 at 6:11 PM, Николаев Алексей <
alexeynikolaev.post(a)yandex.ru> wrote:
Please advise, where I can find some materials how oVirt (or maybe VDSM)
works with block storage, multipath?
30.05.2017, 11:57, "Николаев Алексей" <alexeynikolaev.post(a)yandex.ru>:
Hi, community!
After import DATA DOMAIN some VM disks not visible in oVirt.
Can I manually copy VM disks from data domain at FC block device?
,
_______________________________________________
Users mailing list
Users(a)ovirt.org
http://lists.ovirt.org/mailman/listinfo/users
_______________________________________________
Users mailing list
Users(a)ovirt.org
http://lists.ovirt.org/mailman/listinfo/users
--
Adam Litke
--
Adam Litke