[Users] ovirt / 2 iscsi storage domains / same LUN IDs
Alex Leonhardt
alex.tuxx at gmail.com
Fri Mar 1 09:16:52 UTC 2013
ok, so it looks like ovirt is creating PVs, VGs, and LVs associated to a
iSCSI disk ... now, when i try to add the other storage with the same LUN
ID - this is how it looks like :
[root at TESTHV01 ~]# pvs
/dev/mapper/1IET_00010001: lseek 73266102272 failed: Invalid argument
/dev/mapper/1IET_00010001: lseek 73266102272 failed: Invalid argument
WARNING: Volume Group 7d0f78ff-aa25-4b64-a2ea-c8a65beda616 is not
consistent
PV VG Fmt Attr
PSize PFree
/dev/mapper/1IET_00010001 7d0f78ff-aa25-4b64-a2ea-c8a65beda616 lvm2 a--
68.00g 54.12g
/dev/sda2 vg_root lvm2 a--
78.12g 0
/dev/sda3 vg_vol lvm2 a--
759.26g 0
[root at TESTHV01 ~]# vgs
/dev/mapper/1IET_00010001: lseek 73266102272 failed: Invalid argument
/dev/mapper/1IET_00010001: lseek 73266102272 failed: Invalid argument
/dev/mapper/1IET_00010001: lseek 73266102272 failed: Invalid argument
/dev/mapper/1IET_00010001: lseek 73266102272 failed: Invalid argument
/dev/mapper/1IET_00010001: lseek 73266102272 failed: Invalid argument
WARNING: Inconsistent metadata found for VG
7d0f78ff-aa25-4b64-a2ea-c8a65beda616 - updating to use version 20
/dev/mapper/1IET_00010001: lseek 73266102272 failed: Invalid argument
Automatic metadata correction failed
Recovery of volume group "7d0f78ff-aa25-4b64-a2ea-c8a65beda616" failed.
/dev/mapper/1IET_00010001: lseek 73266102272 failed: Invalid argument
VG #PV #LV #SN Attr VSize VFree
vg_root 1 2 0 wz--n- 78.12g 0
vg_vol 1 1 0 wz--n- 759.26g 0
[root at TESTHV01 ~]# lvs
/dev/mapper/1IET_00010001: lseek 73266102272 failed: Invalid argument
/dev/mapper/1IET_00010001: lseek 73266102272 failed: Invalid argument
/dev/mapper/1IET_00010001: lseek 73266102272 failed: Invalid argument
/dev/mapper/1IET_00010001: lseek 73266102272 failed: Invalid argument
/dev/mapper/1IET_00010001: lseek 73266102272 failed: Invalid argument
WARNING: Inconsistent metadata found for VG
7d0f78ff-aa25-4b64-a2ea-c8a65beda616 - updating to use version 20
/dev/mapper/1IET_00010001: lseek 73266102272 failed: Invalid argument
Automatic metadata correction failed
Recovery of volume group "7d0f78ff-aa25-4b64-a2ea-c8a65beda616" failed.
Skipping volume group 7d0f78ff-aa25-4b64-a2ea-c8a65beda616
/dev/mapper/1IET_00010001: lseek 73266102272 failed: Invalid argument
LV VG Attr LSize Pool Origin Data% Move Log Copy%
Convert
lv_root vg_root -wi-ao--
68.36g
lv_swap vg_root -wi-ao--
9.77g
lv_vol vg_vol -wi-ao--
759.26g
VS
[root at TESTHV01 ~]# pvs
l PV VG Fmt Attr
PSize PFree
/dev/mapper/1IET_00010001 7d0f78ff-aa25-4b64-a2ea-c8a65beda616 lvm2 a--
68.00g 54.12g
/dev/sda2 vg_root lvm2 a--
78.12g 0
/dev/sda3 vg_vol lvm2 a--
759.26g 0
[root at TESTHV01 ~]# vgs
VG #PV #LV #SN Attr VSize VFree
7d0f78ff-aa25-4b64-a2ea-c8a65beda616 1 7 0 wz--n- 68.00g 54.12g
vg_root 1 2 0 wz--n- 78.12g 0
vg_vol 1 1 0 wz--n- 759.26g 0
[root at TESTHV01 ~]# lvs
LV VG
Attr LSize Pool Origin Data% Move Log Copy% Convert
fe6f8584-b6da-4ef0-8879-bf23022827d7 7d0f78ff-aa25-4b64-a2ea-c8a65beda616
-wi----- 10.00g
ids 7d0f78ff-aa25-4b64-a2ea-c8a65beda616
-wi-a--- 128.00m
inbox 7d0f78ff-aa25-4b64-a2ea-c8a65beda616
-wi-a--- 128.00m
leases 7d0f78ff-aa25-4b64-a2ea-c8a65beda616
-wi-a--- 2.00g
master 7d0f78ff-aa25-4b64-a2ea-c8a65beda616
-wi-ao-- 1.00g
metadata 7d0f78ff-aa25-4b64-a2ea-c8a65beda616
-wi-a--- 512.00m
outbox 7d0f78ff-aa25-4b64-a2ea-c8a65beda616
-wi-a--- 128.00m
lv_root vg_root
-wi-ao-- 68.36g
lv_swap vg_root
-wi-ao-- 9.77g
lv_vol vg_vol
-wi-ao-- 759.26g
Ovirt Node ( HV ) seems to use the LV created with in the same VG for the
storage domain as the disk for the VM. And this is how it'll look like -
various LVs are being created for the "management of the storage domain" (i
guess) :
Disk /dev/sdb: 73.4 GB, 73400320000 bytes
255 heads, 63 sectors/track, 8923 cylinders
Units = cylinders of 16065 * 512 = 8225280 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytes
Disk identifier: 0x00000000
Disk /dev/mapper/1IET_00010001: 73.4 GB, 73400320000 bytes
255 heads, 63 sectors/track, 8923 cylinders
Units = cylinders of 16065 * 512 = 8225280 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytes
Disk identifier: 0x00000000
Disk /dev/mapper/7d0f78ff--aa25--4b64--a2ea--c8a65beda616-metadata: 536 MB,
536870912 bytes
255 heads, 63 sectors/track, 65 cylinders
Units = cylinders of 16065 * 512 = 8225280 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytes
Disk identifier: 0x00000000
Disk /dev/mapper/7d0f78ff--aa25--4b64--a2ea--c8a65beda616-ids: 134 MB,
134217728 bytes
255 heads, 63 sectors/track, 16 cylinders
Units = cylinders of 16065 * 512 = 8225280 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytes
Disk identifier: 0x00000000
Disk /dev/mapper/7d0f78ff--aa25--4b64--a2ea--c8a65beda616-leases: 2147 MB,
2147483648 bytes
255 heads, 63 sectors/track, 261 cylinders
Units = cylinders of 16065 * 512 = 8225280 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytes
Disk identifier: 0x00000000
Disk /dev/mapper/7d0f78ff--aa25--4b64--a2ea--c8a65beda616-inbox: 134 MB,
134217728 bytes
255 heads, 63 sectors/track, 16 cylinders
Units = cylinders of 16065 * 512 = 8225280 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytes
Disk identifier: 0x00000000
Disk /dev/mapper/7d0f78ff--aa25--4b64--a2ea--c8a65beda616-outbox: 134 MB,
134217728 bytes
255 heads, 63 sectors/track, 16 cylinders
Units = cylinders of 16065 * 512 = 8225280 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytes
Disk identifier: 0x00000000
Disk /dev/mapper/7d0f78ff--aa25--4b64--a2ea--c8a65beda616-master: 1073 MB,
1073741824 bytes
255 heads, 63 sectors/track, 130 cylinders
Units = cylinders of 16065 * 512 = 8225280 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytes
Disk identifier: 0x00000000
Disk
/dev/mapper/7d0f78ff--aa25--4b64--a2ea--c8a65beda616-fe6f8584--b6da--4ef0--8879--bf23022827d7:
10.7 GB, 10737418240 bytes
255 heads, 63 sectors/track, 1305 cylinders
Units = cylinders of 16065 * 512 = 8225280 bytes
Sector size (logical/physical): 512 bytes / 512 bytes
I/O size (minimum/optimal): 512 bytes / 512 bytes
Disk identifier: 0x0000c91a
Device Boot Start End Blocks Id System
/dev/mapper/7d0f78ff--aa25--4b64--a2ea--c8a65beda616-fe6f8584--b6da--4ef0--8879
* 1 64 512000 83 Linux
Partition 1 does not end on cylinder boundary.
/dev/mapper/7d0f78ff--aa25--4b64--a2ea--c8a65beda616-fe6f8584--b6da--4ef0--8879
64 1306 9972736 8e Linux LVM
Now I wonder whether there is a MAX on the amount of iSCSI storage domains
you can have ? I'd think that that depends on the max of the device-mapper
? Is there a max ?
Alex
On 28 February 2013 16:30, Alex Leonhardt <alex.tuxx at gmail.com> wrote:
> FWIW, restarting the HV "resolves" the issue and brings the storage domain
> back up; but I dont know whether that is because that's where the target
> (and iscsi initiator) runs or whether vdsmd then clears its cache / routes
> to a/the target(s) ?
>
> Alex
>
>
>
> On 28 February 2013 16:25, Alex Leonhardt <alex.tuxx at gmail.com> wrote:
>
>> another screenshot of how confused it can get
>>
>> alex
>>
>>
>> On 28 February 2013 15:36, Alex Leonhardt <alex.tuxx at gmail.com> wrote:
>>
>>> Hi there,
>>>
>>> I was doing some testing around ovirt and iscsi and found an issue where
>>> as when you use "dd" to create "backing-stores" for iscsi and you point
>>> ovirt to it to discover & login, it thinks the LUN ID is the same although
>>> the target is different and adds additional paths to the config
>>> (automagically?) bringing down the iSCSI storage domain.
>>>
>>> See attached screenshot of what I got when trying to a "new iscsi san
>>> storage domain" to ovirt. The Storage Domain is now down and I cannot get
>>> rid of the config (???) how do I force it to logout of the targets ??
>>>
>>> Also, anyone know how to deal with the duplicate LUN ID issue ?
>>>
>>> Thanks
>>> Alex
>>>
>>> --
>>>
>>> | RHCE | Senior Systems Engineer | www.vcore.co | www.vsearchcloud.com|
>>>
>>
>>
>>
>> --
>>
>> | RHCE | Senior Systems Engineer | www.vcore.co | www.vsearchcloud.com |
>>
>
>
>
> --
>
> | RHCE | Senior Systems Engineer | www.vcore.co | www.vsearchcloud.com |
>
--
| RHCE | Senior Systems Engineer | www.vcore.co | www.vsearchcloud.com |
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.ovirt.org/pipermail/users/attachments/20130301/797f7792/attachment-0001.html>
More information about the Users
mailing list