[Users] [rhev 3] add new domain fails: Could not retrieve LUNs
Haim
hateya at redhat.com
Thu Aug 23 16:22:24 UTC 2012
On 08/23/2012 07:13 PM, Scotto Alberto wrote:
> I was going to reply.. :)
> It looks like it halts after an error due to cciss!c0d1, which btw is displayed by multipath -ll.
> That's just a local disk, isn't it? So it shouldn't even be listed. I may have attached it by mistake, playing with /sys/class/fc_* tools.
> So, if I remove that path, everything should go ok. Do you think so too?
Yes, please remove and check again (make sure to clean device-mapper table with dmsetup remove $dm).
anyhow, vdsm should be more robust, so please file a bug for it(https://bugzilla.redhat.com/enter_bug.cgi?product=oVirt).
Haim
>
>> it appears that vdsm fails to handle a device with '!' in it (cciss!c0d1), but let's make sure its indeed the case
> More than that the fact is that the path scsi_disk/ doesn't exist in /sys/block/cciss!c0d1/device
> And this must be due to the fact that c0d1 is NOT a damn scsi disk
>
>
> Anyway, here is your output
>
> [root at pittor06vhxd020 ~]# ls -l /sys/block/
> total 0
> lrwxrwxrwx. 1 root root 0 2007-06-30 00:37 cciss!c0d0 -> ../devices/pci0000:00/0000:00:03.0/0000:06:00.0/cciss0/c0d0/block/cciss!c0d0
> lrwxrwxrwx. 1 root root 0 2007-06-30 00:32 cciss!c0d1 -> ../devices/pci0000:00/0000:00:03.0/0000:06:00.0/cciss0/c0d1/block/cciss!c0d1
> lrwxrwxrwx. 1 root root 0 2007-06-30 01:17 dm-0 -> ../devices/virtual/block/dm-0
> lrwxrwxrwx. 1 root root 0 2007-06-30 00:40 dm-1 -> ../devices/virtual/block/dm-1
> lrwxrwxrwx. 1 root root 0 2007-06-30 01:17 dm-2 -> ../devices/virtual/block/dm-2
> lrwxrwxrwx. 1 root root 0 2007-06-30 00:49 dm-3 -> ../devices/virtual/block/dm-3
> lrwxrwxrwx. 1 root root 0 2007-06-30 00:49 dm-4 -> ../devices/virtual/block/dm-4
> lrwxrwxrwx. 1 root root 0 2007-06-30 01:00 dm-5 -> ../devices/virtual/block/dm-5
> lrwxrwxrwx. 1 root root 0 2007-06-30 00:49 dm-6 -> ../devices/virtual/block/dm-6
> lrwxrwxrwx. 1 root root 0 2007-06-30 00:10 loop0 -> ../devices/virtual/block/loop0
> lrwxrwxrwx. 1 root root 0 2007-06-30 00:10 loop1 -> ../devices/virtual/block/loop1
> lrwxrwxrwx. 1 root root 0 2007-06-30 00:10 loop2 -> ../devices/virtual/block/loop2
> lrwxrwxrwx. 1 root root 0 2007-06-30 00:10 loop3 -> ../devices/virtual/block/loop3
> lrwxrwxrwx. 1 root root 0 2007-06-30 00:10 loop4 -> ../devices/virtual/block/loop4
> lrwxrwxrwx. 1 root root 0 2007-06-30 00:10 loop5 -> ../devices/virtual/block/loop5
> lrwxrwxrwx. 1 root root 0 2007-06-30 00:10 loop6 -> ../devices/virtual/block/loop6
> lrwxrwxrwx. 1 root root 0 2007-06-30 00:10 loop7 -> ../devices/virtual/block/loop7
> lrwxrwxrwx. 1 root root 0 2007-06-30 00:10 ram0 -> ../devices/virtual/block/ram0
> lrwxrwxrwx. 1 root root 0 2007-06-30 00:10 ram1 -> ../devices/virtual/block/ram1
> lrwxrwxrwx. 1 root root 0 2007-06-30 00:10 ram10 -> ../devices/virtual/block/ram10
> lrwxrwxrwx. 1 root root 0 2007-06-30 00:10 ram11 -> ../devices/virtual/block/ram11
> lrwxrwxrwx. 1 root root 0 2007-06-30 00:10 ram12 -> ../devices/virtual/block/ram12
> lrwxrwxrwx. 1 root root 0 2007-06-30 00:10 ram13 -> ../devices/virtual/block/ram13
> lrwxrwxrwx. 1 root root 0 2007-06-30 00:10 ram14 -> ../devices/virtual/block/ram14
> lrwxrwxrwx. 1 root root 0 2007-06-30 00:10 ram15 -> ../devices/virtual/block/ram15
> lrwxrwxrwx. 1 root root 0 2007-06-30 00:10 ram2 -> ../devices/virtual/block/ram2
> lrwxrwxrwx. 1 root root 0 2007-06-30 00:10 ram3 -> ../devices/virtual/block/ram3
> lrwxrwxrwx. 1 root root 0 2007-06-30 00:10 ram4 -> ../devices/virtual/block/ram4
> lrwxrwxrwx. 1 root root 0 2007-06-30 00:10 ram5 -> ../devices/virtual/block/ram5
> lrwxrwxrwx. 1 root root 0 2007-06-30 00:10 ram6 -> ../devices/virtual/block/ram6
> lrwxrwxrwx. 1 root root 0 2007-06-30 00:10 ram7 -> ../devices/virtual/block/ram7
> lrwxrwxrwx. 1 root root 0 2007-06-30 00:10 ram8 -> ../devices/virtual/block/ram8
> lrwxrwxrwx. 1 root root 0 2007-06-30 00:10 ram9 -> ../devices/virtual/block/ram9
> lrwxrwxrwx. 1 root root 0 2007-06-30 00:36 sda -> ../devices/pci0000:00/0000:00:02.0/0000:09:00.0/0000:0a:00.0/0000:0b:00.0/host2/rport-2:0-0/target2:0:0/2:0:0:0/block/sda
> lrwxrwxrwx. 1 root root 0 2007-06-29 18:59 sdb -> ../devices/pci0000:00/0000:00:02.0/0000:09:00.0/0000:0a:00.0/0000:0b:00.0/host2/rport-2:0-2/target2:0:2/2:0:2:0/block/sdb
> lrwxrwxrwx. 1 root root 0 2007-06-29 18:59 sdc -> ../devices/pci0000:00/0000:00:02.0/0000:09:00.0/0000:0a:00.0/0000:0b:00.0/host2/rport-2:0-3/target2:0:3/2:0:3:0/block/sdc
> lrwxrwxrwx. 1 root root 0 2007-06-29 18:59 sdd -> ../devices/pci0000:00/0000:00:02.0/0000:09:00.0/0000:0a:00.0/0000:0b:00.0/host2/rport-2:0-1/target2:0:1/2:0:1:0/block/sdd
> lrwxrwxrwx. 1 root root 0 2007-06-29 18:59 sde -> ../devices/pci0000:00/0000:00:02.0/0000:09:00.0/0000:0a:00.0/0000:0b:00.1/host3/rport-3:0-0/target3:0:0/3:0:0:0/block/sde
> lrwxrwxrwx. 1 root root 0 2007-06-29 18:59 sdf -> ../devices/pci0000:00/0000:00:02.0/0000:09:00.0/0000:0a:00.0/0000:0b:00.1/host3/rport-3:0-1/target3:0:1/3:0:1:0/block/sdf
> lrwxrwxrwx. 1 root root 0 2007-06-29 18:59 sdg -> ../devices/pci0000:00/0000:00:02.0/0000:09:00.0/0000:0a:00.0/0000:0b:00.1/host3/rport-3:0-2/target3:0:2/3:0:2:0/block/sdg
> lrwxrwxrwx. 1 root root 0 2007-06-29 18:59 sdh -> ../devices/pci0000:00/0000:00:02.0/0000:09:00.0/0000:0a:00.0/0000:0b:00.1/host3/rport-3:0-3/target3:0:3/3:0:3:0/block/sdh
> lrwxrwxrwx. 1 root root 0 2007-06-30 00:49 sr0 -> ../devices/pci0000:00/0000:00:1f.1/host0/target0:0:0/0:0:0:0/block/sr0
>
> [root at pittor06vhxd020 ~]# dmsetup table
> 3600601601cde1d0066b2fb054dece111: 0 1363148800 multipath 1 queue_if_no_path 1 emc 2 1 round-robin 0 4 1 8:0 1 8:48 1 8:64 1 8:80 1 round-robin 0 4 1 8:16 1 8:32 1 8:96 1 8:112 1
> HostVG-Logging: 0 4194304 linear 104:4 24741888
> HostVG-Swap: 0 24723456 linear 104:4 2048
> 3600508b1001035333920202020200005: 0 429925920 multipath 1 queue_if_no_path 0 1 1 round-robin 0 1 1 104:16 1
> HostVG-Data: 0 40624128 linear 104:4 28936192
> HostVG-Config: 0 16384 linear 104:4 24725504
> live-rw: 0 2097152 snapshot 7:1 7:2 P 8
>
> [root at pittor06vhxd020 ~]# lsblk
> NAME MAJ:MIN RM SIZE RO TYPE MOUNTPOINT
> loop0 7:0 0 99.3M 1 loop
> loop1 7:1 0 1G 1 loop
> ââlive-rw (dm-1) 253:1 0 1G 0 dm /
> loop2 7:2 0 512M 0 loop
> ââlive-rw (dm-1) 253:1 0 1G 0 dm /
> cciss!c0d0 104:0 0 33.9G 0 disk
> ââcciss!c0d0p1 104:1 0 243M 0 part
> ââcciss!c0d0p2 104:2 0 244M 0 part
> ââcciss!c0d0p3 104:3 0 244M 0 part
> ââcciss!c0d0p4 104:4 0 33.2G 0 part
> ââHostVG-Swap (dm-3) 253:3 0 11.8G 0 lvm [SWAP]
> ââHostVG-Config (dm-4) 253:4 0 8M 0 lvm /config
> ââHostVG-Logging (dm-5) 253:5 0 2G 0 lvm /var/log
> ââHostVG-Data (dm-6) 253:6 0 19.4G 0 lvm /data
> cciss!c0d1 104:16 0 205G 0 disk
> ââ3600508b1001035333920202020200005 (dm-0) 253:0 0 205G 0 mpath
> sr0 11:0 1 1024M 0 rom
> sdb 8:16 0 650G 0 disk
> ââ3600601601cde1d0066b2fb054dece111 (dm-2) 253:2 0 650G 0 mpath
> sda 8:0 0 650G 0 disk
> ââ3600601601cde1d0066b2fb054dece111 (dm-2) 253:2 0 650G 0 mpath
> sdc 8:32 0 650G 0 disk
> ââ3600601601cde1d0066b2fb054dece111 (dm-2) 253:2 0 650G 0 mpath
> sdd 8:48 0 650G 0 disk
> ââ3600601601cde1d0066b2fb054dece111 (dm-2) 253:2 0 650G 0 mpath
> sde 8:64 0 650G 0 disk
> ââ3600601601cde1d0066b2fb054dece111 (dm-2) 253:2 0 650G 0 mpath
> sdf 8:80 0 650G 0 disk
> ââ3600601601cde1d0066b2fb054dece111 (dm-2) 253:2 0 650G 0 mpath
> sdg 8:96 0 650G 0 disk
> ââ3600601601cde1d0066b2fb054dece111 (dm-2) 253:2 0 650G 0 mpath
> sdh 8:112 0 650G 0 disk
> ââ3600601601cde1d0066b2fb054dece111 (dm-2) 253:2 0 650G 0 mpath
>
>
>
> Alberto Scotto
>
> Blue Reply
> Via Cardinal Massaia, 83
> 10147 - Torino - ITALY
> phone: +39 011 29100
> al.scotto at reply.it
> www.reply.it
>
> -----Original Message-----
> From: Haim [mailto:hateya at redhat.com]
> Sent: giovedì 23 agosto 2012 18:01
> To: Scotto Alberto
> Cc: users at ovirt.org
> Subject: Re: [Users] [rhev 3] add new domain fails: Could not retrieve LUNs
>
> On 08/23/2012 06:20 PM, Scotto Alberto wrote:
>> Here you are
> thanks, can you run the following?
>
> - ls -l /sys/block/
> - dmsetup table
> - lsblk (if exists)
>
> it appears that vdsm fails to handle a device with '!' in it (cciss!c0d1), but let's make sure its indeed the case.
>
>>
>> Thread-47346::DEBUG::2007-06-30
>> 00:37:10,268::clientIF::239::Storage.Dispatcher.Protect::(wrapper)
>> [10.16.250.216]
>> Thread-47346::INFO::2007-06-30
>> 00:37:10,269::dispatcher::94::Storage.Dispatcher.Protect::(run) Run
>> and protect: getDeviceList, args: ()
>> Thread-47346::DEBUG::2007-06-30
>> 00:37:10,269::task::495::TaskManager.Task::(_debug) Task
>> 0be1d461-f8fa-4c20-861d-27fde8124408: moving from state init -> state
>> preparing
>> Thread-47346::DEBUG::2007-06-30
>> 00:37:10,269::misc::1010::SamplingMethod::(__call__) Trying to enter
>> sampling method (storage.sdc.refreshStorage)
>> Thread-47346::DEBUG::2007-06-30
>> 00:37:10,270::misc::1012::SamplingMethod::(__call__) Got in to
>> sampling method
>> Thread-47346::DEBUG::2007-06-30
>> 00:37:10,270::misc::1010::SamplingMethod::(__call__) Trying to enter
>> sampling method (storage.iscsi.rescan)
>> Thread-47346::DEBUG::2007-06-30
>> 00:37:10,270::misc::1012::SamplingMethod::(__call__) Got in to
>> sampling method
>> Thread-47346::DEBUG::2007-06-30
>> 00:37:10,271::iscsi::699::Storage.Misc.excCmd::(rescan) '/usr/bin/sudo
>> -n /sbin/iscsiadm -m session -R' (cwd None)
>> Thread-47346::DEBUG::2007-06-30
>> 00:37:10,300::iscsi::699::Storage.Misc.excCmd::(rescan) FAILED: <err>
>> = 'iscsiadm: No session found.\n'; <rc> = 21
>> Thread-47346::DEBUG::2007-06-30
>> 00:37:10,301::misc::1020::SamplingMethod::(__call__) Returning last
>> result
>> Thread-47346::DEBUG::2007-06-30
>> 00:37:10,661::multipath::61::Storage.Misc.excCmd::(rescan)
>> '/usr/bin/sudo -n /sbin/multipath' (cwd None)
>> Thread-47346::DEBUG::2007-06-30
>> 00:37:10,785::multipath::61::Storage.Misc.excCmd::(rescan) SUCCESS:
>> <err> = ''; <rc> = 0
>> Thread-47346::DEBUG::2007-06-30
>> 00:37:10,786::lvm::547::OperationMutex::(_invalidateAllPvs) Operation
>> 'lvm invalidate operation' got the operation mutex
>> Thread-47346::DEBUG::2007-06-30
>> 00:37:10,786::lvm::549::OperationMutex::(_invalidateAllPvs) Operation
>> 'lvm invalidate operation' released the operation mutex
>> Thread-47346::DEBUG::2007-06-30
>> 00:37:10,786::lvm::559::OperationMutex::(_invalidateAllVgs) Operation
>> 'lvm invalidate operation' got the operation mutex
>> Thread-47346::DEBUG::2007-06-30
>> 00:37:10,787::lvm::561::OperationMutex::(_invalidateAllVgs) Operation
>> 'lvm invalidate operation' released the operation mutex
>> Thread-47346::DEBUG::2007-06-30
>> 00:37:10,787::lvm::580::OperationMutex::(_invalidateAllLvs) Operation
>> 'lvm invalidate operation' got the operation mutex
>> Thread-47346::DEBUG::2007-06-30
>> 00:37:10,788::lvm::582::OperationMutex::(_invalidateAllLvs) Operation
>> 'lvm invalidate operation' released the operation mutex
>> Thread-47346::DEBUG::2007-06-30
>> 00:37:10,788::misc::1020::SamplingMethod::(__call__) Returning last
>> result
>> Thread-47346::DEBUG::2007-06-30
>> 00:37:10,788::lvm::406::OperationMutex::(_reloadpvs) Operation 'lvm
>> reload operation' got the operation mutex
>> Thread-47346::DEBUG::2007-06-30
>> 00:37:10,791::lvm::374::Storage.Misc.excCmd::(cmd) '/usr/bin/sudo -n
>> /sbin/lvm pvs --config " devices { preferred_names =
>> [\\"^/dev/mapper/\\"] ignore_suspended_devices=1 write_cache_state=0
>> disable_after_error_count=3 filter = [
>> \\"a%3600508b1001035333920202020200005|3600601601cde1d0066b2fb054dece1
>> 11%\\", \\"r%.*%\\" ] } global { locking_type=1
>> prioritise_write_locks=1 wait_for_locks=1 } backup { retain_min =
>> 50 retain_days = 0 } " --noheadings --units b --nosuffix --separator
>> | -o
>> uuid,name,size,vg_name,vg_uuid,pe_start,pe_count,pe_alloc_count,mda_co
>> unt,dev_size' (cwd None)
>> Thread-47346::DEBUG::2007-06-30
>> 00:37:10,997::lvm::374::Storage.Misc.excCmd::(cmd) SUCCESS: <err> = '
>> /dev/sdh: read failed after 0 of 4096 at 0: Input/output error\n
>> /dev/sdh: read failed after 0 of 4096 at 697932120064: Input/output
>> error\n /dev/sdh: read failed after 0 of 4096 at 697932177408:
>> Input/output error\n WARNING: Error counts reached a limit of 3.
>> Device /dev/sdh was disabled\n'; <rc> = 0
>> Thread-47346::DEBUG::2007-06-30
>> 00:37:10,998::lvm::429::OperationMutex::(_reloadpvs) Operation 'lvm
>> reload operation' released the operation mutex
>> MainProcess|Thread-47346::DEBUG::2007-06-30
>> MainProcess|00:37:11,005::devicemapper::144::Storage.Misc.excCmd::(_ge
>> MainProcess|tPathsStatus) '/sbin/dmsetup status' (cwd None)
>> MainProcess|Thread-47346::DEBUG::2007-06-30
>> MainProcess|00:37:11,014::devicemapper::144::Storage.Misc.excCmd::(_ge
>> MainProcess|tPathsStatus) SUCCESS: <err> = ''; <rc> = 0
>> MainProcess|Thread-47346::DEBUG::2007-06-30
>> MainProcess|00:37:11,019::multipath::159::Storage.Misc.excCmd::(getScs
>> MainProcess|iSerial) '/sbin/scsi_id --page=0x80 --whitelisted --export
>> MainProcess|--replace-whitespace --device=/dev/dm-0' (cwd None)
>> MainProcess|Thread-47346::DEBUG::2007-06-30
>> MainProcess|00:37:11,026::multipath::159::Storage.Misc.excCmd::(getScs
>> MainProcess|iSerial) SUCCESS: <err> = ''; <rc> = 0
>> Thread-47346::WARNING::2007-06-30
>> 00:37:11,027::multipath::261::Storage.Multipath::(pathListIter) Problem getting hbtl from device `cciss!c0d1` Traceback (most recent call last):
>> File "/usr/share/vdsm/storage/multipath.py", line 259, in pathListIter
>> File "/usr/share/vdsm/storage/multipath.py", line 182, in getHBTL
>> OSError: [Errno 2] No such file or directory: '/sys/block/cciss!c0d1/device/scsi_disk/'
>> Thread-47346::ERROR::2007-06-30
>> 00:37:11,029::task::868::TaskManager.Task::(_setError) Unexpected error Traceback (most recent call last):
>> File "/usr/share/vdsm/storage/task.py", line 876, in _run
>> File "/usr/share/vdsm/storage/hsm.py", line 696, in public_getDeviceList
>> File "/usr/share/vdsm/storage/hsm.py", line 759, in _getDeviceList
>> KeyError: 'hbtl'
>> Thread-47346::DEBUG::2007-06-30
>> 00:37:11,030::task::495::TaskManager.Task::(_debug) Task
>> 0be1d461-f8fa-4c20-861d-27fde8124408: Task._run:
>> 0be1d461-f8fa-4c20-861d-27fde8124408 () {} failed - stopping task
>> Thread-47346::DEBUG::2007-06-30
>> 00:37:11,030::task::495::TaskManager.Task::(_debug) Task
>> 0be1d461-f8fa-4c20-861d-27fde8124408: stopping in state preparing
>> (force False)
>> Thread-47346::DEBUG::2007-06-30
>> 00:37:11,030::task::495::TaskManager.Task::(_debug) Task
>> 0be1d461-f8fa-4c20-861d-27fde8124408: ref 1 aborting True
>> Thread-47346::INFO::2007-06-30
>> 00:37:11,031::task::1171::TaskManager.Task::(prepare) aborting: Task
>> is aborted: "'hbtl'" - code 100
>> Thread-47346::DEBUG::2007-06-30 00:37:11,031::task::495::TaskManager.Task::(_debug) Task 0be1d461-f8fa-4c20-861d-27fde8124408: Prepare: aborted: 'hbtl'
>> Thread-47346::DEBUG::2007-06-30
>> 00:37:11,031::task::495::TaskManager.Task::(_debug) Task
>> 0be1d461-f8fa-4c20-861d-27fde8124408: ref 0 aborting True
>> Thread-47346::DEBUG::2007-06-30
>> 00:37:11,032::task::495::TaskManager.Task::(_debug) Task
>> 0be1d461-f8fa-4c20-861d-27fde8124408: Task._doAbort: force False
>> Thread-47346::DEBUG::2007-06-30
>> 00:37:11,032::resourceManager::821::ResourceManager.Owner::(cancelAll)
>> Owner.cancelAll requests {}
>> Thread-47346::DEBUG::2007-06-30
>> 00:37:11,032::task::495::TaskManager.Task::(_debug) Task
>> 0be1d461-f8fa-4c20-861d-27fde8124408: moving from state preparing ->
>> state aborting
>> Thread-47346::DEBUG::2007-06-30
>> 00:37:11,033::task::495::TaskManager.Task::(_debug) Task
>> 0be1d461-f8fa-4c20-861d-27fde8124408: _aborting: recover policy none
>> Thread-47346::DEBUG::2007-06-30
>> 00:37:11,033::task::495::TaskManager.Task::(_debug) Task
>> 0be1d461-f8fa-4c20-861d-27fde8124408: moving from state aborting ->
>> state failed
>> Thread-47346::DEBUG::2007-06-30
>> 00:37:11,033::resourceManager::786::ResourceManager.Owner::(releaseAll
>> ) Owner.releaseAll requests {} resources {}
>> Thread-47346::DEBUG::2007-06-30
>> 00:37:11,034::resourceManager::821::ResourceManager.Owner::(cancelAll)
>> Owner.cancelAll requests {}
>> Thread-47346::ERROR::2007-06-30 00:37:11,034::dispatcher::106::Storage.Dispatcher.Protect::(run) 'hbtl'
>> Thread-47346::ERROR::2007-06-30 00:37:11,034::dispatcher::107::Storage.Dispatcher.Protect::(run) Traceback (most recent call last):
>> File "/usr/share/vdsm/storage/dispatcher.py", line 96, in run
>> File "/usr/share/vdsm/storage/task.py", line 1178, in prepare
>> KeyError: 'hbtl'
>>
>>
>>
>>
>>
>>
>>
>> Alberto Scotto
>>
>> Blue Reply
>> Via Cardinal Massaia, 83
>> 10147 - Torino - ITALY
>> phone: +39 011 29100
>> al.scotto at reply.it
>> www.reply.it
>>
>> -----Original Message-----
>> From: Haim [mailto:hateya at redhat.com]
>> Sent: giovedì 23 agosto 2012 17:00
>> To: Scotto Alberto
>> Cc: users at ovirt.org
>> Subject: Re: [Users] [rhev 3] add new domain fails: Could not retrieve
>> LUNs
>>
>> On 08/23/2012 05:54 PM, Scotto Alberto wrote:
>>
>> hi,
>>
>> can you attach full vdsm log during the execution of getDeviceList command?
>>> Hi all,
>>>
>>> I'm trying to configure a FCP storage domain on RHEV 3.
>>>
>>> I try to add a new domain from the console, but it can't find any
>>> LUNs: "Could not retrieve LUNs, please check your storage"
>>>
>>> Here is the output from /var/log/rhevm/rhevm.log:
>>>
>>> ------------------------------------
>>>
>>> 2007-06-29 21:50:07,811 WARN
>>> [org.ovirt.engine.core.bll.GetConfigurationValueQuery]
>>> (http-0.0.0.0-8443-1) calling GetConfigurationValueQuery with null
>>> version, using default general for version
>>> 2007-06-29 21:50:07,911 INFO
>>> [org.ovirt.engine.core.vdsbroker.vdsbroker.GetDeviceListVDSCommand]
>>> (http-0.0.0.0-8443-1) START, GetDeviceListVDSCommand(vdsId =
>>> 7e077f4c-25d8-11dc-bbcb-001cc4c2469a, storageType=FCP), log id:
>>> 60bdafe6
>>> 2007-06-29 21:50:08,726 ERROR
>>> [org.ovirt.engine.core.vdsbroker.vdsbroker.BrokerCommandBase]
>>> (http-0.0.0.0-8443-1) Failed in GetDeviceListVDS method
>>> 2007-06-29 21:50:08,727 ERROR
>>> [org.ovirt.engine.core.vdsbroker.vdsbroker.BrokerCommandBase]
>>> (http-0.0.0.0-8443-1) Error code BlockDeviceActionError and error
>>> message VDSGenericException: VDSErrorException: Failed to
>>> GetDeviceListVDS, error = Error block device action: ()
>>> 2007-06-29 21:50:08,727 INFO
>>> [org.ovirt.engine.core.vdsbroker.vdsbroker.BrokerCommandBase]
>>> (http-0.0.0.0-8443-1) Command
>>> org.ovirt.engine.core.vdsbroker.vdsbroker.GetDeviceListVDSCommand
>>> return value
>>>
>>> Class Name:
>>> org.ovirt.engine.core.vdsbroker.vdsbroker.LUNListReturnForXmlRpc
>>> lunList Null
>>> mStatus Class Name:
>>> org.ovirt.engine.core.vdsbroker.vdsbroker.StatusForXmlRpc
>>> mCode 600
>>> mMessage Error block device action: ()
>>>
>>> 2007-06-29 21:50:08,727 INFO
>>> [org.ovirt.engine.core.vdsbroker.vdsbroker.BrokerCommandBase]
>>> (http-0.0.0.0-8443-1) Vds: pittor06vhxd020
>>> 2007-06-29 21:50:08,727 ERROR
>>> [org.ovirt.engine.core.vdsbroker.VDSCommandBase]
>>> (http-0.0.0.0-8443-1) Command GetDeviceListVDS execution failed. Exception:
>>> VDSErrorException: VDSGenericException: VDSErrorException: Failed to
>>> GetDeviceListVDS, error = Error block device action: ()
>>> 2007-06-29 21:50:08,727 INFO
>>> [org.ovirt.engine.core.vdsbroker.vdsbroker.GetDeviceListVDSCommand]
>>> (http-0.0.0.0-8443-1) FINISH, GetDeviceListVDSCommand, log id:
>>> 60bdafe6
>>> 2007-06-29 21:50:08,727 ERROR
>>> [org.ovirt.engine.core.bll.storage.GetDeviceListQuery]
>>> (http-0.0.0.0-8443-1) Query GetDeviceListQuery failed. Exception
>>> message is VdcBLLException:
>>> org.ovirt.engine.core.vdsbroker.vdsbroker.VDSErrorException:
>>> VDSGenericException: VDSErrorException: Failed to GetDeviceListVDS,
>>> error = Error block device action: ()
>>>
>>> ----------------------------------------------
>>>
>>> First question: do LUNs have to be visible from RHEV-H or RHEV-M?
>>>
>>> Currently they are visible only from the hypervisor.
>>>
>>> ----------------------------------------
>>>
>>> [root at pittor06vhxd020 log]# multipath -ll
>>> 3600601601cde1d0066b2fb054dece111 dm-2 DGC,RAID 5 size=650G
>>> features='1 queue_if_no_path' hwhandler='1 emc' wp=rw
>>> |-+- policy='round-robin 0' prio=1 status=active
>>> | |- 2:0:0:0 sda 8:0 active ready running
>>> | |- 2:0:1:0 sdd 8:48 active ready running
>>> | |- 3:0:0:0 sde 8:64 active ready running
>>> | `- 3:0:1:0 sdf 8:80 active ready running
>>> `-+- policy='round-robin 0' prio=0 status=enabled
>>> |- 2:0:2:0 sdb 8:16 active ready running
>>> |- 2:0:3:0 sdc 8:32 active ready running
>>> |- 3:0:2:0 sdg 8:96 active ready running
>>> `- 3:0:3:0 sdh 8:112 active ready running
>>> 3600508b1001035333920202020200005 dm-0 HP,LOGICAL VOLUME size=205G
>>> features='1 queue_if_no_path' hwhandler='0' wp=rw
>>> `-+- policy='round-robin 0' prio=1 status=active
>>> `- 0:0:1:0 cciss!c0d1 104:16 active ready running
>>> ------------------------------------------------------
>>>
>>> Our SAN device is Clariion AX150. Is it compatible with ovirt?
>>>
>>> vdsClient -s 0 getDeviceListgives me:
>>>
>>> Error block device action: ()
>>>
>>> Could it be due to SPM turned off? (I have only one host)
>>>
>>> [root at pittor06vhxd020 log]# ps axu | grep -i spm
>>>
>>> root 16068 0.0 0.0 7888 868 pts/1 R+ 00:04 0:00 grep -i spm
>>>
>>> How can I turn it on? I know the command but I don't know what
>>> paramaters append
>>>
>>> spmStart
>>>
>>> <spUUID> <prevID> <prevLVER> <recoveryMode> <scsiFencing> <maxHostID>
>>> <version>
>>>
>>> Thank you very much for any hints.
>>>
>>> AS
>>>
>>>
>>>
>>> Alberto Scotto
>>>
>>> Blue
>>> Via Cardinal Massaia, 83
>>> 10147 - Torino - ITALY
>>> phone: +39 011 29100
>>> al.scotto at reply.it
>>> www.reply.it
>>>
>>>
>>> ---------------------------------------------------------------------
>>> -
>>> --
>>>
>>> --
>>> The information transmitted is intended for the person or entity to
>>> which it is addressed and may contain confidential and/or privileged
>>> material. Any review, retransmission, dissemination or other use of,
>>> or taking of any action in reliance upon, this information by persons
>>> or entities other than the intended recipient is prohibited. If you
>>> received this in error, please contact the sender and delete the
>>> material from any computer.
>>>
>>>
>>> _______________________________________________
>>> Users mailing list
>>> Users at ovirt.org
>>> http://lists.ovirt.org/mailman/listinfo/users
>>
>>
>> ________________________________
>>
>> --
>> The information transmitted is intended for the person or entity to which it is addressed and may contain confidential and/or privileged material. Any review, retransmission, dissemination or other use of, or taking of any action in reliance upon, this information by persons or entities other than the intended recipient is prohibited. If you received this in error, please contact the sender and delete the material from any computer.
>
>
>
> ________________________________
>
> --
> The information transmitted is intended for the person or entity to which it is addressed and may contain confidential and/or privileged material. Any review, retransmission, dissemination or other use of, or taking of any action in reliance upon, this information by persons or entities other than the intended recipient is prohibited. If you received this in error, please contact the sender and delete the material from any computer.
More information about the Users
mailing list