[ovirt-users] Does my Storage Domain crashed or is this iSCSI LUN's a problem?

Adam Litke alitke at redhat.com
Wed Jan 28 08:45:09 EST 2015


On 26/01/15 13:05 +0100, shimano wrote:
>Hi guys,
>
>I'm trying to run one of my storage domains, which experienced failure.
>Unfortunately, I meet a very nasty error (Storage domain does not exist).
>
>Could someone tell me, how to try to restore this domain?

Could you try moving the host to Maintenance mode and then Activate it
again please.  I've encountered situations where vdsm restarts and
engine does not reconnect storage until an Activate action happens.
Let's see if this is your issue.

>
>P.S.
>It's an oVirt 3.4.2-1.el6
>
>**********************************************************************************
>
>/var/log/messages:
>Jan 26 12:48:49 node002 vdsm TaskManager.Task ERROR
>Task=`10d02993-b585-448f-9a50-bd3e8cda7082`::Unexpected error#012Traceback
>(most recent call last):#012  File "/usr/share/vdsm/storage/task.py", line
>873, in _run#012    return fn(*args, **kargs)#012  File
>"/usr/share/vdsm/logUtils.py", line 45, in wrapper#012    res = f(*args,
>**kwargs)#012  File "/usr/share/vdsm/storage/hsm.py", line 2959, in
>getVGInfo#012    return dict(info=self.__getVGsInfo([vgUUID])[0])#012  File
>"/usr/share/vdsm/storage/hsm.py", line 2892, in __getVGsInfo#012    vgList
>= [lvm.getVGbyUUID(vgUUID) for vgUUID in vgUUIDs]#012  File
>"/usr/share/vdsm/storage/lvm.py", line 894, in getVGbyUUID#012    raise
>se.VolumeGroupDoesNotExist("vg_uuid: %s" %
>vgUUID)#012VolumeGroupDoesNotExist: Volume Group does not exist: ('vg_uuid:
>gyaCWf-6VKi-lI9W-JT6H-IZdy-rIsB-hTvZ4O',)
>Jan 26 12:48:49 node002 kernel: device-mapper: table: 253:26: multipath:
>error getting device
>Jan 26 12:48:49 node002 kernel: device-mapper: ioctl: error adding target
>to table
>
>**********************************************************************************
>
>/var/log/vdsm.log:
>Thread-22::ERROR::2015-01-26
>12:43:03,376::sdc::137::Storage.StorageDomainCache::(_findDomain) looking
>for unfetched domain db52e9cb-7306-43fd-aff3-20831bc2bcaf
>Thread-22::ERROR::2015-01-26
>12:43:03,377::sdc::154::Storage.StorageDomainCache::(_findUnfetchedDomain)
>looking for domain db52e9cb-7306-43fd-aff3-20831bc2bcaf
>Thread-22::DEBUG::2015-01-26
>12:43:03,377::lvm::373::OperationMutex::(_reloadvgs) Operation 'lvm reload
>operation' got the operation mutex
>Thread-22::DEBUG::2015-01-26
>12:43:03,378::lvm::296::Storage.Misc.excCmd::(cmd) u'/usr/bin/sudo -n
>/sbin/lvm vgs --config " devices { preferred_names = [\\"^/dev/mapper/\\"]
>ignore_suspended_devices=1 write_cache_state=0 disable_after_error_count=3
>obtain_device_list_from_udev=0 filter = [
>\'a|/dev/mapper/mpathb|/dev/mapper/mpathc|/dev/mapper/mpathd|/dev/mapper/mpathe|/dev/mapper/mpathf|\',
>\'r|.*|\' ] }  global {  locking_type=1  prioritise_write_locks=1
>wait_for_locks=1  use_lvmetad=0 }  backup {  retain_min = 50  retain_days =
>0 } " --noheadings --units b --nosuffix --separator | -o
>uuid,name,attr,size,free,extent_size,extent_count,free_count,tags,vg_mda_size,vg_mda_free,lv_count,pv_count,pv_name
>db52e9cb-7306-43fd-aff3-20831bc2bcaf' (cwd None)
>Thread-22::DEBUG::2015-01-26
>12:43:03,462::lvm::296::Storage.Misc.excCmd::(cmd) FAILED: <err> = '
>/dev/mapper/mpathc: Checksum error\n  /dev/mapper/mpathc: Checksum error\n
>Volume group "db52e9cb-7306-43fd-aff3-20831bc2bcaf" not found\n  Skipping
>volume group db52e9cb-7306-43fd-aff3-20831bc2bcaf\n'; <rc> = 5
>Thread-22::WARNING::2015-01-26
>12:43:03,466::lvm::378::Storage.LVM::(_reloadvgs) lvm vgs failed: 5 [] ['
>/dev/mapper/mpathc: Checksum error', '  /dev/mapper/mpathc: Checksum
>error', '  Volume group "db52e9cb-7306-43fd-aff3-20831bc2bcaf" not found',
>'  Skipping volume group db52e9cb-7306-43fd-aff3-20831bc2bcaf']
>Thread-22::DEBUG::2015-01-26
>12:43:03,466::lvm::415::OperationMutex::(_reloadvgs) Operation 'lvm reload
>operation' released the operation mutex
>Thread-22::ERROR::2015-01-26
>12:43:03,477::sdc::143::Storage.StorageDomainCache::(_findDomain) domain
>db52e9cb-7306-43fd-aff3-20831bc2bcaf not found
>Traceback (most recent call last):
>  File "/usr/share/vdsm/storage/sdc.py", line 141, in _findDomain
>    dom = findMethod(sdUUID)
>  File "/usr/share/vdsm/storage/sdc.py", line 171, in _findUnfetchedDomain
>    raise se.StorageDomainDoesNotExist(sdUUID)
>StorageDomainDoesNotExist: Storage domain does not exist:
>(u'db52e9cb-7306-43fd-aff3-20831bc2bcaf',)
>Thread-22::ERROR::2015-01-26
>12:43:03,478::domainMonitor::239::Storage.DomainMonitorThread::(_monitorDomain)
>Error while collecting domain db52e9cb-7306-43fd-aff3-20831bc2bcaf
>monitoring information
>Traceback (most recent call last):
>  File "/usr/share/vdsm/storage/domainMonitor.py", line 204, in
>_monitorDomain
>    self.domain = sdCache.produce(self.sdUUID)
>  File "/usr/share/vdsm/storage/sdc.py", line 98, in produce
>    domain.getRealDomain()
>  File "/usr/share/vdsm/storage/sdc.py", line 52, in getRealDomain
>    return self._cache._realProduce(self._sdUUID)
>  File "/usr/share/vdsm/storage/sdc.py", line 122, in _realProduce
>    domain = self._findDomain(sdUUID)
>  File "/usr/share/vdsm/storage/sdc.py", line 141, in _findDomain
>    dom = findMethod(sdUUID)
>  File "/usr/share/vdsm/storage/sdc.py", line 171, in _findUnfetchedDomain
>    raise se.StorageDomainDoesNotExist(sdUUID)
>StorageDomainDoesNotExist: Storage domain does not exist:
>(u'db52e9cb-7306-43fd-aff3-20831bc2bcaf',)
>Thread-13::DEBUG::2015-01-26
>12:43:05,102::task::595::TaskManager.Task::(_updateState)
>Task=`b4e85e37-b216-4d29-a448-0711e370a246`::moving from state init ->
>state preparing
>Thread-13::INFO::2015-01-26
>12:43:05,102::logUtils::44::dispatcher::(wrapper) Run and protect:
>repoStats(options=None)
>Thread-13::INFO::2015-01-26
>12:43:05,103::logUtils::47::dispatcher::(wrapper) Run and protect:
>repoStats, Return response: {u'7969d636-1a02-42ba-a50b-2528765cf3d5':
>{'code': 0, 'version': 0, 'acquired': True, 'delay': '0.000457574',
>'lastCheck': '7.5', 'valid': True},
>u'5e1ca1b6-4706-4c79-8924-b8db741c929f': {'code': 0, 'version': 3,
>'acquired': True, 'delay': '0.00100094', 'lastCheck': '6.3', 'valid':
>True}, u'cb85e6cd-df54-4151-8f3b-7e6d72b7372d': {'code': 0, 'version': 3,
>'acquired': True, 'delay': '0.463061', 'lastCheck': '4.9', 'valid': True},
>u'db52e9cb-7306-43fd-aff3-20831bc2bcaf': {'code': 358, 'version': -1,
>'acquired': False, 'delay': '0', 'lastCheck': '1.6', 'valid': False},
>u'5f595801-aaa5-42c7-b829-7a34a636407e': {'code': 0, 'version': 3,
>'acquired': True, 'delay': '0.000942979', 'lastCheck': '7.9', 'valid':
>True}, u'c1ebd0f8-fa32-4fe3-8569-fb7d4ad8faf4': {'code': 0, 'version': 0,
>'acquired': True, 'delay': '0.000424499', 'lastCheck': '7.3', 'valid':
>True}}
>Thread-13::DEBUG::2015-01-26
>12:43:05,103::task::1185::TaskManager.Task::(prepare)
>Task=`b4e85e37-b216-4d29-a448-0711e370a246`::finished:
>{u'7969d636-1a02-42ba-a50b-2528765cf3d5': {'code': 0, 'version': 0,
>'acquired': True, 'delay': '0.000457574', 'lastCheck': '7.5', 'valid':
>True}, u'5e1ca1b6-4706-4c79-8924-b8db741c929f': {'code': 0, 'version': 3,
>'acquired': True, 'delay': '0.00100094', 'lastCheck': '6.3', 'valid':
>True}, u'cb85e6cd-df54-4151-8f3b-7e6d72b7372d': {'code': 0, 'version': 3,
>'acquired': True, 'delay': '0.463061', 'lastCheck': '4.9', 'valid': True},
>u'db52e9cb-7306-43fd-aff3-20831bc2bcaf': {'code': 358, 'version': -1,
>'acquired': False, 'delay': '0', 'lastCheck': '1.6', 'valid': False},
>u'5f595801-aaa5-42c7-b829-7a34a636407e': {'code': 0, 'version': 3,
>'acquired': True, 'delay': '0.000942979', 'lastCheck': '7.9', 'valid':
>True}, u'c1ebd0f8-fa32-4fe3-8569-fb7d4ad8faf4': {'code': 0, 'version': 0,
>'acquired': True, 'delay': '0.000424499', 'lastCheck': '7.3', 'valid':
>True}}
>
>**********************************************************************************
>
>[root at node002 shim]# multipath -ll
>mpathe (1NODE_001_LUN01) dm-6 SHIMI,VIRTUAL-DISK
>size=977G features='0' hwhandler='0' wp=rw
>`-+- policy='round-robin 0' prio=1 status=active
>  `- 21:0:0:1 sdg 8:96 active ready  running
>mpathd (1NODE_003_LUN01) dm-7 SHIMI,VIRTUAL-DISK
>size=977G features='0' hwhandler='0' wp=rw
>`-+- policy='round-robin 0' prio=1 status=active
>  `- 20:0:0:1 sdf 8:80 active ready  running
>mpathc (1NODE_002_LUN01) dm-4 SHIMI,VIRTUAL-DISK
>size=977G features='0' hwhandler='0' wp=rw
>`-+- policy='round-robin 0' prio=1 status=active
>  `- 18:0:0:1 sdd 8:48 active ready  running
>mpathb (1ATA_MARVELL_Raid_VD_0_1c3c8ecf5cf00010) dm-1 ATA,MARVELL Raid VD
>size=1.8T features='0' hwhandler='0' wp=rw
>`-+- policy='round-robin 0' prio=1 status=active
>  `- 0:0:0:0  sda 8:0  active ready  running
>mpathf (1MANAGER_LUN01) dm-5 SHIMI,VIRTUAL-DISK
>size=500G features='0' hwhandler='0' wp=rw
>`-+- policy='round-robin 0' prio=1 status=active
>  `- 19:0:0:1 sde 8:64 active ready  running
>
>**********************************************************************************
>
>[root at node002 shim]# lsblk
>NAME                                                          MAJ:MIN RM
>SIZE RO TYPE  MOUNTPOINT
>sdb                                                             8:16   0
>298.1G  0 disk
>├─sdb1                                                          8:17
>0     1G  0 part  /boot
>├─sdb2                                                          8:18
>0     4G  0 part  [SWAP]
>└─sdb3                                                          8:19   0
>293.1G  0 part
>  └─vg_node002-LogVol00 (dm-0)                                253:0    0
>293.1G  0 lvm   /
>sda                                                             8:0    0
>1.8T  0 disk
>└─sda1                                                          8:1    0
>1.8T  0 part
>sdd                                                             8:48   0
>976.6G  0 disk
>└─mpathc (dm-4)                                               253:4    0
>976.6G  0 mpath
>sde                                                             8:64   0
>500G  0 disk
>└─mpathf (dm-5)                                               253:5    0
>500G  0 mpath
>  ├─cb85e6cd--df54--4151--8f3b--7e6d72b7372d-metadata (dm-15) 253:15   0
>512M  0 lvm
>  ├─cb85e6cd--df54--4151--8f3b--7e6d72b7372d-ids (dm-16)      253:16   0
>128M  0 lvm
>  ├─cb85e6cd--df54--4151--8f3b--7e6d72b7372d-leases (dm-18)   253:18
>0     2G  0 lvm
>  ├─cb85e6cd--df54--4151--8f3b--7e6d72b7372d-outbox (dm-20)   253:20   0
>128M  0 lvm
>  ├─cb85e6cd--df54--4151--8f3b--7e6d72b7372d-inbox (dm-21)    253:21   0
>128M  0 lvm
>  └─cb85e6cd--df54--4151--8f3b--7e6d72b7372d-master (dm-22)   253:22
>0     1G  0 lvm
>sdf                                                             8:80   0
>976.6G  0 disk
>└─mpathd (dm-7)                                               253:7    0
>976.6G  0 mpath
>  ├─5e1ca1b6--4706--4c79--8924--b8db741c929f-metadata (dm-14) 253:14   0
>512M  0 lvm
>  ├─5e1ca1b6--4706--4c79--8924--b8db741c929f-ids (dm-17)      253:17   0
>128M  0 lvm
>  ├─5e1ca1b6--4706--4c79--8924--b8db741c929f-leases (dm-19)   253:19
>0     2G  0 lvm
>  ├─5e1ca1b6--4706--4c79--8924--b8db741c929f-outbox (dm-23)   253:23   0
>128M  0 lvm
>  ├─5e1ca1b6--4706--4c79--8924--b8db741c929f-inbox (dm-24)    253:24   0
>128M  0 lvm
>  └─5e1ca1b6--4706--4c79--8924--b8db741c929f-master (dm-25)   253:25
>0     1G  0 lvm
>sdg                                                             8:96   0
>976.6G  0 disk
>└─mpathe (dm-6)                                               253:6    0
>976.6G  0 mpath
>  ├─5f595801--aaa5--42c7--b829--7a34a636407e-metadata (dm-8)  253:8    0
>512M  0 lvm
>  ├─5f595801--aaa5--42c7--b829--7a34a636407e-ids (dm-9)       253:9    0
>128M  0 lvm
>  ├─5f595801--aaa5--42c7--b829--7a34a636407e-leases (dm-10)   253:10
>0     2G  0 lvm
>  ├─5f595801--aaa5--42c7--b829--7a34a636407e-outbox (dm-11)   253:11   0
>128M  0 lvm
>  ├─5f595801--aaa5--42c7--b829--7a34a636407e-inbox (dm-12)    253:12   0
>128M  0 lvm
>  └─5f595801--aaa5--42c7--b829--7a34a636407e-master (dm-13)   253:13
>0     1G  0 lvm
>
>**********************************************************************************
>
>[root at node002 shim]# multipath -v3
>Jan 26 12:46:28 | ram0: device node name blacklisted
>Jan 26 12:46:28 | ram1: device node name blacklisted
>Jan 26 12:46:28 | ram2: device node name blacklisted
>Jan 26 12:46:28 | ram3: device node name blacklisted
>Jan 26 12:46:28 | ram4: device node name blacklisted
>Jan 26 12:46:28 | ram5: device node name blacklisted
>Jan 26 12:46:28 | ram6: device node name blacklisted
>Jan 26 12:46:28 | ram7: device node name blacklisted
>Jan 26 12:46:28 | ram8: device node name blacklisted
>Jan 26 12:46:28 | ram9: device node name blacklisted
>Jan 26 12:46:28 | ram10: device node name blacklisted
>Jan 26 12:46:28 | ram11: device node name blacklisted
>Jan 26 12:46:28 | ram12: device node name blacklisted
>Jan 26 12:46:28 | ram13: device node name blacklisted
>Jan 26 12:46:28 | ram14: device node name blacklisted
>Jan 26 12:46:28 | ram15: device node name blacklisted
>Jan 26 12:46:28 | loop0: device node name blacklisted
>Jan 26 12:46:28 | loop1: device node name blacklisted
>Jan 26 12:46:28 | loop2: device node name blacklisted
>Jan 26 12:46:28 | loop3: device node name blacklisted
>Jan 26 12:46:28 | loop4: device node name blacklisted
>Jan 26 12:46:28 | loop5: device node name blacklisted
>Jan 26 12:46:28 | loop6: device node name blacklisted
>Jan 26 12:46:28 | loop7: device node name blacklisted
>Jan 26 12:46:28 | sdb: not found in pathvec
>Jan 26 12:46:28 | sdb: mask = 0x3f
>Jan 26 12:46:28 | sdb: dev_t = 8:16
>Jan 26 12:46:28 | sdb: size = 625142448
>Jan 26 12:46:28 | sdb: subsystem = scsi
>Jan 26 12:46:28 | sdb: vendor = ATA
>Jan 26 12:46:28 | sdb: product = WDC WD3200AAJS-6
>Jan 26 12:46:28 | sdb: rev = 03.0
>Jan 26 12:46:28 | sdb: h:b:t:l = 10:0:0:0
>Jan 26 12:46:28 | sdb: serial =      WD-WMAV2HM46197
>Jan 26 12:46:28 | sdb: get_state
>Jan 26 12:46:28 | sdb: path checker = directio (config file default)
>Jan 26 12:46:28 | sdb: checker timeout = 30000 ms (sysfs setting)
>Jan 26 12:46:28 | sdb: state = running
>Jan 26 12:46:28 | directio: starting new request
>Jan 26 12:46:28 | directio: io finished 4096/0
>Jan 26 12:46:28 | sdb: state = 3
>Jan 26 12:46:28 | sdb: getuid = /sbin/scsi_id --whitelisted
>--replace-whitespace --device=/dev/%n (config file default)
>Jan 26 12:46:28 | sdb: uid = 1ATA_WDC_WD3200AAJS-60Z0A0_WD-WMAV2HM46197
>(callout)
>Jan 26 12:46:28 | sdb: state = running
>Jan 26 12:46:28 | sdb: detect_prio = 1 (config file default)
>Jan 26 12:46:28 | sdb: prio = const (config file default)
>Jan 26 12:46:28 | sdb: const prio = 1
>Jan 26 12:46:28 | sda: not found in pathvec
>Jan 26 12:46:28 | sda: mask = 0x3f
>Jan 26 12:46:28 | sda: dev_t = 8:0
>Jan 26 12:46:28 | sda: size = 3904897024
>Jan 26 12:46:28 | sda: subsystem = scsi
>Jan 26 12:46:28 | sda: vendor = ATA
>Jan 26 12:46:28 | sda: product = MARVELL Raid VD
>Jan 26 12:46:28 | sda: rev = MV.R
>Jan 26 12:46:28 | sda: h:b:t:l = 0:0:0:0
>Jan 26 12:46:28 | sda: serial = 1c3c8ecf5cf00010
>Jan 26 12:46:28 | sda: get_state
>Jan 26 12:46:28 | sda: path checker = directio (config file default)
>Jan 26 12:46:28 | sda: checker timeout = 30000 ms (sysfs setting)
>Jan 26 12:46:28 | sda: state = running
>Jan 26 12:46:28 | directio: starting new request
>Jan 26 12:46:28 | directio: io finished 4096/0
>Jan 26 12:46:28 | sda: state = 3
>Jan 26 12:46:28 | sda: getuid = /sbin/scsi_id --whitelisted
>--replace-whitespace --device=/dev/%n (config file default)
>Jan 26 12:46:28 | sda: uid = 1ATA_MARVELL_Raid_VD_0_1c3c8ecf5cf00010
>(callout)
>Jan 26 12:46:28 | sda: state = running
>Jan 26 12:46:28 | sda: detect_prio = 1 (config file default)
>Jan 26 12:46:28 | sda: prio = const (config file default)
>Jan 26 12:46:28 | sda: const prio = 1
>Jan 26 12:46:28 | dm-0: device node name blacklisted
>Jan 26 12:46:28 | sdc: not found in pathvec
>Jan 26 12:46:28 | sdc: mask = 0x3f
>Jan 26 12:46:28 | sdc: dev_t = 8:32
>Jan 26 12:46:28 | sdc: size = 0
>Jan 26 12:46:28 | sdc: subsystem = scsi
>Jan 26 12:46:28 | sdc: vendor = Multi
>Jan 26 12:46:28 | sdc: product = Flash Reader
>Jan 26 12:46:28 | sdc: rev = 1.00
>Jan 26 12:46:28 | sdc: h:b:t:l = 12:0:0:0
>Jan 26 12:46:28 | dm-1: device node name blacklisted
>Jan 26 12:46:28 | dm-2: device node name blacklisted
>Jan 26 12:46:28 | dm-3: device node name blacklisted
>Jan 26 12:46:28 | sdd: not found in pathvec
>Jan 26 12:46:28 | sdd: mask = 0x3f
>Jan 26 12:46:28 | sdd: dev_t = 8:48
>Jan 26 12:46:28 | sdd: size = 2048000000
>Jan 26 12:46:28 | sdd: subsystem = scsi
>Jan 26 12:46:28 | sdd: vendor = SHIMI
>Jan 26 12:46:28 | sdd: product = VIRTUAL-DISK
>Jan 26 12:46:28 | sdd: rev = 0001
>Jan 26 12:46:28 | sdd: h:b:t:l = 18:0:0:1
>Jan 26 12:46:28 | sdd: tgt_node_name = pl.mycomp.shimi:node002.target0
>Jan 26 12:46:28 | sdd: serial =                               beaf11
>Jan 26 12:46:28 | sdd: get_state
>Jan 26 12:46:28 | sdd: path checker = directio (config file default)
>Jan 26 12:46:28 | sdd: checker timeout = 30000 ms (sysfs setting)
>Jan 26 12:46:28 | sdd: state = running
>Jan 26 12:46:28 | directio: starting new request
>Jan 26 12:46:28 | directio: io finished 4096/0
>Jan 26 12:46:28 | sdd: state = 3
>Jan 26 12:46:28 | sdd: getuid = /sbin/scsi_id --whitelisted
>--replace-whitespace --device=/dev/%n (config file default)
>Jan 26 12:46:28 | sdd: uid = 1NODE_002_LUN01 (callout)
>Jan 26 12:46:28 | sdd: state = running
>Jan 26 12:46:28 | sdd: detect_prio = 1 (config file default)
>Jan 26 12:46:28 | sdd: prio = const (config file default)
>Jan 26 12:46:28 | sdd: const prio = 1
>Jan 26 12:46:28 | dm-4: device node name blacklisted
>Jan 26 12:46:28 | sde: not found in pathvec
>Jan 26 12:46:28 | sde: mask = 0x3f
>Jan 26 12:46:28 | sde: dev_t = 8:64
>Jan 26 12:46:28 | sde: size = 1048576000
>Jan 26 12:46:28 | sde: subsystem = scsi
>Jan 26 12:46:28 | sde: vendor = SHIMI
>Jan 26 12:46:28 | sde: product = VIRTUAL-DISK
>Jan 26 12:46:28 | sde: rev = 0001
>Jan 26 12:46:28 | sde: h:b:t:l = 19:0:0:1
>Jan 26 12:46:28 | sde: tgt_node_name = pl.mycomp.shimi:manager.target0
>Jan 26 12:46:28 | sde: serial =                               beaf11
>Jan 26 12:46:28 | sde: get_state
>Jan 26 12:46:28 | sde: path checker = directio (config file default)
>Jan 26 12:46:28 | sde: checker timeout = 30000 ms (sysfs setting)
>Jan 26 12:46:28 | sde: state = running
>Jan 26 12:46:28 | directio: starting new request
>Jan 26 12:46:28 | directio: io finished 4096/0
>Jan 26 12:46:28 | sde: state = 3
>Jan 26 12:46:28 | sde: getuid = /sbin/scsi_id --whitelisted
>--replace-whitespace --device=/dev/%n (config file default)
>Jan 26 12:46:28 | sde: uid = 1MANAGER_LUN01 (callout)
>Jan 26 12:46:28 | sde: state = running
>Jan 26 12:46:28 | sde: detect_prio = 1 (config file default)
>Jan 26 12:46:28 | sde: prio = const (config file default)
>Jan 26 12:46:28 | sde: const prio = 1
>Jan 26 12:46:28 | sdf: not found in pathvec
>Jan 26 12:46:28 | sdf: mask = 0x3f
>Jan 26 12:46:28 | sdf: dev_t = 8:80
>Jan 26 12:46:28 | sdf: size = 2048000000
>Jan 26 12:46:28 | sdf: subsystem = scsi
>Jan 26 12:46:28 | sdf: vendor = SHIMI
>Jan 26 12:46:28 | sdf: product = VIRTUAL-DISK
>Jan 26 12:46:28 | sdf: rev = 0001
>Jan 26 12:46:28 | sdf: h:b:t:l = 20:0:0:1
>Jan 26 12:46:28 | sdf: tgt_node_name = pl.mycomp.shimi:node003.target0
>Jan 26 12:46:28 | sdf: serial =                               beaf11
>Jan 26 12:46:28 | sdf: get_state
>Jan 26 12:46:28 | sdf: path checker = directio (config file default)
>Jan 26 12:46:28 | sdf: checker timeout = 30000 ms (sysfs setting)
>Jan 26 12:46:28 | sdf: state = running
>Jan 26 12:46:28 | directio: starting new request
>Jan 26 12:46:28 | directio: io finished 4096/0
>Jan 26 12:46:28 | sdf: state = 3
>Jan 26 12:46:28 | sdf: getuid = /sbin/scsi_id --whitelisted
>--replace-whitespace --device=/dev/%n (config file default)
>Jan 26 12:46:28 | sdf: uid = 1NODE_003_LUN01 (callout)
>Jan 26 12:46:28 | sdf: state = running
>Jan 26 12:46:28 | sdf: detect_prio = 1 (config file default)
>Jan 26 12:46:28 | sdf: prio = const (config file default)
>Jan 26 12:46:28 | sdf: const prio = 1
>Jan 26 12:46:28 | sdg: not found in pathvec
>Jan 26 12:46:28 | sdg: mask = 0x3f
>Jan 26 12:46:28 | sdg: dev_t = 8:96
>Jan 26 12:46:28 | sdg: size = 2048000000
>Jan 26 12:46:28 | sdg: subsystem = scsi
>Jan 26 12:46:28 | sdg: vendor = SHIMI
>Jan 26 12:46:28 | sdg: product = VIRTUAL-DISK
>Jan 26 12:46:28 | sdg: rev = 0001
>Jan 26 12:46:28 | sdg: h:b:t:l = 21:0:0:1
>Jan 26 12:46:28 | sdg: tgt_node_name = pl.mycomp.shimi:node001.target0
>Jan 26 12:46:28 | sdg: serial =                               beaf11
>Jan 26 12:46:28 | sdg: get_state
>Jan 26 12:46:28 | sdg: path checker = directio (config file default)
>Jan 26 12:46:28 | sdg: checker timeout = 30000 ms (sysfs setting)
>Jan 26 12:46:28 | sdg: state = running
>Jan 26 12:46:28 | directio: starting new request
>Jan 26 12:46:28 | directio: io finished 4096/0
>Jan 26 12:46:28 | sdg: state = 3
>Jan 26 12:46:28 | sdg: getuid = /sbin/scsi_id --whitelisted
>--replace-whitespace --device=/dev/%n (config file default)
>Jan 26 12:46:28 | sdg: uid = 1NODE_001_LUN01 (callout)
>Jan 26 12:46:28 | sdg: state = running
>Jan 26 12:46:28 | sdg: detect_prio = 1 (config file default)
>Jan 26 12:46:28 | sdg: prio = const (config file default)
>Jan 26 12:46:28 | sdg: const prio = 1
>Jan 26 12:46:28 | dm-5: device node name blacklisted
>Jan 26 12:46:28 | dm-6: device node name blacklisted
>Jan 26 12:46:28 | dm-7: device node name blacklisted
>Jan 26 12:46:28 | dm-8: device node name blacklisted
>Jan 26 12:46:28 | dm-9: device node name blacklisted
>Jan 26 12:46:28 | dm-10: device node name blacklisted
>Jan 26 12:46:28 | dm-11: device node name blacklisted
>Jan 26 12:46:28 | dm-12: device node name blacklisted
>Jan 26 12:46:28 | dm-13: device node name blacklisted
>Jan 26 12:46:28 | dm-14: device node name blacklisted
>Jan 26 12:46:28 | dm-15: device node name blacklisted
>Jan 26 12:46:28 | dm-16: device node name blacklisted
>Jan 26 12:46:28 | dm-17: device node name blacklisted
>Jan 26 12:46:28 | dm-18: device node name blacklisted
>Jan 26 12:46:28 | dm-19: device node name blacklisted
>Jan 26 12:46:28 | dm-20: device node name blacklisted
>Jan 26 12:46:28 | dm-21: device node name blacklisted
>Jan 26 12:46:28 | dm-22: device node name blacklisted
>Jan 26 12:46:28 | dm-23: device node name blacklisted
>Jan 26 12:46:28 | dm-24: device node name blacklisted
>Jan 26 12:46:28 | dm-25: device node name blacklisted
>===== paths list =====
>uuid                                       hcil     dev dev_t pri dm_st
>chk_st
>1ATA_WDC_WD3200AAJS-60Z0A0_WD-WMAV2HM46197 10:0:0:0 sdb 8:16  1   undef
>ready
>1ATA_MARVELL_Raid_VD_0_1c3c8ecf5cf00010    0:0:0:0  sda 8:0   1   undef
>ready
>                                           12:0:0:0 sdc 8:32  -1  undef
>faulty
>1NODE_002_LUN01                            18:0:0:1 sdd 8:48  1   undef
>ready
>1MANAGER_LUN01                             19:0:0:1 sde 8:64  1   undef
>ready
>1NODE_003_LUN01                            20:0:0:1 sdf 8:80  1   undef
>ready
>1NODE_001_LUN01                            21:0:0:1 sdg 8:96  1   undef
>ready
>Jan 26 12:46:28 | params = 0 0 1 1 round-robin 0 1 1 8:96 1
>Jan 26 12:46:28 | status = 2 0 0 0 1 1 A 0 1 0 8:96 A 0
>Jan 26 12:46:28 | params = 0 0 1 1 round-robin 0 1 1 8:80 1
>Jan 26 12:46:28 | status = 2 0 0 0 1 1 A 0 1 0 8:80 A 0
>Jan 26 12:46:28 | params = 0 0 1 1 round-robin 0 1 1 8:48 1
>Jan 26 12:46:28 | status = 2 0 0 0 1 1 A 0 1 0 8:48 A 0
>Jan 26 12:46:28 | params = 0 0 1 1 round-robin 0 1 1 8:0 1
>Jan 26 12:46:28 | status = 2 0 0 0 1 1 A 0 1 0 8:0 A 0
>Jan 26 12:46:28 | params = 0 0 1 1 round-robin 0 1 1 8:64 1
>Jan 26 12:46:28 | status = 2 0 0 0 1 1 A 0 1 0 8:64 A 0
>Jan 26 12:46:28 | Found matching wwid
>[1ATA_WDC_WD3200AAJS-60Z0A0_WD-WMAV2HM46197] in bindings file. Setting
>alias to mpatha
>Jan 26 12:46:28 | sdb: ownership set to mpatha
>Jan 26 12:46:28 | sdb: not found in pathvec
>Jan 26 12:46:28 | sdb: mask = 0xc
>Jan 26 12:46:28 | sdb: get_state
>Jan 26 12:46:28 | sdb: state = running
>Jan 26 12:46:28 | directio: starting new request
>Jan 26 12:46:28 | directio: io finished 4096/0
>Jan 26 12:46:28 | sdb: state = 3
>Jan 26 12:46:28 | sdb: state = running
>Jan 26 12:46:28 | sdb: const prio = 1
>Jan 26 12:46:28 | mpatha: pgfailover = -1 (internal default)
>Jan 26 12:46:28 | mpatha: pgpolicy = failover (internal default)
>Jan 26 12:46:28 | mpatha: selector = round-robin 0 (internal default)
>Jan 26 12:46:28 | mpatha: features = 0 (internal default)
>Jan 26 12:46:28 | mpatha: hwhandler = 0 (internal default)
>Jan 26 12:46:28 | mpatha: rr_weight = 1 (internal default)
>Jan 26 12:46:28 | mpatha: minio = 1 rq (config file default)
>Jan 26 12:46:28 | mpatha: no_path_retry = -1 (config file default)
>Jan 26 12:46:28 | pg_timeout = NONE (internal default)
>Jan 26 12:46:28 | mpatha: fast_io_fail_tmo = 5 (config file default)
>Jan 26 12:46:28 | mpatha: dev_loss_tmo = 30 (config file default)
>Jan 26 12:46:28 | mpatha: retain_attached_hw_handler = 1 (config file
>default)
>Jan 26 12:46:28 | failed to find rport_id for target10:0:0
>Jan 26 12:46:28 | mpatha: set ACT_CREATE (map does not exist)
>Jan 26 12:46:28 | mpatha: domap (0) failure for create/reload map
>Jan 26 12:46:28 | mpatha: ignoring map
>
>**********************************************************************************
>
>[root at node002 shim]# iscsiadm -m session -o show
>tcp: [6] 192.168.1.12:3260,1 pl.mycomp.shimi:node002.target0
>tcp: [7] 192.168.1.11:3260,1 pl.mycomp.shimi:manager.target0
>tcp: [8] 192.168.1.14:3260,1 pl.mycomp.shimi:node003.target0
>tcp: [9] 192.168.1.13:3260,1 pl.mycomp.shimi:node001.target0
>
>**********************************************************************************
>
>[root at node002 shim]# iptables -L
>Chain INPUT (policy ACCEPT)
>target     prot opt source               destination
>
>Chain FORWARD (policy ACCEPT)
>target     prot opt source               destination
>
>Chain OUTPUT (policy ACCEPT)
>target     prot opt source               destination
>
>**********************************************************************************
>
>[root at node002 shim]# sestatus
>SELinux status:                 disabled

>_______________________________________________
>Users mailing list
>Users at ovirt.org
>http://lists.ovirt.org/mailman/listinfo/users


-- 
Adam Litke


More information about the Users mailing list