[ovirt-users] Does my Storage Domain crashed or is this iSCSI LUN's a problem?
Adam Litke
alitke at redhat.com
Wed Jan 28 13:45:09 UTC 2015
On 26/01/15 13:05 +0100, shimano wrote:
>Hi guys,
>
>I'm trying to run one of my storage domains, which experienced failure.
>Unfortunately, I meet a very nasty error (Storage domain does not exist).
>
>Could someone tell me, how to try to restore this domain?
Could you try moving the host to Maintenance mode and then Activate it
again please. I've encountered situations where vdsm restarts and
engine does not reconnect storage until an Activate action happens.
Let's see if this is your issue.
>
>P.S.
>It's an oVirt 3.4.2-1.el6
>
>**********************************************************************************
>
>/var/log/messages:
>Jan 26 12:48:49 node002 vdsm TaskManager.Task ERROR
>Task=`10d02993-b585-448f-9a50-bd3e8cda7082`::Unexpected error#012Traceback
>(most recent call last):#012 File "/usr/share/vdsm/storage/task.py", line
>873, in _run#012 return fn(*args, **kargs)#012 File
>"/usr/share/vdsm/logUtils.py", line 45, in wrapper#012 res = f(*args,
>**kwargs)#012 File "/usr/share/vdsm/storage/hsm.py", line 2959, in
>getVGInfo#012 return dict(info=self.__getVGsInfo([vgUUID])[0])#012 File
>"/usr/share/vdsm/storage/hsm.py", line 2892, in __getVGsInfo#012 vgList
>= [lvm.getVGbyUUID(vgUUID) for vgUUID in vgUUIDs]#012 File
>"/usr/share/vdsm/storage/lvm.py", line 894, in getVGbyUUID#012 raise
>se.VolumeGroupDoesNotExist("vg_uuid: %s" %
>vgUUID)#012VolumeGroupDoesNotExist: Volume Group does not exist: ('vg_uuid:
>gyaCWf-6VKi-lI9W-JT6H-IZdy-rIsB-hTvZ4O',)
>Jan 26 12:48:49 node002 kernel: device-mapper: table: 253:26: multipath:
>error getting device
>Jan 26 12:48:49 node002 kernel: device-mapper: ioctl: error adding target
>to table
>
>**********************************************************************************
>
>/var/log/vdsm.log:
>Thread-22::ERROR::2015-01-26
>12:43:03,376::sdc::137::Storage.StorageDomainCache::(_findDomain) looking
>for unfetched domain db52e9cb-7306-43fd-aff3-20831bc2bcaf
>Thread-22::ERROR::2015-01-26
>12:43:03,377::sdc::154::Storage.StorageDomainCache::(_findUnfetchedDomain)
>looking for domain db52e9cb-7306-43fd-aff3-20831bc2bcaf
>Thread-22::DEBUG::2015-01-26
>12:43:03,377::lvm::373::OperationMutex::(_reloadvgs) Operation 'lvm reload
>operation' got the operation mutex
>Thread-22::DEBUG::2015-01-26
>12:43:03,378::lvm::296::Storage.Misc.excCmd::(cmd) u'/usr/bin/sudo -n
>/sbin/lvm vgs --config " devices { preferred_names = [\\"^/dev/mapper/\\"]
>ignore_suspended_devices=1 write_cache_state=0 disable_after_error_count=3
>obtain_device_list_from_udev=0 filter = [
>\'a|/dev/mapper/mpathb|/dev/mapper/mpathc|/dev/mapper/mpathd|/dev/mapper/mpathe|/dev/mapper/mpathf|\',
>\'r|.*|\' ] } global { locking_type=1 prioritise_write_locks=1
>wait_for_locks=1 use_lvmetad=0 } backup { retain_min = 50 retain_days =
>0 } " --noheadings --units b --nosuffix --separator | -o
>uuid,name,attr,size,free,extent_size,extent_count,free_count,tags,vg_mda_size,vg_mda_free,lv_count,pv_count,pv_name
>db52e9cb-7306-43fd-aff3-20831bc2bcaf' (cwd None)
>Thread-22::DEBUG::2015-01-26
>12:43:03,462::lvm::296::Storage.Misc.excCmd::(cmd) FAILED: <err> = '
>/dev/mapper/mpathc: Checksum error\n /dev/mapper/mpathc: Checksum error\n
>Volume group "db52e9cb-7306-43fd-aff3-20831bc2bcaf" not found\n Skipping
>volume group db52e9cb-7306-43fd-aff3-20831bc2bcaf\n'; <rc> = 5
>Thread-22::WARNING::2015-01-26
>12:43:03,466::lvm::378::Storage.LVM::(_reloadvgs) lvm vgs failed: 5 [] ['
>/dev/mapper/mpathc: Checksum error', ' /dev/mapper/mpathc: Checksum
>error', ' Volume group "db52e9cb-7306-43fd-aff3-20831bc2bcaf" not found',
>' Skipping volume group db52e9cb-7306-43fd-aff3-20831bc2bcaf']
>Thread-22::DEBUG::2015-01-26
>12:43:03,466::lvm::415::OperationMutex::(_reloadvgs) Operation 'lvm reload
>operation' released the operation mutex
>Thread-22::ERROR::2015-01-26
>12:43:03,477::sdc::143::Storage.StorageDomainCache::(_findDomain) domain
>db52e9cb-7306-43fd-aff3-20831bc2bcaf not found
>Traceback (most recent call last):
> File "/usr/share/vdsm/storage/sdc.py", line 141, in _findDomain
> dom = findMethod(sdUUID)
> File "/usr/share/vdsm/storage/sdc.py", line 171, in _findUnfetchedDomain
> raise se.StorageDomainDoesNotExist(sdUUID)
>StorageDomainDoesNotExist: Storage domain does not exist:
>(u'db52e9cb-7306-43fd-aff3-20831bc2bcaf',)
>Thread-22::ERROR::2015-01-26
>12:43:03,478::domainMonitor::239::Storage.DomainMonitorThread::(_monitorDomain)
>Error while collecting domain db52e9cb-7306-43fd-aff3-20831bc2bcaf
>monitoring information
>Traceback (most recent call last):
> File "/usr/share/vdsm/storage/domainMonitor.py", line 204, in
>_monitorDomain
> self.domain = sdCache.produce(self.sdUUID)
> File "/usr/share/vdsm/storage/sdc.py", line 98, in produce
> domain.getRealDomain()
> File "/usr/share/vdsm/storage/sdc.py", line 52, in getRealDomain
> return self._cache._realProduce(self._sdUUID)
> File "/usr/share/vdsm/storage/sdc.py", line 122, in _realProduce
> domain = self._findDomain(sdUUID)
> File "/usr/share/vdsm/storage/sdc.py", line 141, in _findDomain
> dom = findMethod(sdUUID)
> File "/usr/share/vdsm/storage/sdc.py", line 171, in _findUnfetchedDomain
> raise se.StorageDomainDoesNotExist(sdUUID)
>StorageDomainDoesNotExist: Storage domain does not exist:
>(u'db52e9cb-7306-43fd-aff3-20831bc2bcaf',)
>Thread-13::DEBUG::2015-01-26
>12:43:05,102::task::595::TaskManager.Task::(_updateState)
>Task=`b4e85e37-b216-4d29-a448-0711e370a246`::moving from state init ->
>state preparing
>Thread-13::INFO::2015-01-26
>12:43:05,102::logUtils::44::dispatcher::(wrapper) Run and protect:
>repoStats(options=None)
>Thread-13::INFO::2015-01-26
>12:43:05,103::logUtils::47::dispatcher::(wrapper) Run and protect:
>repoStats, Return response: {u'7969d636-1a02-42ba-a50b-2528765cf3d5':
>{'code': 0, 'version': 0, 'acquired': True, 'delay': '0.000457574',
>'lastCheck': '7.5', 'valid': True},
>u'5e1ca1b6-4706-4c79-8924-b8db741c929f': {'code': 0, 'version': 3,
>'acquired': True, 'delay': '0.00100094', 'lastCheck': '6.3', 'valid':
>True}, u'cb85e6cd-df54-4151-8f3b-7e6d72b7372d': {'code': 0, 'version': 3,
>'acquired': True, 'delay': '0.463061', 'lastCheck': '4.9', 'valid': True},
>u'db52e9cb-7306-43fd-aff3-20831bc2bcaf': {'code': 358, 'version': -1,
>'acquired': False, 'delay': '0', 'lastCheck': '1.6', 'valid': False},
>u'5f595801-aaa5-42c7-b829-7a34a636407e': {'code': 0, 'version': 3,
>'acquired': True, 'delay': '0.000942979', 'lastCheck': '7.9', 'valid':
>True}, u'c1ebd0f8-fa32-4fe3-8569-fb7d4ad8faf4': {'code': 0, 'version': 0,
>'acquired': True, 'delay': '0.000424499', 'lastCheck': '7.3', 'valid':
>True}}
>Thread-13::DEBUG::2015-01-26
>12:43:05,103::task::1185::TaskManager.Task::(prepare)
>Task=`b4e85e37-b216-4d29-a448-0711e370a246`::finished:
>{u'7969d636-1a02-42ba-a50b-2528765cf3d5': {'code': 0, 'version': 0,
>'acquired': True, 'delay': '0.000457574', 'lastCheck': '7.5', 'valid':
>True}, u'5e1ca1b6-4706-4c79-8924-b8db741c929f': {'code': 0, 'version': 3,
>'acquired': True, 'delay': '0.00100094', 'lastCheck': '6.3', 'valid':
>True}, u'cb85e6cd-df54-4151-8f3b-7e6d72b7372d': {'code': 0, 'version': 3,
>'acquired': True, 'delay': '0.463061', 'lastCheck': '4.9', 'valid': True},
>u'db52e9cb-7306-43fd-aff3-20831bc2bcaf': {'code': 358, 'version': -1,
>'acquired': False, 'delay': '0', 'lastCheck': '1.6', 'valid': False},
>u'5f595801-aaa5-42c7-b829-7a34a636407e': {'code': 0, 'version': 3,
>'acquired': True, 'delay': '0.000942979', 'lastCheck': '7.9', 'valid':
>True}, u'c1ebd0f8-fa32-4fe3-8569-fb7d4ad8faf4': {'code': 0, 'version': 0,
>'acquired': True, 'delay': '0.000424499', 'lastCheck': '7.3', 'valid':
>True}}
>
>**********************************************************************************
>
>[root at node002 shim]# multipath -ll
>mpathe (1NODE_001_LUN01) dm-6 SHIMI,VIRTUAL-DISK
>size=977G features='0' hwhandler='0' wp=rw
>`-+- policy='round-robin 0' prio=1 status=active
> `- 21:0:0:1 sdg 8:96 active ready running
>mpathd (1NODE_003_LUN01) dm-7 SHIMI,VIRTUAL-DISK
>size=977G features='0' hwhandler='0' wp=rw
>`-+- policy='round-robin 0' prio=1 status=active
> `- 20:0:0:1 sdf 8:80 active ready running
>mpathc (1NODE_002_LUN01) dm-4 SHIMI,VIRTUAL-DISK
>size=977G features='0' hwhandler='0' wp=rw
>`-+- policy='round-robin 0' prio=1 status=active
> `- 18:0:0:1 sdd 8:48 active ready running
>mpathb (1ATA_MARVELL_Raid_VD_0_1c3c8ecf5cf00010) dm-1 ATA,MARVELL Raid VD
>size=1.8T features='0' hwhandler='0' wp=rw
>`-+- policy='round-robin 0' prio=1 status=active
> `- 0:0:0:0 sda 8:0 active ready running
>mpathf (1MANAGER_LUN01) dm-5 SHIMI,VIRTUAL-DISK
>size=500G features='0' hwhandler='0' wp=rw
>`-+- policy='round-robin 0' prio=1 status=active
> `- 19:0:0:1 sde 8:64 active ready running
>
>**********************************************************************************
>
>[root at node002 shim]# lsblk
>NAME MAJ:MIN RM
>SIZE RO TYPE MOUNTPOINT
>sdb 8:16 0
>298.1G 0 disk
>├─sdb1 8:17
>0 1G 0 part /boot
>├─sdb2 8:18
>0 4G 0 part [SWAP]
>└─sdb3 8:19 0
>293.1G 0 part
> └─vg_node002-LogVol00 (dm-0) 253:0 0
>293.1G 0 lvm /
>sda 8:0 0
>1.8T 0 disk
>└─sda1 8:1 0
>1.8T 0 part
>sdd 8:48 0
>976.6G 0 disk
>└─mpathc (dm-4) 253:4 0
>976.6G 0 mpath
>sde 8:64 0
>500G 0 disk
>└─mpathf (dm-5) 253:5 0
>500G 0 mpath
> ├─cb85e6cd--df54--4151--8f3b--7e6d72b7372d-metadata (dm-15) 253:15 0
>512M 0 lvm
> ├─cb85e6cd--df54--4151--8f3b--7e6d72b7372d-ids (dm-16) 253:16 0
>128M 0 lvm
> ├─cb85e6cd--df54--4151--8f3b--7e6d72b7372d-leases (dm-18) 253:18
>0 2G 0 lvm
> ├─cb85e6cd--df54--4151--8f3b--7e6d72b7372d-outbox (dm-20) 253:20 0
>128M 0 lvm
> ├─cb85e6cd--df54--4151--8f3b--7e6d72b7372d-inbox (dm-21) 253:21 0
>128M 0 lvm
> └─cb85e6cd--df54--4151--8f3b--7e6d72b7372d-master (dm-22) 253:22
>0 1G 0 lvm
>sdf 8:80 0
>976.6G 0 disk
>└─mpathd (dm-7) 253:7 0
>976.6G 0 mpath
> ├─5e1ca1b6--4706--4c79--8924--b8db741c929f-metadata (dm-14) 253:14 0
>512M 0 lvm
> ├─5e1ca1b6--4706--4c79--8924--b8db741c929f-ids (dm-17) 253:17 0
>128M 0 lvm
> ├─5e1ca1b6--4706--4c79--8924--b8db741c929f-leases (dm-19) 253:19
>0 2G 0 lvm
> ├─5e1ca1b6--4706--4c79--8924--b8db741c929f-outbox (dm-23) 253:23 0
>128M 0 lvm
> ├─5e1ca1b6--4706--4c79--8924--b8db741c929f-inbox (dm-24) 253:24 0
>128M 0 lvm
> └─5e1ca1b6--4706--4c79--8924--b8db741c929f-master (dm-25) 253:25
>0 1G 0 lvm
>sdg 8:96 0
>976.6G 0 disk
>└─mpathe (dm-6) 253:6 0
>976.6G 0 mpath
> ├─5f595801--aaa5--42c7--b829--7a34a636407e-metadata (dm-8) 253:8 0
>512M 0 lvm
> ├─5f595801--aaa5--42c7--b829--7a34a636407e-ids (dm-9) 253:9 0
>128M 0 lvm
> ├─5f595801--aaa5--42c7--b829--7a34a636407e-leases (dm-10) 253:10
>0 2G 0 lvm
> ├─5f595801--aaa5--42c7--b829--7a34a636407e-outbox (dm-11) 253:11 0
>128M 0 lvm
> ├─5f595801--aaa5--42c7--b829--7a34a636407e-inbox (dm-12) 253:12 0
>128M 0 lvm
> └─5f595801--aaa5--42c7--b829--7a34a636407e-master (dm-13) 253:13
>0 1G 0 lvm
>
>**********************************************************************************
>
>[root at node002 shim]# multipath -v3
>Jan 26 12:46:28 | ram0: device node name blacklisted
>Jan 26 12:46:28 | ram1: device node name blacklisted
>Jan 26 12:46:28 | ram2: device node name blacklisted
>Jan 26 12:46:28 | ram3: device node name blacklisted
>Jan 26 12:46:28 | ram4: device node name blacklisted
>Jan 26 12:46:28 | ram5: device node name blacklisted
>Jan 26 12:46:28 | ram6: device node name blacklisted
>Jan 26 12:46:28 | ram7: device node name blacklisted
>Jan 26 12:46:28 | ram8: device node name blacklisted
>Jan 26 12:46:28 | ram9: device node name blacklisted
>Jan 26 12:46:28 | ram10: device node name blacklisted
>Jan 26 12:46:28 | ram11: device node name blacklisted
>Jan 26 12:46:28 | ram12: device node name blacklisted
>Jan 26 12:46:28 | ram13: device node name blacklisted
>Jan 26 12:46:28 | ram14: device node name blacklisted
>Jan 26 12:46:28 | ram15: device node name blacklisted
>Jan 26 12:46:28 | loop0: device node name blacklisted
>Jan 26 12:46:28 | loop1: device node name blacklisted
>Jan 26 12:46:28 | loop2: device node name blacklisted
>Jan 26 12:46:28 | loop3: device node name blacklisted
>Jan 26 12:46:28 | loop4: device node name blacklisted
>Jan 26 12:46:28 | loop5: device node name blacklisted
>Jan 26 12:46:28 | loop6: device node name blacklisted
>Jan 26 12:46:28 | loop7: device node name blacklisted
>Jan 26 12:46:28 | sdb: not found in pathvec
>Jan 26 12:46:28 | sdb: mask = 0x3f
>Jan 26 12:46:28 | sdb: dev_t = 8:16
>Jan 26 12:46:28 | sdb: size = 625142448
>Jan 26 12:46:28 | sdb: subsystem = scsi
>Jan 26 12:46:28 | sdb: vendor = ATA
>Jan 26 12:46:28 | sdb: product = WDC WD3200AAJS-6
>Jan 26 12:46:28 | sdb: rev = 03.0
>Jan 26 12:46:28 | sdb: h:b:t:l = 10:0:0:0
>Jan 26 12:46:28 | sdb: serial = WD-WMAV2HM46197
>Jan 26 12:46:28 | sdb: get_state
>Jan 26 12:46:28 | sdb: path checker = directio (config file default)
>Jan 26 12:46:28 | sdb: checker timeout = 30000 ms (sysfs setting)
>Jan 26 12:46:28 | sdb: state = running
>Jan 26 12:46:28 | directio: starting new request
>Jan 26 12:46:28 | directio: io finished 4096/0
>Jan 26 12:46:28 | sdb: state = 3
>Jan 26 12:46:28 | sdb: getuid = /sbin/scsi_id --whitelisted
>--replace-whitespace --device=/dev/%n (config file default)
>Jan 26 12:46:28 | sdb: uid = 1ATA_WDC_WD3200AAJS-60Z0A0_WD-WMAV2HM46197
>(callout)
>Jan 26 12:46:28 | sdb: state = running
>Jan 26 12:46:28 | sdb: detect_prio = 1 (config file default)
>Jan 26 12:46:28 | sdb: prio = const (config file default)
>Jan 26 12:46:28 | sdb: const prio = 1
>Jan 26 12:46:28 | sda: not found in pathvec
>Jan 26 12:46:28 | sda: mask = 0x3f
>Jan 26 12:46:28 | sda: dev_t = 8:0
>Jan 26 12:46:28 | sda: size = 3904897024
>Jan 26 12:46:28 | sda: subsystem = scsi
>Jan 26 12:46:28 | sda: vendor = ATA
>Jan 26 12:46:28 | sda: product = MARVELL Raid VD
>Jan 26 12:46:28 | sda: rev = MV.R
>Jan 26 12:46:28 | sda: h:b:t:l = 0:0:0:0
>Jan 26 12:46:28 | sda: serial = 1c3c8ecf5cf00010
>Jan 26 12:46:28 | sda: get_state
>Jan 26 12:46:28 | sda: path checker = directio (config file default)
>Jan 26 12:46:28 | sda: checker timeout = 30000 ms (sysfs setting)
>Jan 26 12:46:28 | sda: state = running
>Jan 26 12:46:28 | directio: starting new request
>Jan 26 12:46:28 | directio: io finished 4096/0
>Jan 26 12:46:28 | sda: state = 3
>Jan 26 12:46:28 | sda: getuid = /sbin/scsi_id --whitelisted
>--replace-whitespace --device=/dev/%n (config file default)
>Jan 26 12:46:28 | sda: uid = 1ATA_MARVELL_Raid_VD_0_1c3c8ecf5cf00010
>(callout)
>Jan 26 12:46:28 | sda: state = running
>Jan 26 12:46:28 | sda: detect_prio = 1 (config file default)
>Jan 26 12:46:28 | sda: prio = const (config file default)
>Jan 26 12:46:28 | sda: const prio = 1
>Jan 26 12:46:28 | dm-0: device node name blacklisted
>Jan 26 12:46:28 | sdc: not found in pathvec
>Jan 26 12:46:28 | sdc: mask = 0x3f
>Jan 26 12:46:28 | sdc: dev_t = 8:32
>Jan 26 12:46:28 | sdc: size = 0
>Jan 26 12:46:28 | sdc: subsystem = scsi
>Jan 26 12:46:28 | sdc: vendor = Multi
>Jan 26 12:46:28 | sdc: product = Flash Reader
>Jan 26 12:46:28 | sdc: rev = 1.00
>Jan 26 12:46:28 | sdc: h:b:t:l = 12:0:0:0
>Jan 26 12:46:28 | dm-1: device node name blacklisted
>Jan 26 12:46:28 | dm-2: device node name blacklisted
>Jan 26 12:46:28 | dm-3: device node name blacklisted
>Jan 26 12:46:28 | sdd: not found in pathvec
>Jan 26 12:46:28 | sdd: mask = 0x3f
>Jan 26 12:46:28 | sdd: dev_t = 8:48
>Jan 26 12:46:28 | sdd: size = 2048000000
>Jan 26 12:46:28 | sdd: subsystem = scsi
>Jan 26 12:46:28 | sdd: vendor = SHIMI
>Jan 26 12:46:28 | sdd: product = VIRTUAL-DISK
>Jan 26 12:46:28 | sdd: rev = 0001
>Jan 26 12:46:28 | sdd: h:b:t:l = 18:0:0:1
>Jan 26 12:46:28 | sdd: tgt_node_name = pl.mycomp.shimi:node002.target0
>Jan 26 12:46:28 | sdd: serial = beaf11
>Jan 26 12:46:28 | sdd: get_state
>Jan 26 12:46:28 | sdd: path checker = directio (config file default)
>Jan 26 12:46:28 | sdd: checker timeout = 30000 ms (sysfs setting)
>Jan 26 12:46:28 | sdd: state = running
>Jan 26 12:46:28 | directio: starting new request
>Jan 26 12:46:28 | directio: io finished 4096/0
>Jan 26 12:46:28 | sdd: state = 3
>Jan 26 12:46:28 | sdd: getuid = /sbin/scsi_id --whitelisted
>--replace-whitespace --device=/dev/%n (config file default)
>Jan 26 12:46:28 | sdd: uid = 1NODE_002_LUN01 (callout)
>Jan 26 12:46:28 | sdd: state = running
>Jan 26 12:46:28 | sdd: detect_prio = 1 (config file default)
>Jan 26 12:46:28 | sdd: prio = const (config file default)
>Jan 26 12:46:28 | sdd: const prio = 1
>Jan 26 12:46:28 | dm-4: device node name blacklisted
>Jan 26 12:46:28 | sde: not found in pathvec
>Jan 26 12:46:28 | sde: mask = 0x3f
>Jan 26 12:46:28 | sde: dev_t = 8:64
>Jan 26 12:46:28 | sde: size = 1048576000
>Jan 26 12:46:28 | sde: subsystem = scsi
>Jan 26 12:46:28 | sde: vendor = SHIMI
>Jan 26 12:46:28 | sde: product = VIRTUAL-DISK
>Jan 26 12:46:28 | sde: rev = 0001
>Jan 26 12:46:28 | sde: h:b:t:l = 19:0:0:1
>Jan 26 12:46:28 | sde: tgt_node_name = pl.mycomp.shimi:manager.target0
>Jan 26 12:46:28 | sde: serial = beaf11
>Jan 26 12:46:28 | sde: get_state
>Jan 26 12:46:28 | sde: path checker = directio (config file default)
>Jan 26 12:46:28 | sde: checker timeout = 30000 ms (sysfs setting)
>Jan 26 12:46:28 | sde: state = running
>Jan 26 12:46:28 | directio: starting new request
>Jan 26 12:46:28 | directio: io finished 4096/0
>Jan 26 12:46:28 | sde: state = 3
>Jan 26 12:46:28 | sde: getuid = /sbin/scsi_id --whitelisted
>--replace-whitespace --device=/dev/%n (config file default)
>Jan 26 12:46:28 | sde: uid = 1MANAGER_LUN01 (callout)
>Jan 26 12:46:28 | sde: state = running
>Jan 26 12:46:28 | sde: detect_prio = 1 (config file default)
>Jan 26 12:46:28 | sde: prio = const (config file default)
>Jan 26 12:46:28 | sde: const prio = 1
>Jan 26 12:46:28 | sdf: not found in pathvec
>Jan 26 12:46:28 | sdf: mask = 0x3f
>Jan 26 12:46:28 | sdf: dev_t = 8:80
>Jan 26 12:46:28 | sdf: size = 2048000000
>Jan 26 12:46:28 | sdf: subsystem = scsi
>Jan 26 12:46:28 | sdf: vendor = SHIMI
>Jan 26 12:46:28 | sdf: product = VIRTUAL-DISK
>Jan 26 12:46:28 | sdf: rev = 0001
>Jan 26 12:46:28 | sdf: h:b:t:l = 20:0:0:1
>Jan 26 12:46:28 | sdf: tgt_node_name = pl.mycomp.shimi:node003.target0
>Jan 26 12:46:28 | sdf: serial = beaf11
>Jan 26 12:46:28 | sdf: get_state
>Jan 26 12:46:28 | sdf: path checker = directio (config file default)
>Jan 26 12:46:28 | sdf: checker timeout = 30000 ms (sysfs setting)
>Jan 26 12:46:28 | sdf: state = running
>Jan 26 12:46:28 | directio: starting new request
>Jan 26 12:46:28 | directio: io finished 4096/0
>Jan 26 12:46:28 | sdf: state = 3
>Jan 26 12:46:28 | sdf: getuid = /sbin/scsi_id --whitelisted
>--replace-whitespace --device=/dev/%n (config file default)
>Jan 26 12:46:28 | sdf: uid = 1NODE_003_LUN01 (callout)
>Jan 26 12:46:28 | sdf: state = running
>Jan 26 12:46:28 | sdf: detect_prio = 1 (config file default)
>Jan 26 12:46:28 | sdf: prio = const (config file default)
>Jan 26 12:46:28 | sdf: const prio = 1
>Jan 26 12:46:28 | sdg: not found in pathvec
>Jan 26 12:46:28 | sdg: mask = 0x3f
>Jan 26 12:46:28 | sdg: dev_t = 8:96
>Jan 26 12:46:28 | sdg: size = 2048000000
>Jan 26 12:46:28 | sdg: subsystem = scsi
>Jan 26 12:46:28 | sdg: vendor = SHIMI
>Jan 26 12:46:28 | sdg: product = VIRTUAL-DISK
>Jan 26 12:46:28 | sdg: rev = 0001
>Jan 26 12:46:28 | sdg: h:b:t:l = 21:0:0:1
>Jan 26 12:46:28 | sdg: tgt_node_name = pl.mycomp.shimi:node001.target0
>Jan 26 12:46:28 | sdg: serial = beaf11
>Jan 26 12:46:28 | sdg: get_state
>Jan 26 12:46:28 | sdg: path checker = directio (config file default)
>Jan 26 12:46:28 | sdg: checker timeout = 30000 ms (sysfs setting)
>Jan 26 12:46:28 | sdg: state = running
>Jan 26 12:46:28 | directio: starting new request
>Jan 26 12:46:28 | directio: io finished 4096/0
>Jan 26 12:46:28 | sdg: state = 3
>Jan 26 12:46:28 | sdg: getuid = /sbin/scsi_id --whitelisted
>--replace-whitespace --device=/dev/%n (config file default)
>Jan 26 12:46:28 | sdg: uid = 1NODE_001_LUN01 (callout)
>Jan 26 12:46:28 | sdg: state = running
>Jan 26 12:46:28 | sdg: detect_prio = 1 (config file default)
>Jan 26 12:46:28 | sdg: prio = const (config file default)
>Jan 26 12:46:28 | sdg: const prio = 1
>Jan 26 12:46:28 | dm-5: device node name blacklisted
>Jan 26 12:46:28 | dm-6: device node name blacklisted
>Jan 26 12:46:28 | dm-7: device node name blacklisted
>Jan 26 12:46:28 | dm-8: device node name blacklisted
>Jan 26 12:46:28 | dm-9: device node name blacklisted
>Jan 26 12:46:28 | dm-10: device node name blacklisted
>Jan 26 12:46:28 | dm-11: device node name blacklisted
>Jan 26 12:46:28 | dm-12: device node name blacklisted
>Jan 26 12:46:28 | dm-13: device node name blacklisted
>Jan 26 12:46:28 | dm-14: device node name blacklisted
>Jan 26 12:46:28 | dm-15: device node name blacklisted
>Jan 26 12:46:28 | dm-16: device node name blacklisted
>Jan 26 12:46:28 | dm-17: device node name blacklisted
>Jan 26 12:46:28 | dm-18: device node name blacklisted
>Jan 26 12:46:28 | dm-19: device node name blacklisted
>Jan 26 12:46:28 | dm-20: device node name blacklisted
>Jan 26 12:46:28 | dm-21: device node name blacklisted
>Jan 26 12:46:28 | dm-22: device node name blacklisted
>Jan 26 12:46:28 | dm-23: device node name blacklisted
>Jan 26 12:46:28 | dm-24: device node name blacklisted
>Jan 26 12:46:28 | dm-25: device node name blacklisted
>===== paths list =====
>uuid hcil dev dev_t pri dm_st
>chk_st
>1ATA_WDC_WD3200AAJS-60Z0A0_WD-WMAV2HM46197 10:0:0:0 sdb 8:16 1 undef
>ready
>1ATA_MARVELL_Raid_VD_0_1c3c8ecf5cf00010 0:0:0:0 sda 8:0 1 undef
>ready
> 12:0:0:0 sdc 8:32 -1 undef
>faulty
>1NODE_002_LUN01 18:0:0:1 sdd 8:48 1 undef
>ready
>1MANAGER_LUN01 19:0:0:1 sde 8:64 1 undef
>ready
>1NODE_003_LUN01 20:0:0:1 sdf 8:80 1 undef
>ready
>1NODE_001_LUN01 21:0:0:1 sdg 8:96 1 undef
>ready
>Jan 26 12:46:28 | params = 0 0 1 1 round-robin 0 1 1 8:96 1
>Jan 26 12:46:28 | status = 2 0 0 0 1 1 A 0 1 0 8:96 A 0
>Jan 26 12:46:28 | params = 0 0 1 1 round-robin 0 1 1 8:80 1
>Jan 26 12:46:28 | status = 2 0 0 0 1 1 A 0 1 0 8:80 A 0
>Jan 26 12:46:28 | params = 0 0 1 1 round-robin 0 1 1 8:48 1
>Jan 26 12:46:28 | status = 2 0 0 0 1 1 A 0 1 0 8:48 A 0
>Jan 26 12:46:28 | params = 0 0 1 1 round-robin 0 1 1 8:0 1
>Jan 26 12:46:28 | status = 2 0 0 0 1 1 A 0 1 0 8:0 A 0
>Jan 26 12:46:28 | params = 0 0 1 1 round-robin 0 1 1 8:64 1
>Jan 26 12:46:28 | status = 2 0 0 0 1 1 A 0 1 0 8:64 A 0
>Jan 26 12:46:28 | Found matching wwid
>[1ATA_WDC_WD3200AAJS-60Z0A0_WD-WMAV2HM46197] in bindings file. Setting
>alias to mpatha
>Jan 26 12:46:28 | sdb: ownership set to mpatha
>Jan 26 12:46:28 | sdb: not found in pathvec
>Jan 26 12:46:28 | sdb: mask = 0xc
>Jan 26 12:46:28 | sdb: get_state
>Jan 26 12:46:28 | sdb: state = running
>Jan 26 12:46:28 | directio: starting new request
>Jan 26 12:46:28 | directio: io finished 4096/0
>Jan 26 12:46:28 | sdb: state = 3
>Jan 26 12:46:28 | sdb: state = running
>Jan 26 12:46:28 | sdb: const prio = 1
>Jan 26 12:46:28 | mpatha: pgfailover = -1 (internal default)
>Jan 26 12:46:28 | mpatha: pgpolicy = failover (internal default)
>Jan 26 12:46:28 | mpatha: selector = round-robin 0 (internal default)
>Jan 26 12:46:28 | mpatha: features = 0 (internal default)
>Jan 26 12:46:28 | mpatha: hwhandler = 0 (internal default)
>Jan 26 12:46:28 | mpatha: rr_weight = 1 (internal default)
>Jan 26 12:46:28 | mpatha: minio = 1 rq (config file default)
>Jan 26 12:46:28 | mpatha: no_path_retry = -1 (config file default)
>Jan 26 12:46:28 | pg_timeout = NONE (internal default)
>Jan 26 12:46:28 | mpatha: fast_io_fail_tmo = 5 (config file default)
>Jan 26 12:46:28 | mpatha: dev_loss_tmo = 30 (config file default)
>Jan 26 12:46:28 | mpatha: retain_attached_hw_handler = 1 (config file
>default)
>Jan 26 12:46:28 | failed to find rport_id for target10:0:0
>Jan 26 12:46:28 | mpatha: set ACT_CREATE (map does not exist)
>Jan 26 12:46:28 | mpatha: domap (0) failure for create/reload map
>Jan 26 12:46:28 | mpatha: ignoring map
>
>**********************************************************************************
>
>[root at node002 shim]# iscsiadm -m session -o show
>tcp: [6] 192.168.1.12:3260,1 pl.mycomp.shimi:node002.target0
>tcp: [7] 192.168.1.11:3260,1 pl.mycomp.shimi:manager.target0
>tcp: [8] 192.168.1.14:3260,1 pl.mycomp.shimi:node003.target0
>tcp: [9] 192.168.1.13:3260,1 pl.mycomp.shimi:node001.target0
>
>**********************************************************************************
>
>[root at node002 shim]# iptables -L
>Chain INPUT (policy ACCEPT)
>target prot opt source destination
>
>Chain FORWARD (policy ACCEPT)
>target prot opt source destination
>
>Chain OUTPUT (policy ACCEPT)
>target prot opt source destination
>
>**********************************************************************************
>
>[root at node002 shim]# sestatus
>SELinux status: disabled
>_______________________________________________
>Users mailing list
>Users at ovirt.org
>http://lists.ovirt.org/mailman/listinfo/users
--
Adam Litke
More information about the Users
mailing list