[ovirt-users] Does my Storage Domain crashed or is this iSCSI LUN's a problem?

shimano shimano at go2.pl
Tue Jan 27 08:43:59 UTC 2015


Hi guys,

I'm trying to run one of my storage domains, which experienced failure.
Unfortunately, I meet a very nasty error (Storage domain does not exist).

Could someone tell me, how to try to restore this domain?

P.S.
It's an oVirt 3.4.2-1.el6

******************************
****************************************************

/var/log/messages:
Jan 26 12:48:49 node002 vdsm TaskManager.Task ERROR
Task=`10d02993-b585-448f-9a50-bd3e8cda7082`::Unexpected error#012Traceback
(most recent call last):#012  File "/usr/share/vdsm/storage/task.py", line
873, in _run#012    return fn(*args, **kargs)#012  File
"/usr/share/vdsm/logUtils.py", line 45, in wrapper#012    res = f(*args,
**kwargs)#012  File "/usr/share/vdsm/storage/hsm.py", line 2959, in
getVGInfo#012    return dict(info=self.__getVGsInfo([vgUUID])[0])#012  File
"/usr/share/vdsm/storage/hsm.py", line 2892, in __getVGsInfo#012    vgList
= [lvm.getVGbyUUID(vgUUID) for vgUUID in vgUUIDs]#012  File
"/usr/share/vdsm/storage/lvm.py", line 894, in getVGbyUUID#012    raise
se.VolumeGroupDoesNotExist("vg_uuid: %s" %
vgUUID)#012VolumeGroupDoesNotExist: Volume Group does not exist: ('vg_uuid:
gyaCWf-6VKi-lI9W-JT6H-IZdy-rIsB-hTvZ4O',)
Jan 26 12:48:49 node002 kernel: device-mapper: table: 253:26: multipath:
error getting device
Jan 26 12:48:49 node002 kernel: device-mapper: ioctl: error adding target
to table

**********************************************************************************

/var/log/vdsm.log:
Thread-22::ERROR::2015-01-26
12:43:03,376::sdc::137::Storage.StorageDomainCache::(_findDomain) looking
for unfetched domain db52e9cb-7306-43fd-aff3-20831bc2bcaf
Thread-22::ERROR::2015-01-26
12:43:03,377::sdc::154::Storage.StorageDomainCache::(_findUnfetchedDomain)
looking for domain db52e9cb-7306-43fd-aff3-20831bc2bcaf
Thread-22::DEBUG::2015-01-26
12:43:03,377::lvm::373::OperationMutex::(_reloadvgs) Operation 'lvm reload
operation' got the operation mutex
Thread-22::DEBUG::2015-01-26
12:43:03,378::lvm::296::Storage.Misc.excCmd::(cmd) u'/usr/bin/sudo -n
/sbin/lvm vgs --config " devices { preferred_names = [\\"^/dev/mapper/\\"]
ignore_suspended_devices=1 write_cache_state=0 disable_after_error_count=3
obtain_device_list_from_udev=0 filter = [
\'a|/dev/mapper/mpathb|/dev/mapper/mpathc|/dev/mapper/mpathd|/dev/mapper/mpathe|/dev/mapper/mpathf|\',
\'r|.*|\' ] }  global {  locking_type=1  prioritise_write_locks=1
wait_for_locks=1  use_lvmetad=0 }  backup {  retain_min = 50  retain_days =
0 } " --noheadings --units b --nosuffix --separator | -o
uuid,name,attr,size,free,extent_size,extent_count,free_count,tags,vg_mda_size,vg_mda_free,lv_count,pv_count,pv_name
db52e9cb-7306-43fd-aff3-20831bc2bcaf' (cwd None)
Thread-22::DEBUG::2015-01-26
12:43:03,462::lvm::296::Storage.Misc.excCmd::(cmd) FAILED: <err> = '
/dev/mapper/mpathc: Checksum error\n  /dev/mapper/mpathc: Checksum error\n
Volume group "db52e9cb-7306-43fd-aff3-20831bc2bcaf" not found\n  Skipping
volume group db52e9cb-7306-43fd-aff3-20831bc2bcaf\n'; <rc> = 5
Thread-22::WARNING::2015-01-26
12:43:03,466::lvm::378::Storage.LVM::(_reloadvgs) lvm vgs failed: 5 [] ['
/dev/mapper/mpathc: Checksum error', '  /dev/mapper/mpathc: Checksum
error', '  Volume group "db52e9cb-7306-43fd-aff3-20831bc2bcaf" not found',
'  Skipping volume group db52e9cb-7306-43fd-aff3-20831bc2bcaf']
Thread-22::DEBUG::2015-01-26
12:43:03,466::lvm::415::OperationMutex::(_reloadvgs) Operation 'lvm reload
operation' released the operation mutex
Thread-22::ERROR::2015-01-26
12:43:03,477::sdc::143::Storage.StorageDomainCache::(_findDomain) domain
db52e9cb-7306-43fd-aff3-20831bc2bcaf not found
Traceback (most recent call last):
  File "/usr/share/vdsm/storage/sdc.py", line 141, in _findDomain
    dom = findMethod(sdUUID)
  File "/usr/share/vdsm/storage/sdc.py", line 171, in _findUnfetchedDomain
    raise se.StorageDomainDoesNotExist(sdUUID)
StorageDomainDoesNotExist: Storage domain does not exist:
(u'db52e9cb-7306-43fd-aff3-20831bc2bcaf',)
Thread-22::ERROR::2015-01-26
12:43:03,478::domainMonitor::239::Storage.DomainMonitorThread::(_monitorDomain)
Error while collecting domain db52e9cb-7306-43fd-aff3-20831bc2bcaf
monitoring information
Traceback (most recent call last):
  File "/usr/share/vdsm/storage/domainMonitor.py", line 204, in
_monitorDomain
    self.domain = sdCache.produce(self.sdUUID)
  File "/usr/share/vdsm/storage/sdc.py", line 98, in produce
    domain.getRealDomain()
  File "/usr/share/vdsm/storage/sdc.py", line 52, in getRealDomain
    return self._cache._realProduce(self._sdUUID)
  File "/usr/share/vdsm/storage/sdc.py", line 122, in _realProduce
    domain = self._findDomain(sdUUID)
  File "/usr/share/vdsm/storage/sdc.py", line 141, in _findDomain
    dom = findMethod(sdUUID)
  File "/usr/share/vdsm/storage/sdc.py", line 171, in _findUnfetchedDomain
    raise se.StorageDomainDoesNotExist(sdUUID)
StorageDomainDoesNotExist: Storage domain does not exist:
(u'db52e9cb-7306-43fd-aff3-20831bc2bcaf',)
Thread-13::DEBUG::2015-01-26
12:43:05,102::task::595::TaskManager.Task::(_updateState)
Task=`b4e85e37-b216-4d29-a448-0711e370a246`::moving from state init ->
state preparing
Thread-13::INFO::2015-01-26
12:43:05,102::logUtils::44::dispatcher::(wrapper) Run and protect:
repoStats(options=None)
Thread-13::INFO::2015-01-26
12:43:05,103::logUtils::47::dispatcher::(wrapper) Run and protect:
repoStats, Return response: {u'7969d636-1a02-42ba-a50b-2528765cf3d5':
{'code': 0, 'version': 0, 'acquired': True, 'delay': '0.000457574',
'lastCheck': '7.5', 'valid': True},
u'5e1ca1b6-4706-4c79-8924-b8db741c929f': {'code': 0, 'version': 3,
'acquired': True, 'delay': '0.00100094', 'lastCheck': '6.3', 'valid':
True}, u'cb85e6cd-df54-4151-8f3b-7e6d72b7372d': {'code': 0, 'version': 3,
'acquired': True, 'delay': '0.463061', 'lastCheck': '4.9', 'valid': True},
u'db52e9cb-7306-43fd-aff3-20831bc2bcaf': {'code': 358, 'version': -1,
'acquired': False, 'delay': '0', 'lastCheck': '1.6', 'valid': False},
u'5f595801-aaa5-42c7-b829-7a34a636407e': {'code': 0, 'version': 3,
'acquired': True, 'delay': '0.000942979', 'lastCheck': '7.9', 'valid':
True}, u'c1ebd0f8-fa32-4fe3-8569-fb7d4ad8faf4': {'code': 0, 'version': 0,
'acquired': True, 'delay': '0.000424499', 'lastCheck': '7.3', 'valid':
True}}
Thread-13::DEBUG::2015-01-26
12:43:05,103::task::1185::TaskManager.Task::(prepare)
Task=`b4e85e37-b216-4d29-a448-0711e370a246`::finished:
{u'7969d636-1a02-42ba-a50b-2528765cf3d5': {'code': 0, 'version': 0,
'acquired': True, 'delay': '0.000457574', 'lastCheck': '7.5', 'valid':
True}, u'5e1ca1b6-4706-4c79-8924-b8db741c929f': {'code': 0, 'version': 3,
'acquired': True, 'delay': '0.00100094', 'lastCheck': '6.3', 'valid':
True}, u'cb85e6cd-df54-4151-8f3b-7e6d72b7372d': {'code': 0, 'version': 3,
'acquired': True, 'delay': '0.463061', 'lastCheck': '4.9', 'valid': True},
u'db52e9cb-7306-43fd-aff3-20831bc2bcaf': {'code': 358, 'version': -1,
'acquired': False, 'delay': '0', 'lastCheck': '1.6', 'valid': False},
u'5f595801-aaa5-42c7-b829-7a34a636407e': {'code': 0, 'version': 3,
'acquired': True, 'delay': '0.000942979', 'lastCheck': '7.9', 'valid':
True}, u'c1ebd0f8-fa32-4fe3-8569-fb7d4ad8faf4': {'code': 0, 'version': 0,
'acquired': True, 'delay': '0.000424499', 'lastCheck': '7.3', 'valid':
True}}

**********************************************************************************

[root at node002 shim]# multipath -ll
mpathe (1NODE_001_LUN01) dm-6 SHIMI,VIRTUAL-DISK
size=977G features='0' hwhandler='0' wp=rw
`-+- policy='round-robin 0' prio=1 status=active
  `- 21:0:0:1 sdg 8:96 active ready  running
mpathd (1NODE_003_LUN01) dm-7 SHIMI,VIRTUAL-DISK
size=977G features='0' hwhandler='0' wp=rw
`-+- policy='round-robin 0' prio=1 status=active
  `- 20:0:0:1 sdf 8:80 active ready  running
mpathc (1NODE_002_LUN01) dm-4 SHIMI,VIRTUAL-DISK
size=977G features='0' hwhandler='0' wp=rw
`-+- policy='round-robin 0' prio=1 status=active
  `- 18:0:0:1 sdd 8:48 active ready  running
mpathb (1ATA_MARVELL_Raid_VD_0_1c3c8ecf5cf00010) dm-1 ATA,MARVELL Raid VD
size=1.8T features='0' hwhandler='0' wp=rw
`-+- policy='round-robin 0' prio=1 status=active
  `- 0:0:0:0  sda 8:0  active ready  running
mpathf (1MANAGER_LUN01) dm-5 SHIMI,VIRTUAL-DISK
size=500G features='0' hwhandler='0' wp=rw
`-+- policy='round-robin 0' prio=1 status=active
  `- 19:0:0:1 sde 8:64 active ready  running

**********************************************************************************

[root at node002 shim]# lsblk
NAME                                                          MAJ:MIN RM
SIZE RO TYPE  MOUNTPOINT
sdb                                                             8:16   0
298.1G  0 disk
├─sdb1                                                          8:17
0     1G  0 part  /boot
├─sdb2                                                          8:18
0     4G  0 part  [SWAP]
└─sdb3                                                          8:19   0
293.1G  0 part
  └─vg_node002-LogVol00 (dm-0)                                253:0    0
293.1G  0 lvm   /
sda                                                             8:0    0
1.8T  0 disk
└─sda1                                                          8:1    0
1.8T  0 part
sdd                                                             8:48   0
976.6G  0 disk
└─mpathc (dm-4)                                               253:4    0
976.6G  0 mpath
sde                                                             8:64   0
500G  0 disk
└─mpathf (dm-5)                                               253:5    0
500G  0 mpath
  ├─cb85e6cd--df54--4151--8f3b--7e6d72b7372d-metadata (dm-15) 253:15   0
512M  0 lvm
  ├─cb85e6cd--df54--4151--8f3b--7e6d72b7372d-ids (dm-16)      253:16   0
128M  0 lvm
  ├─cb85e6cd--df54--4151--8f3b--7e6d72b7372d-leases (dm-18)   253:18
0     2G  0 lvm
  ├─cb85e6cd--df54--4151--8f3b--7e6d72b7372d-outbox (dm-20)   253:20   0
128M  0 lvm
  ├─cb85e6cd--df54--4151--8f3b--7e6d72b7372d-inbox (dm-21)    253:21   0
128M  0 lvm
  └─cb85e6cd--df54--4151--8f3b--7e6d72b7372d-master (dm-22)   253:22
0     1G  0 lvm
sdf                                                             8:80   0
976.6G  0 disk
└─mpathd (dm-7)                                               253:7    0
976.6G  0 mpath
  ├─5e1ca1b6--4706--4c79--8924--b8db741c929f-metadata (dm-14) 253:14   0
512M  0 lvm
  ├─5e1ca1b6--4706--4c79--8924--b8db741c929f-ids (dm-17)      253:17   0
128M  0 lvm
  ├─5e1ca1b6--4706--4c79--8924--b8db741c929f-leases (dm-19)   253:19
0     2G  0 lvm
  ├─5e1ca1b6--4706--4c79--8924--b8db741c929f-outbox (dm-23)   253:23   0
128M  0 lvm
  ├─5e1ca1b6--4706--4c79--8924--b8db741c929f-inbox (dm-24)    253:24   0
128M  0 lvm
  └─5e1ca1b6--4706--4c79--8924--b8db741c929f-master (dm-25)   253:25
0     1G  0 lvm
sdg                                                             8:96   0
976.6G  0 disk
└─mpathe (dm-6)                                               253:6    0
976.6G  0 mpath
  ├─5f595801--aaa5--42c7--b829--7a34a636407e-metadata (dm-8)  253:8    0
512M  0 lvm
  ├─5f595801--aaa5--42c7--b829--7a34a636407e-ids (dm-9)       253:9    0
128M  0 lvm
  ├─5f595801--aaa5--42c7--b829--7a34a636407e-leases (dm-10)   253:10
0     2G  0 lvm
  ├─5f595801--aaa5--42c7--b829--7a34a636407e-outbox (dm-11)   253:11   0
128M  0 lvm
  ├─5f595801--aaa5--42c7--b829--7a34a636407e-inbox (dm-12)    253:12   0
128M  0 lvm
  └─5f595801--aaa5--42c7--b829--7a34a636407e-master (dm-13)   253:13
0     1G  0 lvm

**********************************************************************************

[root at node002 shim]# multipath -v3
Jan 26 12:46:28 | ram0: device node name blacklisted
Jan 26 12:46:28 | ram1: device node name blacklisted
Jan 26 12:46:28 | ram2: device node name blacklisted
Jan 26 12:46:28 | ram3: device node name blacklisted
Jan 26 12:46:28 | ram4: device node name blacklisted
Jan 26 12:46:28 | ram5: device node name blacklisted
Jan 26 12:46:28 | ram6: device node name blacklisted
Jan 26 12:46:28 | ram7: device node name blacklisted
Jan 26 12:46:28 | ram8: device node name blacklisted
Jan 26 12:46:28 | ram9: device node name blacklisted
Jan 26 12:46:28 | ram10: device node name blacklisted
Jan 26 12:46:28 | ram11: device node name blacklisted
Jan 26 12:46:28 | ram12: device node name blacklisted
Jan 26 12:46:28 | ram13: device node name blacklisted
Jan 26 12:46:28 | ram14: device node name blacklisted
Jan 26 12:46:28 | ram15: device node name blacklisted
Jan 26 12:46:28 | loop0: device node name blacklisted
Jan 26 12:46:28 | loop1: device node name blacklisted
Jan 26 12:46:28 | loop2: device node name blacklisted
Jan 26 12:46:28 | loop3: device node name blacklisted
Jan 26 12:46:28 | loop4: device node name blacklisted
Jan 26 12:46:28 | loop5: device node name blacklisted
Jan 26 12:46:28 | loop6: device node name blacklisted
Jan 26 12:46:28 | loop7: device node name blacklisted
Jan 26 12:46:28 | sdb: not found in pathvec
Jan 26 12:46:28 | sdb: mask = 0x3f
Jan 26 12:46:28 | sdb: dev_t = 8:16
Jan 26 12:46:28 | sdb: size = 625142448
Jan 26 12:46:28 | sdb: subsystem = scsi
Jan 26 12:46:28 | sdb: vendor = ATA
Jan 26 12:46:28 | sdb: product = WDC WD3200AAJS-6
Jan 26 12:46:28 | sdb: rev = 03.0
Jan 26 12:46:28 | sdb: h:b:t:l = 10:0:0:0
Jan 26 12:46:28 | sdb: serial =      WD-WMAV2HM46197
Jan 26 12:46:28 | sdb: get_state
Jan 26 12:46:28 | sdb: path checker = directio (config file default)
Jan 26 12:46:28 | sdb: checker timeout = 30000 ms (sysfs setting)
Jan 26 12:46:28 | sdb: state = running
Jan 26 12:46:28 | directio: starting new request
Jan 26 12:46:28 | directio: io finished 4096/0
Jan 26 12:46:28 | sdb: state = 3
Jan 26 12:46:28 | sdb: getuid = /sbin/scsi_id --whitelisted
--replace-whitespace --device=/dev/%n (config file default)
Jan 26 12:46:28 | sdb: uid = 1ATA_WDC_WD3200AAJS-60Z0A0_WD-WMAV2HM46197
(callout)
Jan 26 12:46:28 | sdb: state = running
Jan 26 12:46:28 | sdb: detect_prio = 1 (config file default)
Jan 26 12:46:28 | sdb: prio = const (config file default)
Jan 26 12:46:28 | sdb: const prio = 1
Jan 26 12:46:28 | sda: not found in pathvec
Jan 26 12:46:28 | sda: mask = 0x3f
Jan 26 12:46:28 | sda: dev_t = 8:0
Jan 26 12:46:28 | sda: size = 3904897024
Jan 26 12:46:28 | sda: subsystem = scsi
Jan 26 12:46:28 | sda: vendor = ATA
Jan 26 12:46:28 | sda: product = MARVELL Raid VD
Jan 26 12:46:28 | sda: rev = MV.R
Jan 26 12:46:28 | sda: h:b:t:l = 0:0:0:0
Jan 26 12:46:28 | sda: serial = 1c3c8ecf5cf00010
Jan 26 12:46:28 | sda: get_state
Jan 26 12:46:28 | sda: path checker = directio (config file default)
Jan 26 12:46:28 | sda: checker timeout = 30000 ms (sysfs setting)
Jan 26 12:46:28 | sda: state = running
Jan 26 12:46:28 | directio: starting new request
Jan 26 12:46:28 | directio: io finished 4096/0
Jan 26 12:46:28 | sda: state = 3
Jan 26 12:46:28 | sda: getuid = /sbin/scsi_id --whitelisted
--replace-whitespace --device=/dev/%n (config file default)
Jan 26 12:46:28 | sda: uid = 1ATA_MARVELL_Raid_VD_0_1c3c8ecf5cf00010
(callout)
Jan 26 12:46:28 | sda: state = running
Jan 26 12:46:28 | sda: detect_prio = 1 (config file default)
Jan 26 12:46:28 | sda: prio = const (config file default)
Jan 26 12:46:28 | sda: const prio = 1
Jan 26 12:46:28 | dm-0: device node name blacklisted
Jan 26 12:46:28 | sdc: not found in pathvec
Jan 26 12:46:28 | sdc: mask = 0x3f
Jan 26 12:46:28 | sdc: dev_t = 8:32
Jan 26 12:46:28 | sdc: size = 0
Jan 26 12:46:28 | sdc: subsystem = scsi
Jan 26 12:46:28 | sdc: vendor = Multi
Jan 26 12:46:28 | sdc: product = Flash Reader
Jan 26 12:46:28 | sdc: rev = 1.00
Jan 26 12:46:28 | sdc: h:b:t:l = 12:0:0:0
Jan 26 12:46:28 | dm-1: device node name blacklisted
Jan 26 12:46:28 | dm-2: device node name blacklisted
Jan 26 12:46:28 | dm-3: device node name blacklisted
Jan 26 12:46:28 | sdd: not found in pathvec
Jan 26 12:46:28 | sdd: mask = 0x3f
Jan 26 12:46:28 | sdd: dev_t = 8:48
Jan 26 12:46:28 | sdd: size = 2048000000
Jan 26 12:46:28 | sdd: subsystem = scsi
Jan 26 12:46:28 | sdd: vendor = SHIMI
Jan 26 12:46:28 | sdd: product = VIRTUAL-DISK
Jan 26 12:46:28 | sdd: rev = 0001
Jan 26 12:46:28 | sdd: h:b:t:l = 18:0:0:1
Jan 26 12:46:28 | sdd: tgt_node_name = pl.mycomp.shimi:node002.target0
Jan 26 12:46:28 | sdd: serial =                               beaf11
Jan 26 12:46:28 | sdd: get_state
Jan 26 12:46:28 | sdd: path checker = directio (config file default)
Jan 26 12:46:28 | sdd: checker timeout = 30000 ms (sysfs setting)
Jan 26 12:46:28 | sdd: state = running
Jan 26 12:46:28 | directio: starting new request
Jan 26 12:46:28 | directio: io finished 4096/0
Jan 26 12:46:28 | sdd: state = 3
Jan 26 12:46:28 | sdd: getuid = /sbin/scsi_id --whitelisted
--replace-whitespace --device=/dev/%n (config file default)
Jan 26 12:46:28 | sdd: uid = 1NODE_002_LUN01 (callout)
Jan 26 12:46:28 | sdd: state = running
Jan 26 12:46:28 | sdd: detect_prio = 1 (config file default)
Jan 26 12:46:28 | sdd: prio = const (config file default)
Jan 26 12:46:28 | sdd: const prio = 1
Jan 26 12:46:28 | dm-4: device node name blacklisted
Jan 26 12:46:28 | sde: not found in pathvec
Jan 26 12:46:28 | sde: mask = 0x3f
Jan 26 12:46:28 | sde: dev_t = 8:64
Jan 26 12:46:28 | sde: size = 1048576000
Jan 26 12:46:28 | sde: subsystem = scsi
Jan 26 12:46:28 | sde: vendor = SHIMI
Jan 26 12:46:28 | sde: product = VIRTUAL-DISK
Jan 26 12:46:28 | sde: rev = 0001
Jan 26 12:46:28 | sde: h:b:t:l = 19:0:0:1
Jan 26 12:46:28 | sde: tgt_node_name = pl.mycomp.shimi:manager.target0
Jan 26 12:46:28 | sde: serial =                               beaf11
Jan 26 12:46:28 | sde: get_state
Jan 26 12:46:28 | sde: path checker = directio (config file default)
Jan 26 12:46:28 | sde: checker timeout = 30000 ms (sysfs setting)
Jan 26 12:46:28 | sde: state = running
Jan 26 12:46:28 | directio: starting new request
Jan 26 12:46:28 | directio: io finished 4096/0
Jan 26 12:46:28 | sde: state = 3
Jan 26 12:46:28 | sde: getuid = /sbin/scsi_id --whitelisted
--replace-whitespace --device=/dev/%n (config file default)
Jan 26 12:46:28 | sde: uid = 1MANAGER_LUN01 (callout)
Jan 26 12:46:28 | sde: state = running
Jan 26 12:46:28 | sde: detect_prio = 1 (config file default)
Jan 26 12:46:28 | sde: prio = const (config file default)
Jan 26 12:46:28 | sde: const prio = 1
Jan 26 12:46:28 | sdf: not found in pathvec
Jan 26 12:46:28 | sdf: mask = 0x3f
Jan 26 12:46:28 | sdf: dev_t = 8:80
Jan 26 12:46:28 | sdf: size = 2048000000
Jan 26 12:46:28 | sdf: subsystem = scsi
Jan 26 12:46:28 | sdf: vendor = SHIMI
Jan 26 12:46:28 | sdf: product = VIRTUAL-DISK
Jan 26 12:46:28 | sdf: rev = 0001
Jan 26 12:46:28 | sdf: h:b:t:l = 20:0:0:1
Jan 26 12:46:28 | sdf: tgt_node_name = pl.mycomp.shimi:node003.target0
Jan 26 12:46:28 | sdf: serial =                               beaf11
Jan 26 12:46:28 | sdf: get_state
Jan 26 12:46:28 | sdf: path checker = directio (config file default)
Jan 26 12:46:28 | sdf: checker timeout = 30000 ms (sysfs setting)
Jan 26 12:46:28 | sdf: state = running
Jan 26 12:46:28 | directio: starting new request
Jan 26 12:46:28 | directio: io finished 4096/0
Jan 26 12:46:28 | sdf: state = 3
Jan 26 12:46:28 | sdf: getuid = /sbin/scsi_id --whitelisted
--replace-whitespace --device=/dev/%n (config file default)
Jan 26 12:46:28 | sdf: uid = 1NODE_003_LUN01 (callout)
Jan 26 12:46:28 | sdf: state = running
Jan 26 12:46:28 | sdf: detect_prio = 1 (config file default)
Jan 26 12:46:28 | sdf: prio = const (config file default)
Jan 26 12:46:28 | sdf: const prio = 1
Jan 26 12:46:28 | sdg: not found in pathvec
Jan 26 12:46:28 | sdg: mask = 0x3f
Jan 26 12:46:28 | sdg: dev_t = 8:96
Jan 26 12:46:28 | sdg: size = 2048000000
Jan 26 12:46:28 | sdg: subsystem = scsi
Jan 26 12:46:28 | sdg: vendor = SHIMI
Jan 26 12:46:28 | sdg: product = VIRTUAL-DISK
Jan 26 12:46:28 | sdg: rev = 0001
Jan 26 12:46:28 | sdg: h:b:t:l = 21:0:0:1
Jan 26 12:46:28 | sdg: tgt_node_name = pl.mycomp.shimi:node001.target0
Jan 26 12:46:28 | sdg: serial =                               beaf11
Jan 26 12:46:28 | sdg: get_state
Jan 26 12:46:28 | sdg: path checker = directio (config file default)
Jan 26 12:46:28 | sdg: checker timeout = 30000 ms (sysfs setting)
Jan 26 12:46:28 | sdg: state = running
Jan 26 12:46:28 | directio: starting new request
Jan 26 12:46:28 | directio: io finished 4096/0
Jan 26 12:46:28 | sdg: state = 3
Jan 26 12:46:28 | sdg: getuid = /sbin/scsi_id --whitelisted
--replace-whitespace --device=/dev/%n (config file default)
Jan 26 12:46:28 | sdg: uid = 1NODE_001_LUN01 (callout)
Jan 26 12:46:28 | sdg: state = running
Jan 26 12:46:28 | sdg: detect_prio = 1 (config file default)
Jan 26 12:46:28 | sdg: prio = const (config file default)
Jan 26 12:46:28 | sdg: const prio = 1
Jan 26 12:46:28 | dm-5: device node name blacklisted
Jan 26 12:46:28 | dm-6: device node name blacklisted
Jan 26 12:46:28 | dm-7: device node name blacklisted
Jan 26 12:46:28 | dm-8: device node name blacklisted
Jan 26 12:46:28 | dm-9: device node name blacklisted
Jan 26 12:46:28 | dm-10: device node name blacklisted
Jan 26 12:46:28 | dm-11: device node name blacklisted
Jan 26 12:46:28 | dm-12: device node name blacklisted
Jan 26 12:46:28 | dm-13: device node name blacklisted
Jan 26 12:46:28 | dm-14: device node name blacklisted
Jan 26 12:46:28 | dm-15: device node name blacklisted
Jan 26 12:46:28 | dm-16: device node name blacklisted
Jan 26 12:46:28 | dm-17: device node name blacklisted
Jan 26 12:46:28 | dm-18: device node name blacklisted
Jan 26 12:46:28 | dm-19: device node name blacklisted
Jan 26 12:46:28 | dm-20: device node name blacklisted
Jan 26 12:46:28 | dm-21: device node name blacklisted
Jan 26 12:46:28 | dm-22: device node name blacklisted
Jan 26 12:46:28 | dm-23: device node name blacklisted
Jan 26 12:46:28 | dm-24: device node name blacklisted
Jan 26 12:46:28 | dm-25: device node name blacklisted
===== paths list =====
uuid                                       hcil     dev dev_t pri dm_st
chk_st
1ATA_WDC_WD3200AAJS-60Z0A0_WD-WMAV2HM46197 10:0:0:0 sdb 8:16  1   undef
ready
1ATA_MARVELL_Raid_VD_0_1c3c8ecf5cf00010    0:0:0:0  sda 8:0   1   undef
ready
                                           12:0:0:0 sdc 8:32  -1  undef
faulty
1NODE_002_LUN01                            18:0:0:1 sdd 8:48  1   undef
ready
1MANAGER_LUN01                             19:0:0:1 sde 8:64  1   undef
ready
1NODE_003_LUN01                            20:0:0:1 sdf 8:80  1   undef
ready
1NODE_001_LUN01                            21:0:0:1 sdg 8:96  1   undef
ready
Jan 26 12:46:28 | params = 0 0 1 1 round-robin 0 1 1 8:96 1
Jan 26 12:46:28 | status = 2 0 0 0 1 1 A 0 1 0 8:96 A 0
Jan 26 12:46:28 | params = 0 0 1 1 round-robin 0 1 1 8:80 1
Jan 26 12:46:28 | status = 2 0 0 0 1 1 A 0 1 0 8:80 A 0
Jan 26 12:46:28 | params = 0 0 1 1 round-robin 0 1 1 8:48 1
Jan 26 12:46:28 | status = 2 0 0 0 1 1 A 0 1 0 8:48 A 0
Jan 26 12:46:28 | params = 0 0 1 1 round-robin 0 1 1 8:0 1
Jan 26 12:46:28 | status = 2 0 0 0 1 1 A 0 1 0 8:0 A 0
Jan 26 12:46:28 | params = 0 0 1 1 round-robin 0 1 1 8:64 1
Jan 26 12:46:28 | status = 2 0 0 0 1 1 A 0 1 0 8:64 A 0
Jan 26 12:46:28 | Found matching wwid
[1ATA_WDC_WD3200AAJS-60Z0A0_WD-WMAV2HM46197] in bindings file. Setting
alias to mpatha
Jan 26 12:46:28 | sdb: ownership set to mpatha
Jan 26 12:46:28 | sdb: not found in pathvec
Jan 26 12:46:28 | sdb: mask = 0xc
Jan 26 12:46:28 | sdb: get_state
Jan 26 12:46:28 | sdb: state = running
Jan 26 12:46:28 | directio: starting new request
Jan 26 12:46:28 | directio: io finished 4096/0
Jan 26 12:46:28 | sdb: state = 3
Jan 26 12:46:28 | sdb: state = running
Jan 26 12:46:28 | sdb: const prio = 1
Jan 26 12:46:28 | mpatha: pgfailover = -1 (internal default)
Jan 26 12:46:28 | mpatha: pgpolicy = failover (internal default)
Jan 26 12:46:28 | mpatha: selector = round-robin 0 (internal default)
Jan 26 12:46:28 | mpatha: features = 0 (internal default)
Jan 26 12:46:28 | mpatha: hwhandler = 0 (internal default)
Jan 26 12:46:28 | mpatha: rr_weight = 1 (internal default)
Jan 26 12:46:28 | mpatha: minio = 1 rq (config file default)
Jan 26 12:46:28 | mpatha: no_path_retry = -1 (config file default)
Jan 26 12:46:28 | pg_timeout = NONE (internal default)
Jan 26 12:46:28 | mpatha: fast_io_fail_tmo = 5 (config file default)
Jan 26 12:46:28 | mpatha: dev_loss_tmo = 30 (config file default)
Jan 26 12:46:28 | mpatha: retain_attached_hw_handler = 1 (config file
default)
Jan 26 12:46:28 | failed to find rport_id for target10:0:0
Jan 26 12:46:28 | mpatha: set ACT_CREATE (map does not exist)
Jan 26 12:46:28 | mpatha: domap (0) failure for create/reload map
Jan 26 12:46:28 | mpatha: ignoring map

**********************************************************************************

[root at node002 shim]# iscsiadm -m session -o show
tcp: [6] 192.168.1.12:3260,1 pl.mycomp.shimi:node002.target0
tcp: [7] 192.168.1.11:3260,1 pl.mycomp.shimi:manager.target0
tcp: [8] 192.168.1.14:3260,1 pl.mycomp.shimi:node003.target0
tcp: [9] 192.168.1.13:3260,1 pl.mycomp.shimi:node001.target0

**********************************************************************************

[root at node002 shim]# iptables -L
Chain INPUT (policy ACCEPT)
target     prot opt source               destination

Chain FORWARD (policy ACCEPT)
target     prot opt source               destination

Chain OUTPUT (policy ACCEPT)
target     prot opt source               destination

**********************************************************************************

[root at node002 shim]# sestatus
SELinux status:                 disabled
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.ovirt.org/pipermail/users/attachments/20150127/46125f57/attachment-0001.html>


More information about the Users mailing list