/> ls
o- / ......................................................................................................................... [...]
o- backstores .............................................................................................................. [...]
| o- block .................................................................................................. [Storage Objects: 0]
| o- fileio ................................................................................................. [Storage Objects: 4]
| | o- MAISON-core2-iscsi-dsk01 ....................................... [/MAISON/core2/iscsi/dsk01 (20.0GiB) write-thru activated]
| | o- MAISON-core2-iscsi-dsk02 ....................................... [/MAISON/core2/iscsi/dsk02 (20.0GiB) write-thru activated]
| | o- dsk01 ........................................................ [/MAISON/core2/iscsi/dsk01 (20.0GiB) write-thru deactivated]
| | o- dsk02 ........................................................ [/MAISON/core2/iscsi/dsk02 (20.0GiB) write-thru deactivated]
| o- pscsi .................................................................................................. [Storage Objects: 0]
| o- ramdisk ................................................................................................ [Storage Objects: 0]
| o- user ................................................................................................... [Storage Objects: 0]
o- iscsi ............................................................................................................ [Targets: 1]
| o- iqn.2017-06.stet.iscsi.server:target .............................................................................. [TPGs: 1]
| o- tpg1 ............................................................................................... [no-gen-acls, no-auth]
| o- acls .......................................................................................................... [ACLs: 2]
| | o- iqn.1994-05.com.redhat:287c2b12050 ................................................................... [Mapped LUNs: 2]
| | | o- mapped_lun0 ............................................................. [lun0 fileio/MAISON-core2-iscsi-dsk01 (rw)]
| | | o- mapped_lun1 ............................................................. [lun1 fileio/MAISON-core2-iscsi-dsk02 (rw)]
| | o- iqn.1994-05.com.redhat:f654257ba68 ................................................................... [Mapped LUNs: 2]
| | o- mapped_lun0 ............................................................. [lun0 fileio/MAISON-core2-iscsi-dsk01 (rw)]
| | o- mapped_lun1 ............................................................. [lun1 fileio/MAISON-core2-iscsi-dsk02 (rw)]
| o- luns .......................................................................................................... [LUNs: 2]
| | o- lun0 .................................................... [fileio/MAISON-core2-iscsi-dsk01 (/MAISON/core2/iscsi/dsk01)]
| | o- lun1 .................................................... [fileio/MAISON-core2-iscsi-dsk02 (/MAISON/core2/iscsi/dsk02)]
| o- portals .................................................................................................... [Portals: 2]
| o-
10.10.1.129:3260 ..............................
................................................................... [OK] | o-
10.10.1.1:3260 ..............................
..................................................................... [OK] o- loopback ......................................................................................................... [Targets: 0]
o- vhost ............................................................................................................ [Targets: 0]
/>
Here is our RHEV-H hypervisor which has 3 nic configured.
[root@pdx1cr207 ~]# ip a
1: lo: <LOOPBACK,UP,LOWER_UP> mtu 65536 qdisc noqueue state UNKNOWN qlen 1
link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
inet 127.0.0.1/8 scope host lo
valid_lft forever preferred_lft forever
inet6 ::1/128 scope host
valid_lft forever preferred_lft forever
2: eth0: <BROADCAST,MULTICAST,SLAVE,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast master bond0 state UP qlen 1000
link/ether 52:54:00:1c:25:02 brd ff:ff:ff:ff:ff:ff
3: eth1: <BROADCAST,MULTICAST,SLAVE,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast master bond0 state UP qlen 1000
link/ether 52:54:00:1c:25:02 brd ff:ff:ff:ff:ff:ff
4: eth2: <BROADCAST,MULTICAST,SLAVE,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast master bond1 state UP qlen 1000
link/ether 52:54:00:9a:7d:06 brd ff:ff:ff:ff:ff:ff
5: eth3: <BROADCAST,MULTICAST,SLAVE,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast master bond1 state UP qlen 1000
link/ether 52:54:00:9a:7d:06 brd ff:ff:ff:ff:ff:ff
6: eth4: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN qlen 1000
link/ether 52:54:00:e5:b8:42 brd ff:ff:ff:ff:ff:ff
7: eth5: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN qlen 1000
link/ether 52:54:00:e6:55:59 brd ff:ff:ff:ff:ff:ff
8: eth6: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast state UP qlen 1000
link/ether 52:54:00:9b:80:80 brd ff:ff:ff:ff:ff:ff
inet 10.10.1.11/25 brd 10.10.1.127 scope global eth6
valid_lft forever preferred_lft forever
inet6 fe80::5054:ff:fe9b:8080/64 scope link
valid_lft forever preferred_lft forever
9: eth7: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast state UP qlen 1000
link/ether 52:54:00:f7:5f:f8 brd ff:ff:ff:ff:ff:ff
inet 10.10.1.139/25 brd 10.10.1.255 scope global eth7
valid_lft forever preferred_lft forever
inet6 fe80::5054:ff:fef7:5ff8/64 scope link
valid_lft forever preferred_lft forever
10: bond0: <BROADCAST,MULTICAST,MASTER,UP,LOWER_UP> mtu 1500 qdisc noqueue master ovirtmgmt state UP qlen 1000
link/ether 52:54:00:1c:25:02 brd ff:ff:ff:ff:ff:ff
11: ovirtmgmt: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc noqueue state UP qlen 1000
link/ether 52:54:00:1c:25:02 brd ff:ff:ff:ff:ff:ff
inet 10.11.8.168/27 brd 10.11.8.191 scope global ovirtmgmt
valid_lft forever preferred_lft forever
12: bond1: <BROADCAST,MULTICAST,MASTER,UP,LOWER_UP> mtu 1500 qdisc noqueue state UP qlen 1000
link/ether 52:54:00:9a:7d:06 brd ff:ff:ff:ff:ff:ff
inet6 fe80::5054:ff:fe9a:7d06/64 scope link
valid_lft forever preferred_lft forever
13: ;vdsmdummy;: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN qlen 1000
link/ether 7a:79:ea:86:d6:5d brd ff:ff:ff:ff:ff:ff
Here is our playbook used to prouve the issue. It points 10.10.1.1 (iscsi servers)
---
- hosts: rhevm2
gather_facts: False
remote_user: root
tasks:
- name: authentication
ovirt_auth:
url: https://rhevm2.res1/ovirt-engine/api
username: admin@internal
password: rootroot01
insecure: True
#
- ovirt_storage_domains_facts:
auth:
url: https://rhevm2.res1/ovirt-engine/api
token: "{{ ovirt_auth.token }}"
insecure: True
pattern: name=* and datacenter=default
- debug:
var: ovirt_storage_domains
#
- name: create iscsi_dsk02
ovirt_storage_domains:
auth:
url: https://rhevm2.res1/ovirt-engine/api
token: "{{ ovirt_auth.token }}"
insecure: True
name: iscsi_dsk02
domain_function: data
host: pdx1cr207
data_center: prslog
iscsi:
target: iqn.2017-06.stet.iscsi.server:target
address: 10.10.1.1
lun_id: 3600140550738b53dd774303bfedac122
# lun_id: 36001405ed330f17f8e74ca1a08b7bb04
state: present
destroy: true
register: res_iscsi_dsk02
- debug:
var: res_iscsi_dsk02
We run the playbook and got error message:
ansible @sichuan workspace]$ansible-playbook ds_create6.yml -v
Using /home/papa/environments/workspace/inventories/env_x1rop/ansible.cfg as config file
PLAY [rhevm2] ********************************************************************************************************************************
TASK [authentication] ************************************************************************************************************************
ok: [rhevm2] => {"ansible_facts": {"ovirt_auth": {"ca_file": null, "compress": true, "insecure": true, "kerberos": false, "timeout": 0, "token
": "t1ibTsNmCh13nYhwpFpXJb5eK6YSIH-V4FdXGN2PKWq-g-F2ksAWjl51iYK0dB-2-yWHYvTGL6Xgo_2jtab2tA", "url": "https://rhevm2.res1/ovirt-engine/api"}},
"changed": false}
TASK [ovirt_storage_domains_facts] ***********************************************************************************************************
ok: [rhevm2] => {"ansible_facts": {"ovirt_storage_domains": [{"available": 95563022336, "committed": 0, "critical_space_action_blocker": 5, "d
ata_centers": [{"id": "5950a8cc-03df-013e-02a8-0000000001ff"}], "disk_profiles": [], "disk_snapshots": [], "disks": [], "external_status": "ok
", "href": "/ovirt-engine/api/storagedomains/d2b84bb1-526a-4981-b6b4-f69b7e079d0d", "id": "d2b84bb1-526a-4981-b6b4-f69b7e079d0d", "master": tr
ue, "name": "dm02", "permissions": [], "storage": {"address": "10.11.8.190", "nfs_version": "v3", "path": "/MAISON/nfs/domaines/dm02", "type":
"nfs"}, "storage_connections": [], "storage_format": "v3", "templates": [], "type": "data", "used": 9663676416, "vms": [], "warning_low_space
_indicator": 10, "wipe_after_delete": false}]}, "changed": false}
TASK [debug] *********************************************************************************************************************************
ok: [rhevm2] => {
"ovirt_storage_domains": [
{
"available": 95563022336,
"committed": 0,
"critical_space_action_blocker": 5,
"data_centers": [
{
"id": "5950a8cc-03df-013e-02a8-0000000001ff"
}
],
"disk_profiles": [],
"disk_snapshots": [],
"disks": [],
"external_status": "ok",
"href": "/ovirt-engine/api/storagedomains/d2b84bb1-526a-4981-b6b4-f69b7e079d0d",
"id": "d2b84bb1-526a-4981-b6b4-f69b7e079d0d",
"master": true,
"name": "dm02",
"permissions": [],
"storage": {
"address": "10.11.8.190",
"nfs_version": "v3",
"path": "/MAISON/nfs/domaines/dm02",
"type": "nfs"
},
"storage_connections": [],
"storage_format": "v3",
"templates": [],
"type": "data",
"used": 9663676416,
"vms": [],
"warning_low_space_indicator": 10,
"wipe_after_delete": false
}
]
}
TASK [create iscsi_dsk02] ********************************************************************************************************************
An exception occurred during task execution. To see the full traceback, use -vvv. The error was: Error: Fault reason is "Operation Failed". Fa
ult detail is "[Storage domain cannot be reached. Please ensure it is accessible from the host(s).]". HTTP response code is 400.
fatal: [rhevm2]: FAILED! => {"changed": false, "failed": true, "msg": "Fault reason is \"Operation Failed\". Fault detail is \"[Storage domain
cannot be reached. Please ensure it is accessible from the host(s).]\". HTTP response code is 400."}
to retry, use: --limit @/home/papa/environments/workspace/ds_create6.retry
PLAY RECAP ***********************************************************************************************************************************
rhevm2 : ok=3 changed=0 unreachable=0 failed=1
Here is compressed vdsm et engine log: