*Here is our iscsi target config( on iscsi server), we have 2 iscsi
devices. *
/> ls
o- / ............................................................
............................................................. [...]
o- backstores ............................................................
.................................................. [...]
| o- block ............................................................
...................................... [Storage Objects: 0]
| o- fileio ............................................................
..................................... [Storage Objects: 4]
| | o- MAISON-core2-iscsi-dsk01 .......................................
[/MAISON/core2/iscsi/dsk01 (20.0GiB) write-thru activated]
| | o- MAISON-core2-iscsi-dsk02 .......................................
[/MAISON/core2/iscsi/dsk02 (20.0GiB) write-thru activated]
| | o- dsk01 ........................................................
[/MAISON/core2/iscsi/dsk01 (20.0GiB) write-thru deactivated]
| | o- dsk02 ........................................................
[/MAISON/core2/iscsi/dsk02 (20.0GiB) write-thru deactivated]
| o- pscsi ............................................................
...................................... [Storage Objects: 0]
| o- ramdisk ............................................................
.................................... [Storage Objects: 0]
| o- user ............................................................
....................................... [Storage Objects: 0]
o- iscsi ............................................................
................................................ [Targets: 1]
| o- iqn.2017-06.stet.iscsi.server:target ..............................
................................................ [TPGs: 1]
| o- tpg1 ............................................................
................................... [no-gen-acls, no-auth]
| o- acls ............................................................
.............................................. [ACLs: 2]
| | o- iqn.1994-05.com.redhat:287c2b12050
................................................................... [Mapped
LUNs: 2]
| | | o- mapped_lun0 ..............................
............................... [lun0 fileio/MAISON-core2-iscsi-dsk01 (rw)]
| | | o- mapped_lun1 ..............................
............................... [lun1 fileio/MAISON-core2-iscsi-dsk02 (rw)]
| | o- iqn.1994-05.com.redhat:f654257ba68
................................................................... [Mapped
LUNs: 2]
| | o- mapped_lun0 ..............................
............................... [lun0 fileio/MAISON-core2-iscsi-dsk01 (rw)]
| | o- mapped_lun1 ..............................
............................... [lun1 fileio/MAISON-core2-iscsi-dsk02 (rw)]
| o- luns ............................................................
.............................................. [LUNs: 2]
| | o- lun0 ....................................................
[fileio/MAISON-core2-iscsi-dsk01 (/MAISON/core2/iscsi/dsk01)]
| | o- lun1 ....................................................
[fileio/MAISON-core2-iscsi-dsk02 (/MAISON/core2/iscsi/dsk02)]
| o- portals ..............................
......................................................................
[Portals: 2]
| o- 10.10.1.129:3260 ..............................
................................................................... [OK]
| o- 10.10.1.1:3260 ..............................
..................................................................... [OK]
o- loopback ............................................................
............................................. [Targets: 0]
o- vhost ............................................................
................................................ [Targets: 0]
/>
*Here is our RHEV-H hypervisor which has 3 nic configured.*
*10.11.8.168/27 <
http://10.11.8.168/27> for management and NFS*
*10.10.1.11/25 <
http://10.10.1.11/25> and 10.10.1.139/25
<
http://10.10.1.139/25> are dedicated to ISCSI. The two are separated in
different *
*VLAN so as that 10.10.1.11/25 <
http://10.10.1.11/25> will talk to
10.10.1.1 (iscsi server side) and 10.10.1.139/25 <
http://10.10.1.139/25>
will talk to 10.10.1.12 <
http://10.10.1.129:3260/>9 (server side).*
[root@pdx1cr207 ~]# ip a
1: lo: <LOOPBACK,UP,LOWER_UP> mtu 65536 qdisc noqueue state UNKNOWN qlen 1
link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00
inet 127.0.0.1/8 scope host lo
valid_lft forever preferred_lft forever
inet6 ::1/128 scope host
valid_lft forever preferred_lft forever
2: eth0: <BROADCAST,MULTICAST,SLAVE,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast
master bond0 state UP qlen 1000
link/ether 52:54:00:1c:25:02 brd ff:ff:ff:ff:ff:ff
3: eth1: <BROADCAST,MULTICAST,SLAVE,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast
master bond0 state UP qlen 1000
link/ether 52:54:00:1c:25:02 brd ff:ff:ff:ff:ff:ff
4: eth2: <BROADCAST,MULTICAST,SLAVE,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast
master bond1 state UP qlen 1000
link/ether 52:54:00:9a:7d:06 brd ff:ff:ff:ff:ff:ff
5: eth3: <BROADCAST,MULTICAST,SLAVE,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast
master bond1 state UP qlen 1000
link/ether 52:54:00:9a:7d:06 brd ff:ff:ff:ff:ff:ff
6: eth4: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN qlen 1000
link/ether 52:54:00:e5:b8:42 brd ff:ff:ff:ff:ff:ff
7: eth5: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN qlen 1000
link/ether 52:54:00:e6:55:59 brd ff:ff:ff:ff:ff:ff
8: eth6: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast state
UP qlen 1000
link/ether 52:54:00:9b:80:80 brd ff:ff:ff:ff:ff:ff
inet 10.10.1.11/25 brd 10.10.1.127 scope global eth6
valid_lft forever preferred_lft forever
inet6 fe80::5054:ff:fe9b:8080/64 scope link
valid_lft forever preferred_lft forever
9: eth7: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast state
UP qlen 1000
link/ether 52:54:00:f7:5f:f8 brd ff:ff:ff:ff:ff:ff
inet 10.10.1.139/25 brd 10.10.1.255 scope global eth7
valid_lft forever preferred_lft forever
inet6 fe80::5054:ff:fef7:5ff8/64 scope link
valid_lft forever preferred_lft forever
10: bond0: <BROADCAST,MULTICAST,MASTER,UP,LOWER_UP> mtu 1500 qdisc noqueue
master ovirtmgmt state UP qlen 1000
link/ether 52:54:00:1c:25:02 brd ff:ff:ff:ff:ff:ff
11: ovirtmgmt: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc noqueue
state UP qlen 1000
link/ether 52:54:00:1c:25:02 brd ff:ff:ff:ff:ff:ff
inet 10.11.8.168/27 brd 10.11.8.191 scope global ovirtmgmt
valid_lft forever preferred_lft forever
12: bond1: <BROADCAST,MULTICAST,MASTER,UP,LOWER_UP> mtu 1500 qdisc noqueue
state UP qlen 1000
link/ether 52:54:00:9a:7d:06 brd ff:ff:ff:ff:ff:ff
inet6 fe80::5054:ff:fe9a:7d06/64 scope link
valid_lft forever preferred_lft forever
13: ;vdsmdummy;: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN qlen
1000
link/ether 7a:79:ea:86:d6:5d brd ff:ff:ff:ff:ff:ff
*Here is our playbook used to prouve the issue. It points 10.10.1.1 (iscsi
servers)*
---
- hosts: rhevm2
gather_facts: False
remote_user: root
tasks:
- name: authentication
ovirt_auth:
url:
https://rhevm2.res1/ovirt-engine/api
username: admin@internal
password: rootroot01
insecure: True
#
- ovirt_storage_domains_facts:
auth:
url:
https://rhevm2.res1/ovirt-engine/api
token: "{{ ovirt_auth.token }}"
insecure: True
pattern: name=* and datacenter=default
- debug:
var: ovirt_storage_domains
#
- name: create iscsi_dsk02
ovirt_storage_domains:
auth:
url:
https://rhevm2.res1/ovirt-engine/api
token: "{{ ovirt_auth.token }}"
insecure: True
name: iscsi_dsk02
domain_function: data
host: pdx1cr207
data_center: prslog
iscsi:
target: iqn.2017-06.stet.iscsi.server:target
address: 10.10.1.1
lun_id: 3600140550738b53dd774303bfedac122
# lun_id: 36001405ed330f17f8e74ca1a08b7bb04
state: present
destroy: true
register: res_iscsi_dsk02
- debug:
var: res_iscsi_dsk02
*We run the playbook and got error message:*
ansible @sichuan workspace]$ansible-playbook ds_create6.yml -v
Using /home/papa/environments/workspace/inventories/env_x1rop/ansible.cfg
as config file
PLAY [rhevm2]
********************************************************************************************************************************
TASK [authentication]
************************************************************************************************************************
ok: [rhevm2] => {"ansible_facts": {"ovirt_auth":
{"ca_file": null,
"compress": true, "insecure": true, "kerberos": false,
"timeout": 0, "token
":
"t1ibTsNmCh13nYhwpFpXJb5eK6YSIH-V4FdXGN2PKWq-g-F2ksAWjl51iYK0dB-2-yWHYvTGL6Xgo_2jtab2tA",
"url": "https://rhevm2.res1/ovirt-engine/api"}},
"changed": false}
TASK [ovirt_storage_domains_facts]
***********************************************************************************************************
ok: [rhevm2] => {"ansible_facts": {"ovirt_storage_domains":
[{"available":
95563022336, "committed": 0, "critical_space_action_blocker": 5,
"d
ata_centers": [{"id": "5950a8cc-03df-013e-02a8-0000000001ff"}],
"disk_profiles": [], "disk_snapshots": [], "disks": [],
"external_status":
"ok
", "href":
"/ovirt-engine/api/storagedomains/d2b84bb1-526a-4981-b6b4-f69b7e079d0d",
"id": "d2b84bb1-526a-4981-b6b4-f69b7e079d0d", "master": tr
ue, "name": "dm02", "permissions": [], "storage":
{"address":
"10.11.8.190", "nfs_version": "v3", "path":
"/MAISON/nfs/domaines/dm02",
"type":
"nfs"}, "storage_connections": [], "storage_format":
"v3", "templates": [],
"type": "data", "used": 9663676416, "vms": [],
"warning_low_space
_indicator": 10, "wipe_after_delete": false}]}, "changed":
false}
TASK [debug]
*********************************************************************************************************************************
ok: [rhevm2] => {
"ovirt_storage_domains": [
{
"available": 95563022336,
"committed": 0,
"critical_space_action_blocker": 5,
"data_centers": [
{
"id": "5950a8cc-03df-013e-02a8-0000000001ff"
}
],
"disk_profiles": [],
"disk_snapshots": [],
"disks": [],
"external_status": "ok",
"href":
"/ovirt-engine/api/storagedomains/d2b84bb1-526a-4981-b6b4-f69b7e079d0d",
"id": "d2b84bb1-526a-4981-b6b4-f69b7e079d0d",
"master": true,
"name": "dm02",
"permissions": [],
"storage": {
"address": "10.11.8.190",
"nfs_version": "v3",
"path": "/MAISON/nfs/domaines/dm02",
"type": "nfs"
},
"storage_connections": [],
"storage_format": "v3",
"templates": [],
"type": "data",
"used": 9663676416,
"vms": [],
"warning_low_space_indicator": 10,
"wipe_after_delete": false
}
]
}
TASK [create iscsi_dsk02]
********************************************************************************************************************
An exception occurred during task execution. To see the full traceback, use
-vvv. The error was: Error: Fault reason is "Operation Failed". Fa
ult detail is "[Storage domain cannot be reached. Please ensure it is
accessible from the host(s).]". HTTP response code is 400.
fatal: [rhevm2]: FAILED! => {"changed": false, "failed": true,
"msg":
"Fault reason is \"Operation Failed\". Fault detail is \"[Storage
domain
cannot be reached. Please ensure it is accessible from the host(s).]\".
HTTP response code is 400."}
to retry, use: --limit
@/home/papa/environments/workspace/ds_create6.retry
PLAY RECAP
***********************************************************************************************************************************
rhevm2 : ok=3 changed=0 unreachable=0 failed=1
*Here is compressed vdsm et engine log:*
On Wed, Jun 28, 2017 at 3:28 PM, Yaniv Kaul <ykaul(a)redhat.com> wrote:
Can you share vdsm log?
On Wed, Jun 28, 2017 at 4:16 PM, victor 600833 <victor600833(a)gmail.com>
wrote:
> Hi ondra,
>
> As you pointed. in your previous mail.
>
> The iscsiadm discovered througth 10.11.8.190. It should have probed 10.10.1.1.
> This may be the root cause.
>
>
>
> Victor,
>
>
>
>
>
> If you lookup at our playbook.
>
> - name: create iscsi_dsk02
> ovirt_storage_domains:
> auth:
> url:
https://rhevm2.res1/ovirt-engine/api
> token: "{{ ovirt_auth.token }}"
> insecure: True
> name: iscsi_dsk02
> domain_function: data
> host: pdx1cr207
> data_center: default
> iscsi:
> target: iqn.2017-06.stet.iscsi.server:target
> address: *10.10.1.1*
> lun_id: 3600140550738b53dd774303bfedac122
> state: absent
> destroy: true
> register: res_iscsi_dsk02
> - debug:
> var: res_iscsi_dsk02
>
> On Wed, Jun 28, 2017 at 7:58 AM, Ondra Machacek <notifications(a)github.com
> > wrote:
>
>> It's using the host, which you will pass as host parameter to the task
>> for iscsi discover.
>>
>> As Yaniv said, can you please share your issue on our mailing list:
>>
>> users(a)ovirt.org
>>
>> There are people with knowledge of iSCSI and they will help you.
>>
>> —
>> You are receiving this because you were mentioned.
>> Reply to this email directly, view it on GitHub
>> <
https://github.com/ansible/ansible/issues/25417#issuecomment-311563987>,
>> or mute the thread
>>
<
https://github.com/notifications/unsubscribe-auth/AScoAyV364pDjqXtCPSNodv...
>> .
>>
>
>
> _______________________________________________
> Users mailing list
> Users(a)ovirt.org
>
http://lists.ovirt.org/mailman/listinfo/users
>
>