
*Here is our iscsi target config( on iscsi server), we have 2 iscsi devices. * /> ls o- / ............................................................ ............................................................. [...] o- backstores ............................................................ .................................................. [...] | o- block ............................................................ ...................................... [Storage Objects: 0] | o- fileio ............................................................ ..................................... [Storage Objects: 4] | | o- MAISON-core2-iscsi-dsk01 ....................................... [/MAISON/core2/iscsi/dsk01 (20.0GiB) write-thru activated] | | o- MAISON-core2-iscsi-dsk02 ....................................... [/MAISON/core2/iscsi/dsk02 (20.0GiB) write-thru activated] | | o- dsk01 ........................................................ [/MAISON/core2/iscsi/dsk01 (20.0GiB) write-thru deactivated] | | o- dsk02 ........................................................ [/MAISON/core2/iscsi/dsk02 (20.0GiB) write-thru deactivated] | o- pscsi ............................................................ ...................................... [Storage Objects: 0] | o- ramdisk ............................................................ .................................... [Storage Objects: 0] | o- user ............................................................ ....................................... [Storage Objects: 0] o- iscsi ............................................................ ................................................ [Targets: 1] | o- iqn.2017-06.stet.iscsi.server:target .............................. ................................................ [TPGs: 1] | o- tpg1 ............................................................ ................................... [no-gen-acls, no-auth] | o- acls ............................................................ .............................................. [ACLs: 2] | | o- iqn.1994-05.com.redhat:287c2b12050 ................................................................... [Mapped LUNs: 2] | | | o- mapped_lun0 .............................. ............................... [lun0 fileio/MAISON-core2-iscsi-dsk01 (rw)] | | | o- mapped_lun1 .............................. ............................... [lun1 fileio/MAISON-core2-iscsi-dsk02 (rw)] | | o- iqn.1994-05.com.redhat:f654257ba68 ................................................................... [Mapped LUNs: 2] | | o- mapped_lun0 .............................. ............................... [lun0 fileio/MAISON-core2-iscsi-dsk01 (rw)] | | o- mapped_lun1 .............................. ............................... [lun1 fileio/MAISON-core2-iscsi-dsk02 (rw)] | o- luns ............................................................ .............................................. [LUNs: 2] | | o- lun0 .................................................... [fileio/MAISON-core2-iscsi-dsk01 (/MAISON/core2/iscsi/dsk01)] | | o- lun1 .................................................... [fileio/MAISON-core2-iscsi-dsk02 (/MAISON/core2/iscsi/dsk02)] | o- portals .............................. ...................................................................... [Portals: 2] | o- 10.10.1.129:3260 .............................. ................................................................... [OK] | o- 10.10.1.1:3260 .............................. ..................................................................... [OK] o- loopback ............................................................ ............................................. [Targets: 0] o- vhost ............................................................ ................................................ [Targets: 0] /> *Here is our RHEV-H hypervisor which has 3 nic configured.* *10.11.8.168/27 <http://10.11.8.168/27> for management and NFS* *10.10.1.11/25 <http://10.10.1.11/25> and 10.10.1.139/25 <http://10.10.1.139/25> are dedicated to ISCSI. The two are separated in different * *VLAN so as that 10.10.1.11/25 <http://10.10.1.11/25> will talk to 10.10.1.1 (iscsi server side) and 10.10.1.139/25 <http://10.10.1.139/25> will talk to 10.10.1.12 <http://10.10.1.129:3260/>9 (server side).* [root@pdx1cr207 ~]# ip a 1: lo: <LOOPBACK,UP,LOWER_UP> mtu 65536 qdisc noqueue state UNKNOWN qlen 1 link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00 inet 127.0.0.1/8 scope host lo valid_lft forever preferred_lft forever inet6 ::1/128 scope host valid_lft forever preferred_lft forever 2: eth0: <BROADCAST,MULTICAST,SLAVE,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast master bond0 state UP qlen 1000 link/ether 52:54:00:1c:25:02 brd ff:ff:ff:ff:ff:ff 3: eth1: <BROADCAST,MULTICAST,SLAVE,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast master bond0 state UP qlen 1000 link/ether 52:54:00:1c:25:02 brd ff:ff:ff:ff:ff:ff 4: eth2: <BROADCAST,MULTICAST,SLAVE,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast master bond1 state UP qlen 1000 link/ether 52:54:00:9a:7d:06 brd ff:ff:ff:ff:ff:ff 5: eth3: <BROADCAST,MULTICAST,SLAVE,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast master bond1 state UP qlen 1000 link/ether 52:54:00:9a:7d:06 brd ff:ff:ff:ff:ff:ff 6: eth4: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN qlen 1000 link/ether 52:54:00:e5:b8:42 brd ff:ff:ff:ff:ff:ff 7: eth5: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN qlen 1000 link/ether 52:54:00:e6:55:59 brd ff:ff:ff:ff:ff:ff 8: eth6: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast state UP qlen 1000 link/ether 52:54:00:9b:80:80 brd ff:ff:ff:ff:ff:ff inet 10.10.1.11/25 brd 10.10.1.127 scope global eth6 valid_lft forever preferred_lft forever inet6 fe80::5054:ff:fe9b:8080/64 scope link valid_lft forever preferred_lft forever 9: eth7: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc pfifo_fast state UP qlen 1000 link/ether 52:54:00:f7:5f:f8 brd ff:ff:ff:ff:ff:ff inet 10.10.1.139/25 brd 10.10.1.255 scope global eth7 valid_lft forever preferred_lft forever inet6 fe80::5054:ff:fef7:5ff8/64 scope link valid_lft forever preferred_lft forever 10: bond0: <BROADCAST,MULTICAST,MASTER,UP,LOWER_UP> mtu 1500 qdisc noqueue master ovirtmgmt state UP qlen 1000 link/ether 52:54:00:1c:25:02 brd ff:ff:ff:ff:ff:ff 11: ovirtmgmt: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc noqueue state UP qlen 1000 link/ether 52:54:00:1c:25:02 brd ff:ff:ff:ff:ff:ff inet 10.11.8.168/27 brd 10.11.8.191 scope global ovirtmgmt valid_lft forever preferred_lft forever 12: bond1: <BROADCAST,MULTICAST,MASTER,UP,LOWER_UP> mtu 1500 qdisc noqueue state UP qlen 1000 link/ether 52:54:00:9a:7d:06 brd ff:ff:ff:ff:ff:ff inet6 fe80::5054:ff:fe9a:7d06/64 scope link valid_lft forever preferred_lft forever 13: ;vdsmdummy;: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN qlen 1000 link/ether 7a:79:ea:86:d6:5d brd ff:ff:ff:ff:ff:ff *Here is our playbook used to prouve the issue. It points 10.10.1.1 (iscsi servers)* --- - hosts: rhevm2 gather_facts: False remote_user: root tasks: - name: authentication ovirt_auth: url: https://rhevm2.res1/ovirt-engine/api username: admin@internal password: rootroot01 insecure: True # - ovirt_storage_domains_facts: auth: url: https://rhevm2.res1/ovirt-engine/api token: "{{ ovirt_auth.token }}" insecure: True pattern: name=* and datacenter=default - debug: var: ovirt_storage_domains # - name: create iscsi_dsk02 ovirt_storage_domains: auth: url: https://rhevm2.res1/ovirt-engine/api token: "{{ ovirt_auth.token }}" insecure: True name: iscsi_dsk02 domain_function: data host: pdx1cr207 data_center: prslog iscsi: target: iqn.2017-06.stet.iscsi.server:target address: 10.10.1.1 lun_id: 3600140550738b53dd774303bfedac122 # lun_id: 36001405ed330f17f8e74ca1a08b7bb04 state: present destroy: true register: res_iscsi_dsk02 - debug: var: res_iscsi_dsk02 *We run the playbook and got error message:* ansible @sichuan workspace]$ansible-playbook ds_create6.yml -v Using /home/papa/environments/workspace/inventories/env_x1rop/ansible.cfg as config file PLAY [rhevm2] ******************************************************************************************************************************** TASK [authentication] ************************************************************************************************************************ ok: [rhevm2] => {"ansible_facts": {"ovirt_auth": {"ca_file": null, "compress": true, "insecure": true, "kerberos": false, "timeout": 0, "token ": "t1ibTsNmCh13nYhwpFpXJb5eK6YSIH-V4FdXGN2PKWq-g-F2ksAWjl51iYK0dB-2-yWHYvTGL6Xgo_2jtab2tA", "url": "https://rhevm2.res1/ovirt-engine/api"}}, "changed": false} TASK [ovirt_storage_domains_facts] *********************************************************************************************************** ok: [rhevm2] => {"ansible_facts": {"ovirt_storage_domains": [{"available": 95563022336, "committed": 0, "critical_space_action_blocker": 5, "d ata_centers": [{"id": "5950a8cc-03df-013e-02a8-0000000001ff"}], "disk_profiles": [], "disk_snapshots": [], "disks": [], "external_status": "ok ", "href": "/ovirt-engine/api/storagedomains/d2b84bb1-526a-4981-b6b4-f69b7e079d0d", "id": "d2b84bb1-526a-4981-b6b4-f69b7e079d0d", "master": tr ue, "name": "dm02", "permissions": [], "storage": {"address": "10.11.8.190", "nfs_version": "v3", "path": "/MAISON/nfs/domaines/dm02", "type": "nfs"}, "storage_connections": [], "storage_format": "v3", "templates": [], "type": "data", "used": 9663676416, "vms": [], "warning_low_space _indicator": 10, "wipe_after_delete": false}]}, "changed": false} TASK [debug] ********************************************************************************************************************************* ok: [rhevm2] => { "ovirt_storage_domains": [ { "available": 95563022336, "committed": 0, "critical_space_action_blocker": 5, "data_centers": [ { "id": "5950a8cc-03df-013e-02a8-0000000001ff" } ], "disk_profiles": [], "disk_snapshots": [], "disks": [], "external_status": "ok", "href": "/ovirt-engine/api/storagedomains/d2b84bb1-526a-4981-b6b4-f69b7e079d0d", "id": "d2b84bb1-526a-4981-b6b4-f69b7e079d0d", "master": true, "name": "dm02", "permissions": [], "storage": { "address": "10.11.8.190", "nfs_version": "v3", "path": "/MAISON/nfs/domaines/dm02", "type": "nfs" }, "storage_connections": [], "storage_format": "v3", "templates": [], "type": "data", "used": 9663676416, "vms": [], "warning_low_space_indicator": 10, "wipe_after_delete": false } ] } TASK [create iscsi_dsk02] ******************************************************************************************************************** An exception occurred during task execution. To see the full traceback, use -vvv. The error was: Error: Fault reason is "Operation Failed". Fa ult detail is "[Storage domain cannot be reached. Please ensure it is accessible from the host(s).]". HTTP response code is 400. fatal: [rhevm2]: FAILED! => {"changed": false, "failed": true, "msg": "Fault reason is \"Operation Failed\". Fault detail is \"[Storage domain cannot be reached. Please ensure it is accessible from the host(s).]\". HTTP response code is 400."} to retry, use: --limit @/home/papa/environments/workspace/ds_create6.retry PLAY RECAP *********************************************************************************************************************************** rhevm2 : ok=3 changed=0 unreachable=0 failed=1 *Here is compressed vdsm et engine log:* On Wed, Jun 28, 2017 at 3:28 PM, Yaniv Kaul <ykaul@redhat.com> wrote:
Can you share vdsm log?
On Wed, Jun 28, 2017 at 4:16 PM, victor 600833 <victor600833@gmail.com> wrote:
Hi ondra,
As you pointed. in your previous mail.
The iscsiadm discovered througth 10.11.8.190. It should have probed 10.10.1.1. This may be the root cause.
Victor,
If you lookup at our playbook.
- name: create iscsi_dsk02 ovirt_storage_domains: auth: url: https://rhevm2.res1/ovirt-engine/api token: "{{ ovirt_auth.token }}" insecure: True name: iscsi_dsk02 domain_function: data host: pdx1cr207 data_center: default iscsi: target: iqn.2017-06.stet.iscsi.server:target address: *10.10.1.1* lun_id: 3600140550738b53dd774303bfedac122 state: absent destroy: true register: res_iscsi_dsk02 - debug: var: res_iscsi_dsk02
On Wed, Jun 28, 2017 at 7:58 AM, Ondra Machacek <notifications@github.com
wrote:
It's using the host, which you will pass as host parameter to the task for iscsi discover.
As Yaniv said, can you please share your issue on our mailing list:
users@ovirt.org
There are people with knowledge of iSCSI and they will help you.
— You are receiving this because you were mentioned. Reply to this email directly, view it on GitHub <https://github.com/ansible/ansible/issues/25417#issuecomment-311563987>, or mute the thread <https://github.com/notifications/unsubscribe-auth/AScoAyV364pDjqXtCPSNodvhEeLw3b_Sks5sIet0gaJpZM4NyeaB> .
_______________________________________________ Users mailing list Users@ovirt.org http://lists.ovirt.org/mailman/listinfo/users