On Tue, Apr 2, 2019 at 3:35 PM Callum Smith <callum(a)well.ox.ac.uk> wrote:
Ok so on a clean NFS:
<virthyp04.virt.in.bmrc.ox.ac.uk> ESTABLISH SSH CONNECTION FOR USER: root
<virthyp04.virt.in.bmrc.ox.ac.uk> SSH: EXEC ssh -C -o ControlMaster=auto
-o ControlPersist=60s -o KbdInteractiveAuthentication=no -o
PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey
-o PasswordAuthentication=no -o User=root -o ConnectTimeout=10 -o
ControlPath=/etc/ansible/.ansible/cp/2c1e73363c
virthyp04.virt.in.bmrc.ox.ac.uk '/bin/sh -c '"'"'echo ~root
&& sleep
0'"'"''
<virthyp04.virt.in.bmrc.ox.ac.uk> (0, '/root\n', '')
<virthyp04.virt.in.bmrc.ox.ac.uk> ESTABLISH SSH CONNECTION FOR USER: root
<virthyp04.virt.in.bmrc.ox.ac.uk> SSH: EXEC ssh -C -o ControlMaster=auto
-o ControlPersist=60s -o KbdInteractiveAuthentication=no -o
PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey
-o PasswordAuthentication=no -o User=root -o ConnectTimeout=10 -o
ControlPath=/etc/ansible/.ansible/cp/2c1e73363c
virthyp04.virt.in.bmrc.ox.ac.uk '/bin/sh -c '"'"'( umask 77
&& mkdir -p
"` echo /root/.ansible/tmp/ansible-tmp-1554211873.65-5732618148991 `"
&&
echo ansible-tmp-1554211873.65-5732618148991="` echo
/root/.ansible/tmp/ansible-tmp-1554211873.65-5732618148991 `" ) && sleep
0'"'"''
<virthyp04.virt.in.bmrc.ox.ac.uk> (0,
'ansible-tmp-1554211873.65-5732618148991=/root/.ansible/tmp/ansible-tmp-1554211873.65-5732618148991\n',
'')
Using module file /opt/ansible/lib/ansible/modules/commands/command.py
<virthyp04.virt.in.bmrc.ox.ac.uk> PUT
/etc/ansible/.ansible/tmp/ansible-local-109444FfQz4/tmp0icLBP TO
/root/.ansible/tmp/ansible-tmp-1554211873.65-5732618148991/AnsiballZ_command.py
<virthyp04.virt.in.bmrc.ox.ac.uk> SSH: EXEC sftp -b - -C -o
ControlMaster=auto -o ControlPersist=60s -o KbdInteractiveAuthentication=no
-o
PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey
-o PasswordAuthentication=no -o User=root -o ConnectTimeout=10 -o
ControlPath=/etc/ansible/.ansible/cp/2c1e73363c '[
virthyp04.virt.in.bmrc.ox.ac.uk]'
<virthyp04.virt.in.bmrc.ox.ac.uk> (0, 'sftp> put
/etc/ansible/.ansible/tmp/ansible-local-109444FfQz4/tmp0icLBP
/root/.ansible/tmp/ansible-tmp-1554211873.65-5732618148991/AnsiballZ_command.py\n',
'')
<virthyp04.virt.in.bmrc.ox.ac.uk> ESTABLISH SSH CONNECTION FOR USER: root
<virthyp04.virt.in.bmrc.ox.ac.uk> SSH: EXEC ssh -C -o ControlMaster=auto
-o ControlPersist=60s -o KbdInteractiveAuthentication=no -o
PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey
-o PasswordAuthentication=no -o User=root -o ConnectTimeout=10 -o
ControlPath=/etc/ansible/.ansible/cp/2c1e73363c
virthyp04.virt.in.bmrc.ox.ac.uk '/bin/sh -c '"'"'chmod u+x
/root/.ansible/tmp/ansible-tmp-1554211873.65-5732618148991/
/root/.ansible/tmp/ansible-tmp-1554211873.65-5732618148991/AnsiballZ_command.py
&& sleep 0'"'"''
<virthyp04.virt.in.bmrc.ox.ac.uk> (0, '', '')
<virthyp04.virt.in.bmrc.ox.ac.uk> ESTABLISH SSH CONNECTION FOR USER: root
<virthyp04.virt.in.bmrc.ox.ac.uk> SSH: EXEC ssh -C -o ControlMaster=auto
-o ControlPersist=60s -o KbdInteractiveAuthentication=no -o
PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey
-o PasswordAuthentication=no -o User=root -o ConnectTimeout=10 -o
ControlPath=/etc/ansible/.ansible/cp/2c1e73363c -tt
virthyp04.virt.in.bmrc.ox.ac.uk '/bin/sh -c
'"'"'LANG=en_US.UTF-8
LC_ALL=en_US.UTF-8 LC_MESSAGES=en_US.UTF-8 /usr/bin/python
/root/.ansible/tmp/ansible-tmp-1554211873.65-5732618148991/AnsiballZ_command.py
&& sleep 0'"'"''
<virthyp04.virt.in.bmrc.ox.ac.uk> (1, '\r\n{"changed": true,
"end":
"2019-04-02 13:31:16.103973", "stdout": "",
"cmd": ["dd", "bs=20480",
"count=1", "oflag=direct",
"if=/var/tmp/localvmjOO3X6/1757281c-7c8e-407f-a18b-aecaf13e27a6",
"of=/rhev/data-center/mnt/10.141.15.248:_export_virtman_hosted__storage/d2d2255f-713c-4b32-9f63-680772435e00/images/01be5af0-c79a-402c-a98e-7da0714a23fe/1757281c-7c8e-407f-a18b-aecaf13e27a6"],
"failed": true, "delta": "0:00:00.005304",
"stderr": "dd: failed to open
\\u2018/rhev/data-center/mnt/10.141.15.248:_export_virtman_hosted__storage/d2d2255f-713c-4b32-9f63-680772435e00/images/01be5af0-c79a-402c-a98e-7da0714a23fe/1757281c-7c8e-407f-a18b-aecaf13e27a6\\u2019:
Permission denied", "rc": 1, "invocation":
{"module_args": {"warn": false,
"executable": null, "_uses_shell": false, "_raw_params":
"dd bs=20480
count=1 oflag=direct
if=\\"/var/tmp/localvmjOO3X6/1757281c-7c8e-407f-a18b-aecaf13e27a6\\"
of=\\"/rhev/data-center/mnt/10.141.15.248:_export_virtman_hosted__storage/d2d2255f-713c-4b32-9f63-680772435e00/images/01be5af0-c79a-402c-a98e-7da0714a23fe/1757281c-7c8e-407f-a18b-aecaf13e27a6\\"",
"removes": null, "argv": null, "creates": null,
"chdir": null, "stdin":
null}}, "start": "2019-04-02 13:31:16.098669", "msg":
"non-zero return
code"}\r\n', 'Shared connection to virthyp04.virt.in.bmrc.ox.ac.uk
closed.\r\n')
<virthyp04.virt.in.bmrc.ox.ac.uk> Failed to connect to the host via ssh:
Shared connection to virthyp04.virt.in.bmrc.ox.ac.uk closed.
<virthyp04.virt.in.bmrc.ox.ac.uk> ESTABLISH SSH CONNECTION FOR USER: root
<virthyp04.virt.in.bmrc.ox.ac.uk> SSH: EXEC ssh -C -o ControlMaster=auto
-o ControlPersist=60s -o KbdInteractiveAuthentication=no -o
PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey
-o PasswordAuthentication=no -o User=root -o ConnectTimeout=10 -o
ControlPath=/etc/ansible/.ansible/cp/2c1e73363c
virthyp04.virt.in.bmrc.ox.ac.uk '/bin/sh -c '"'"'rm -f -r
/root/.ansible/tmp/ansible-tmp-1554211873.65-5732618148991/ > /dev/null
2>&1 && sleep 0'"'"''
<virthyp04.virt.in.bmrc.ox.ac.uk> (0, '', '')
fatal: [virthyp04.virt.in.bmrc.ox.ac.uk]: FAILED! => {
"changed": true,
"cmd": [
"dd",
"bs=20480",
"count=1",
"oflag=direct",
"if=/var/tmp/localvmjOO3X6/1757281c-7c8e-407f-a18b-aecaf13e27a6",
"of=/rhev/data-center/mnt/10.141.15.248:
_export_virtman_hosted__storage/d2d2255f-713c-4b32-9f63-680772435e00/images/01be5af0-c79a-402c-a98e-7da0714a23fe/1757281c-7c8e-407f-a18b-aecaf13e27a6"
],
"delta": "0:00:00.005304",
"end": "2019-04-02 13:31:16.103973",
"invocation": {
"module_args": {
"_raw_params": "dd bs=20480 count=1 oflag=direct
if=\"/var/tmp/localvmjOO3X6/1757281c-7c8e-407f-a18b-aecaf13e27a6\"
of=\"/rhev/data-center/mnt/10.141.15.248:
_export_virtman_hosted__storage/d2d2255f-713c-4b32-9f63-680772435e00/images/01be5af0-c79a-402c-a98e-7da0714a23fe/1757281c-7c8e-407f-a18b-aecaf13e27a6\"",
"_uses_shell": false,
"argv": null,
"chdir": null,
"creates": null,
"executable": null,
"removes": null,
"stdin": null,
"warn": false
}
},
"msg": "non-zero return code",
"rc": 1,
"start": "2019-04-02 13:31:16.098669",
"stderr": "dd: failed to open
‘/rhev/data-center/mnt/10.141.15.248:_export_virtman_hosted__storage/d2d2255f-713c-4b32-9f63-680772435e00/images/01be5af0-c79a-402c-a98e-7da0714a23fe/1757281c-7c8e-407f-a18b-aecaf13e27a6’:
Permission denied",
"stderr_lines": [
"dd: failed to open
‘/rhev/data-center/mnt/10.141.15.248:_export_virtman_hosted__storage/d2d2255f-713c-4b32-9f63-680772435e00/images/01be5af0-c79a-402c-a98e-7da0714a23fe/1757281c-7c8e-407f-a18b-aecaf13e27a6’:
Permission denied"
],
"stdout": "",
"stdout_lines": []
}
ls -laZ /rhev/data-center/mnt/10.141.15.248:
_export_virtman_hosted__storage/d2d2255f-713c-4b32-9f63-680772435e00/images/01be5af0-c79a-402c-a98e-7da0714a23fe/
drwxrwxrwx. vdsm kvm system_u:object_r:nfs_t:s0 .
drwxrwxrwx. vdsm kvm system_u:object_r:nfs_t:s0 ..
-rw-rw----. vdsm kvm system_u:object_r:nfs_t:s0
1757281c-7c8e-407f-a18b-aecaf13e27a6
-rw-rw----. vdsm kvm system_u:object_r:nfs_t:s0
1757281c-7c8e-407f-a18b-aecaf13e27a6.lease
-rwxrwxrwx. vdsm kvm system_u:object_r:nfs_t:s0
1757281c-7c8e-407f-a18b-aecaf13e27a6.meta
Should this task be running as the `vdsm` user - I imagine adding a
become: true; and become_user: vdsm; would work?
Are you forcing
become: true
become_user: root
or something like that at playbook level?
I'd expect that the same directive at role level will win being more
specific but honestly I'm not 100% sure.
Regards,
Callum
--
Callum Smith
Research Computing Core
Wellcome Trust Centre for Human Genetics
University of Oxford
e. callum(a)well.ox.ac.uk
On 2 Apr 2019, at 10:21, Simone Tiraboschi <stirabos(a)redhat.com> wrote:
On Tue, Apr 2, 2019 at 11:18 AM Callum Smith <callum(a)well.ox.ac.uk> wrote:
> No, the NFS is full of artefacts - should i be rm -rf the whole thing
> every time?
>
Yes, right.
>
> Regards,
> Callum
>
> --
>
> Callum Smith
> Research Computing Core
> Wellcome Trust Centre for Human Genetics
> University of Oxford
> e. callum(a)well.ox.ac.uk
>
> On 2 Apr 2019, at 10:09, Simone Tiraboschi <stirabos(a)redhat.com> wrote:
>
> TASK [ovirt.hosted_engine_setup : Activate storage domain]
> ******************************************
> ...
> Error: Fault reason is "Operation Failed". Fault detail is "[]".
HTTP
> response code is 400.
>
> usually means that the engine failed to activate that storage domain;
> unfortunately engine error messages are not always that clear (see
>
https://bugzilla.redhat.com/1554922
> <
https://bugzilla.redhat.com/show_bug.cgi?id=1554922> ) but this is
> often due to fact the the NFS share or the iSCSI lun or whatever you used
> wasn't really clean.
> Are you manually cleaning it between one attempt and the next one?
>
> On Tue, Apr 2, 2019 at 10:50 AM Callum Smith <callum(a)well.ox.ac.uk>
> wrote:
>
>> Dear Simone,
>>
>> With no changes, we're now seeing this baffling error:
>>
>> TASK [ovirt.hosted_engine_setup : Parse OVF]
>> ********************************************************
>> task path:
>>
/etc/ansible/playbook/ovirt-ansible-hosted-engine-setup/tasks/create_storage_domain.yml:120
>> <virthyp04.virt.in.bmrc.ox.ac.uk> ESTABLISH SSH CONNECTION FOR USER:
>> root
>> <virthyp04.virt.in.bmrc.ox.ac.uk> SSH: EXEC ssh -C -o
>> ControlMaster=auto -o ControlPersist=60s -o KbdInteractiveAuthentication=no
>> -o PreferredAuthentications=
>> gssapi-with-mic,gssapi-keyex,hostbased,publickey -o
>> PasswordAuthentication=no -o User=root -o ConnectTimeout=10 -o
>> ControlPath=/etc/ansible/.ansible/cp/2c1e73
>> 363c virthyp04.virt.in.bmrc.ox.ac.uk '/bin/sh -c
'"'"'echo ~root &&
>> sleep 0'"'"''
>> <virthyp04.virt.in.bmrc.ox.ac.uk> (0, '/root\n', '')
>> <virthyp04.virt.in.bmrc.ox.ac.uk> ESTABLISH SSH CONNECTION FOR USER:
>> root
>> <virthyp04.virt.in.bmrc.ox.ac.uk> SSH: EXEC ssh -C -o
>> ControlMaster=auto -o ControlPersist=60s -o KbdInteractiveAuthentication=no
>> -o PreferredAuthentications=
>> gssapi-with-mic,gssapi-keyex,hostbased,publickey -o
>> PasswordAuthentication=no -o User=root -o ConnectTimeout=10 -o
>> ControlPath=/etc/ansible/.ansible/cp/2c1e73
>> 363c virthyp04.virt.in.bmrc.ox.ac.uk '/bin/sh -c '"'"'(
umask 77 &&
>> mkdir -p "` echo
>> /root/.ansible/tmp/ansible-tmp-1553937522.31-129798476242320 `" &&
echo a
>> nsible-tmp-1553937522.31-129798476242320="` echo
>> /root/.ansible/tmp/ansible-tmp-1553937522.31-129798476242320 `" ) &&
sleep
>> 0'"'"''
>> <virthyp04.virt.in.bmrc.ox.ac.uk> (0,
>>
'ansible-tmp-1553937522.31-129798476242320=/root/.ansible/tmp/ansible-tmp-1553937522.31-129798476242320\n',
>> '')
>> Using module file /opt/ansible/lib/ansible/modules/files/xml.py
>> <virthyp04.virt.in.bmrc.ox.ac.uk> PUT
>> /etc/ansible/.ansible/tmp/ansible-local-32213KmUe6/tmp8wMU8o TO
>> /root/.ansible/tmp/ansible-tmp-1553937522.31-12979847624
>> 2320/AnsiballZ_xml.py
>> <virthyp04.virt.in.bmrc.ox.ac.uk> SSH: EXEC sftp -b - -C -o
>> ControlMaster=auto -o ControlPersist=60s -o KbdInteractiveAuthentication=no
>> -o PreferredAuthentica
>> tions=gssapi-with-mic,gssapi-keyex,hostbased,publickey -o
>> PasswordAuthentication=no -o User=root -o ConnectTimeout=10 -o
>> ControlPath=/etc/ansible/.ansible/cp/
>> 2c1e73363c '[virthyp04.virt.in.bmrc.ox.ac.uk]'
>> <virthyp04.virt.in.bmrc.ox.ac.uk> (0, 'sftp> put
>> /etc/ansible/.ansible/tmp/ansible-local-32213KmUe6/tmp8wMU8o
>> /root/.ansible/tmp/ansible-tmp-1553937522.31-129
>> 798476242320/AnsiballZ_xml.py\n', '')
>> <virthyp04.virt.in.bmrc.ox.ac.uk> ESTABLISH SSH CONNECTION FOR USER:
>> root
>> <virthyp04.virt.in.bmrc.ox.ac.uk> SSH: EXEC ssh -C -o
>> ControlMaster=auto -o ControlPersist=60s -o KbdInteractiveAuthentication=no
>> -o PreferredAuthentications=
>> gssapi-with-mic,gssapi-keyex,hostbased,publickey -o
>> PasswordAuthentication=no -o User=root -o ConnectTimeout=10 -o
>> ControlPath=/etc/ansible/.ansible/cp/2c1e73
>> 363c virthyp04.virt.in.bmrc.ox.ac.uk '/bin/sh -c
'"'"'chmod u+x
>> /root/.ansible/tmp/ansible-tmp-1553937522.31-129798476242320/
>> /root/.ansible/tmp/ansible-tmp-1
>> 553937522.31-129798476242320/AnsiballZ_xml.py && sleep
0'"'"''
>> <virthyp04.virt.in.bmrc.ox.ac.uk> (0, '', '')
>> <virthyp04.virt.in.bmrc.ox.ac.uk> ESTABLISH SSH CONNECTION FOR USER:
>> root
>> <virthyp04.virt.in.bmrc.ox.ac.uk> SSH: EXEC ssh -C -o
>> ControlMaster=auto -o ControlPersist=60s -o KbdInteractiveAuthentication=no
>> -o PreferredAuthentications=
>> gssapi-with-mic,gssapi-keyex,hostbased,publickey -o
>> PasswordAuthentication=no -o User=root -o ConnectTimeout=10 -o
>> ControlPath=/etc/ansible/.ansible/cp/2c1e73
>> 363c -tt virthyp04.virt.in.bmrc.ox.ac.uk '/bin/sh -c
>> '"'"'/usr/bin/python
>> /root/.ansible/tmp/ansible-tmp-1553937522.31-129798476242320/AnsiballZ_xml.py
>> && sle
>> ep 0'"'"''
>> <virthyp04.virt.in.bmrc.ox.ac.uk> (0, '\r\n{"count": 1,
"matches":
>> [{"Disk":
{"{http://schemas.dmtf.org/ovf/envelope/1/}wipe-after-delete":
>> "false", "{http://
>>
schemas.dmtf.org/ovf/envelope/1/}format": "
>>
http://www.vmware.com/specifications/vmdk.html#sparse", "{
>>
http://schemas.dmtf.org/ovf/envelope/1/}vm_snapshot_id":
>> "5f2be758-82d7-4c07-a220-9060e782dc7a", "{
>>
http://schemas.dmtf.org/ovf/envelope/1/}parentRef": "", "{
>>
http://schemas.dmtf.org/ovf/envelope/1/}fileRef": "6f76686
>> b-199c-4cb3-bbbe-86fc34365745/72bc3948-5d8d-4877-bac8-7db4995045b5",
"{
>>
http://schemas.dmtf.org/ovf/envelope/1/}actual_size": "51",
"{
>>
http://schemas.dmtf.org/o
>> vf/envelope/1/}volume-format": "COW", "{
>>
http://schemas.dmtf.org/ovf/envelope/1/}boot": "true", "{
>>
http://schemas.dmtf.org/ovf/envelope/1/}size": "51",
"{http:/
>> /schemas.dmtf.org/ovf/envelope/1/}volume-type": "Sparse", "{
>>
http://schemas.dmtf.org/ovf/envelope/1/}disk-type": "System",
"{
>>
http://schemas.dmtf.org/ovf/envelo
>> pe/1/}diskId": "72bc3948-5d8d-4877-bac8-7db4995045b5", "{
>>
http://schemas.dmtf.org/ovf/envelope/1/}disk-interface":
"VirtIO"}}],
>> "changed": false, "actions": {"
>> xpath": "/ovf:Envelope/Section/Disk", "state":
"present", "namespaces":
>> {"vssd": "
>>
http://schemas.dmtf.org/wbem/wscim/1/cim-schema/2/CIM_VirtualSystemSettingDa
>> ta", "rasd": "
>>
http://schemas.dmtf.org/wbem/wscim/1/cim-schema/2/CIM_ResourceAllocationS...;,
>> "xsi": "http://www.w3.org/2001/XMLSchema-instance",
"ovf": "
>>
http://schemas.dmtf.org/ovf/envelope/1/"}}, "msg": 1,
"invocation":
>> {"module_args": {"xpath":
"/ovf:Envelope/Section/Disk", "count": false,
>> "set_children": null, "xmlstring": null,
"strip_cdata_tags": false,
>> "attribute": "ovf:size", "pretty_print": false,
"add_children": null,
>> "value": null, "content": "attribute",
"state": "present", "namespaces":
>> {"vssd": "
>>
http://schemas.dmtf.org/wbem/wscim/1/cim-schema/2/CIM_VirtualSystemSettin...;,
>> "rasd": "
>>
http://schemas.dmtf.org/wbem/wscim/1/cim-schema/2/CIM_ResourceAllocationS...;,
>> "xsi": "http://www.w3.org/2001/XMLSchema-instance",
"ovf": "
>>
http://schemas.dmtf.org/ovf/envelope/1/"}, "input_type":
"yaml",
>> "print_match": false, "path":
>>
"/var/tmp/localvmMNxnwL/master/vms/074a62d4-44f9-4ffe-a172-2702a9fe96df/074a62d4-44f9-4ffe-a172-2702a9fe96df.ovf",
>> "backup": false}}}\r\n', 'Shared connection to
>> virthyp04.virt.in.bmrc.ox.ac.uk closed.\r\n')
>> <virthyp04.virt.in.bmrc.ox.ac.uk> ESTABLISH SSH CONNECTION FOR USER:
>> root
>> <virthyp04.virt.in.bmrc.ox.ac.uk> SSH: EXEC ssh -C -o
>> ControlMaster=auto -o ControlPersist=60s -o KbdInteractiveAuthentication=no
>> -o
>> PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey
>> -o PasswordAuthentication=no -o User=root -o ConnectTimeout=10 -o
>> ControlPath=/etc/ansible/.ansible/cp/2c1e73363c
>> virthyp04.virt.in.bmrc.ox.ac.uk '/bin/sh -c '"'"'rm -f
-r
>> /root/.ansible/tmp/ansible-tmp-1553937522.31-129798476242320/ > /dev/null
>> 2>&1 && sleep 0'"'"''
>> <virthyp04.virt.in.bmrc.ox.ac.uk> (0, '', '')
>> ok: [virthyp04.virt.in.bmrc.ox.ac.uk] => {
>> "actions": {
>> "namespaces": {
>> "ovf":
"http://schemas.dmtf.org/ovf/envelope/1/",
>> "rasd": "
>>
http://schemas.dmtf.org/wbem/wscim/1/cim-schema/2/CIM_ResourceAllocationS...
>> ",
>> "vssd": "
>>
http://schemas.dmtf.org/wbem/wscim/1/cim-schema/2/CIM_VirtualSystemSettin...
>> ",
>> "xsi":
"http://www.w3.org/2001/XMLSchema-instance"
>> },
>> "state": "present",
>> "xpath": "/ovf:Envelope/Section/Disk"
>> },
>> "changed": false,
>> "count": 1,
>> "invocation": {
>> "module_args": {
>> "add_children": null,
>> "attribute": "ovf:size",
>> "backup": false,
>> "content": "attribute",
>> "count": false,
>> "input_type": "yaml",
>> "namespaces": {
>> "ovf":
"http://schemas.dmtf.org/ovf/envelope/1/",
>> "rasd": "
>>
http://schemas.dmtf.org/wbem/wscim/1/cim-schema/2/CIM_ResourceAllocationS...
>> ",
>> "vssd": "
>>
http://schemas.dmtf.org/wbem/wscim/1/cim-schema/2/CIM_VirtualSystemSettin...
>> ",
>> "xsi":
"http://www.w3.org/2001/XMLSchema-instance"
>> },
>> "path":
>>
"/var/tmp/localvmMNxnwL/master/vms/074a62d4-44f9-4ffe-a172-2702a9fe96df/074a62d4-44f9-4ffe-a172-2702a9fe96df.ovf",
>> "pretty_print": false,
>> "print_match": false,
>> "set_children": null,
>> "state": "present",
>> "strip_cdata_tags": false,
>> "value": null,
>> "xmlstring": null,
>> "xpath": "/ovf:Envelope/Section/Disk"
>> }
>> },
>> "matches": [
>> {
>> "Disk": {
>>
"{http://schemas.dmtf.org/ovf/envelope/1/}actual_size":
>> "51",
>> "{http://schemas.dmtf.org/ovf/envelope/1/}boot":
"true",
>>
"{http://schemas.dmtf.org/ovf/envelope/1/}disk-interface":
>> "VirtIO",
>> "{http://schemas.dmtf.org/ovf/envelope/1/}disk-type":
>> "System",
>> "{http://schemas.dmtf.org/ovf/envelope/1/}diskId":
>> "72bc3948-5d8d-4877-bac8-7db4995045b5",
>> "{http://schemas.dmtf.org/ovf/envelope/1/}fileRef":
>>
"6f76686b-199c-4cb3-bbbe-86fc34365745/72bc3948-5d8d-4877-bac8-7db4995045b5",
>> "{http://schemas.dmtf.org/ovf/envelope/1/}format":
"
>>
http://www.vmware.com/specifications/vmdk.html#sparse",
>> "{http://schemas.dmtf.org/ovf/envelope/1/}parentRef":
>> "",
>> "{http://schemas.dmtf.org/ovf/envelope/1/}size":
"51",
>>
"{http://schemas.dmtf.org/ovf/envelope/1/}vm_snapshot_id":
>> "5f2be758-82d7-4c07-a220-9060e782dc7a",
>>
"{http://schemas.dmtf.org/ovf/envelope/1/}volume-format":
>> "COW",
>>
"{http://schemas.dmtf.org/ovf/envelope/1/}volume-type":
>> "Sparse",
>>
"{http://schemas.dmtf.org/ovf/envelope/1/}wipe-after-delete":
>> "false"
>> }
>> }
>> ],
>> "msg": 1
>> }
>>
>> TASK [ovirt.hosted_engine_setup : Get required size]
>> ************************************************
>> task path:
>>
/etc/ansible/playbook/ovirt-ansible-hosted-engine-setup/tasks/create_storage_domain.yml:132
>> ok: [virthyp04.virt.in.bmrc.ox.ac.uk] => {
>> "ansible_facts": {
>> "required_size": "65498251264"
>> },
>> "changed": false
>> }
>>
>> TASK [ovirt.hosted_engine_setup : debug]
>> ************************************************************
>> task path:
>>
/etc/ansible/playbook/ovirt-ansible-hosted-engine-setup/tasks/create_storage_domain.yml:139
>> ok: [virthyp04.virt.in.bmrc.ox.ac.uk] => {
>> "required_size": "65498251264"
>> }
>>
>> TASK [ovirt.hosted_engine_setup : Remove unsuitable storage domain]
>> *********************************
>> task path:
>>
/etc/ansible/playbook/ovirt-ansible-hosted-engine-setup/tasks/create_storage_domain.yml:140
>> skipping: [virthyp04.virt.in.bmrc.ox.ac.uk] => {
>> "changed": false,
>> "skip_reason": "Conditional result was False"
>> }
>>
>> TASK [ovirt.hosted_engine_setup : debug]
>> ************************************************************
>> task path:
>>
/etc/ansible/playbook/ovirt-ansible-hosted-engine-setup/tasks/create_storage_domain.yml:151
>> ok: [virthyp04.virt.in.bmrc.ox.ac.uk] => {
>> "remove_storage_domain_details": {
>> "changed": false,
>> "skip_reason": "Conditional result was False",
>> "skipped": true
>> }
>> }
>>
>> TASK [ovirt.hosted_engine_setup : Check storage domain free space]
>> **********************************
>> task path:
>>
/etc/ansible/playbook/ovirt-ansible-hosted-engine-setup/tasks/create_storage_domain.yml:152
>> skipping: [virthyp04.virt.in.bmrc.ox.ac.uk] => {
>> "changed": false,
>> "skip_reason": "Conditional result was False"
>> }
>>
>> TASK [ovirt.hosted_engine_setup : Activate storage domain]
>> ******************************************
>> task path:
>>
/etc/ansible/playbook/ovirt-ansible-hosted-engine-setup/tasks/create_storage_domain.yml:161
>> <virthyp04.virt.in.bmrc.ox.ac.uk> ESTABLISH SSH CONNECTION FOR USER:
>> root
>> <virthyp04.virt.in.bmrc.ox.ac.uk> SSH: EXEC ssh -C -o
>> ControlMaster=auto -o ControlPersist=60s -o KbdInteractiveAuthentication=no
>> -o
>> PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey
>> -o PasswordAuthentication=no -o User=root -o ConnectTimeout=10 -o
>> ControlPath=/etc/ansible/.ansible/cp/2c1e73363c
>> virthyp04.virt.in.bmrc.ox.ac.uk '/bin/sh -c '"'"'echo
~root && sleep
>> 0'"'"''
>> <virthyp04.virt.in.bmrc.ox.ac.uk> (0, '/root\n', '')
>> <virthyp04.virt.in.bmrc.ox.ac.uk> ESTABLISH SSH CONNECTION FOR USER:
>> root
>> <virthyp04.virt.in.bmrc.ox.ac.uk> SSH: EXEC ssh -C -o
>> ControlMaster=auto -o ControlPersist=60s -o KbdInteractiveAuthentication=no
>> -o
>> PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey
>> -o PasswordAuthentication=no -o User=root -o ConnectTimeout=10 -o
>> ControlPath=/etc/ansible/.ansible/cp/2c1e73363c
>> virthyp04.virt.in.bmrc.ox.ac.uk '/bin/sh -c '"'"'(
umask 77 && mkdir -p
>> "` echo /root/.ansible/tmp/ansible-tmp-1553937525.89-247224387363848 `"
&&
>> echo ansible-tmp-1553937525.89-247224387363848="` echo
>> /root/.ansible/tmp/ansible-tmp-1553937525.89-247224387363848 `" ) &&
sleep
>> 0'"'"''
>> <virthyp04.virt.in.bmrc.ox.ac.uk> (0,
>>
'ansible-tmp-1553937525.89-247224387363848=/root/.ansible/tmp/ansible-tmp-1553937525.89-247224387363848\n',
>> '')
>> Using module file
>> /opt/ansible/lib/ansible/modules/cloud/ovirt/ovirt_storage_domain.py
>> <virthyp04.virt.in.bmrc.ox.ac.uk> PUT
>> /etc/ansible/.ansible/tmp/ansible-local-32213KmUe6/tmpQtVJtM TO
>>
/root/.ansible/tmp/ansible-tmp-1553937525.89-247224387363848/AnsiballZ_ovirt_storage_domain.py
>> <virthyp04.virt.in.bmrc.ox.ac.uk> SSH: EXEC sftp -b - -C -o
>> ControlMaster=auto -o ControlPersist=60s -o KbdInteractiveAuthentication=no
>> -o
>> PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey
>> -o PasswordAuthentication=no -o User=root -o ConnectTimeout=10 -o
>> ControlPath=/etc/ansible/.ansible/cp/2c1e73363c '[
>> virthyp04.virt.in.bmrc.ox.ac.uk]'
>> <virthyp04.virt.in.bmrc.ox.ac.uk> (0, 'sftp> put
>> /etc/ansible/.ansible/tmp/ansible-local-32213KmUe6/tmpQtVJtM
>>
/root/.ansible/tmp/ansible-tmp-1553937525.89-247224387363848/AnsiballZ_ovirt_storage_domain.py\n',
>> '')
>> <virthyp04.virt.in.bmrc.ox.ac.uk> ESTABLISH SSH CONNECTION FOR USER:
>> root
>> <virthyp04.virt.in.bmrc.ox.ac.uk> SSH: EXEC ssh -C -o
>> ControlMaster=auto -o ControlPersist=60s -o KbdInteractiveAuthentication=no
>> -o
>> PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey
>> -o PasswordAuthentication=no -o User=root -o ConnectTimeout=10 -o
>> ControlPath=/etc/ansible/.ansible/cp/2c1e73363c
>> virthyp04.virt.in.bmrc.ox.ac.uk '/bin/sh -c '"'"'chmod
u+x
>> /root/.ansible/tmp/ansible-tmp-1553937525.89-247224387363848/
>>
/root/.ansible/tmp/ansible-tmp-1553937525.89-247224387363848/AnsiballZ_ovirt_storage_domain.py
>> && sleep 0'"'"''
>> <virthyp04.virt.in.bmrc.ox.ac.uk> (0, '', '')
>> <virthyp04.virt.in.bmrc.ox.ac.uk> ESTABLISH SSH CONNECTION FOR USER:
>> root
>> <virthyp04.virt.in.bmrc.ox.ac.uk> SSH: EXEC ssh -C -o
>> ControlMaster=auto -o ControlPersist=60s -o KbdInteractiveAuthentication=no
>> -o
>> PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey
>> -o PasswordAuthentication=no -o User=root -o ConnectTimeout=10 -o
>> ControlPath=/etc/ansible/.ansible/cp/2c1e73363c -tt
>> virthyp04.virt.in.bmrc.ox.ac.uk '/bin/sh -c
'"'"'/usr/bin/python
>>
/root/.ansible/tmp/ansible-tmp-1553937525.89-247224387363848/AnsiballZ_ovirt_storage_domain.py
>> && sleep 0'"'"''
>> <virthyp04.virt.in.bmrc.ox.ac.uk> (1, '\r\n{"msg":
"Fault reason is
>> \\"Operation Failed\\". Fault detail is \\"[]\\". HTTP
response code is
>> 400.", "failed": true, "exception": "Traceback
(most recent call last):\\n
>> File
\\"/tmp/ansible_ovirt_storage_domain_payload_w8oO0Y/__main__.py\\",
>> line 664, in main\\n storage_domains_module.post_create_check(sd_id)\\n
>> File
\\"/tmp/ansible_ovirt_storage_domain_payload_w8oO0Y/__main__.py\\",
>> line 526, in post_create_check\\n id=storage_domain.id,\\n File
>> \\"/usr/lib64/python2.7/site-packages/ovirtsdk4/services.py\\", line
3053,
>> in add\\n return self._internal_add(storage_domain, headers, query,
>> wait)\\n File
>> \\"/usr/lib64/python2.7/site-packages/ovirtsdk4/service.py\\", line
232, in
>> _internal_add\\n return future.wait() if wait else future\\n File
>> \\"/usr/lib64/python2.7/site-packages/ovirtsdk4/service.py\\", line 55,
in
>> wait\\n return self._code(response)\\n File
>> \\"/usr/lib64/python2.7/site-packages/ovirtsdk4/service.py\\", line
229, in
>> callback\\n self._check_fault(response)\\n File
>> \\"/usr/lib64/python2.7/site-packages/ovirtsdk4/service.py\\", line
132, in
>> _check_fault\\n self._raise_error(response, body)\\n File
>> \\"/usr/lib64/python2.7/site-packages/ovirtsdk4/service.py\\", line
118, in
>> _raise_error\\n raise error\\nError: Fault reason is \\"Operation
>> Failed\\". Fault detail is \\"[]\\". HTTP response code is
400.\\n",
>> "invocation": {"module_args": {"comment": null,
"warning_low_space": null,
>> "fetch_nested": false, "localfs": null,
"data_center": "Default", "id":
>> null, "iscsi": null, "state": "present",
"wipe_after_delete": null,
>> "destroy": null, "fcp": null, "description": null,
"format": null,
>> "nested_attributes": [], "host":
"virthyp04.virt.in.bmrc.ox.ac.uk",
>> "discard_after_delete": null, "wait": true,
"domain_function": "data",
>> "name": "hosted_storage",
"critical_space_action_blocker": null, "posixfs":
>> null, "poll_interval": 3, "glusterfs": null, "nfs":
null, "timeout": 180,
>> "backup": null}}}\r\n', 'Shared connection to
>> virthyp04.virt.in.bmrc.ox.ac.uk closed.\r\n')
>> <virthyp04.virt.in.bmrc.ox.ac.uk> Failed to connect to the host via
>> ssh: Shared connection to virthyp04.virt.in.bmrc.ox.ac.uk closed.
>> <virthyp04.virt.in.bmrc.ox.ac.uk> ESTABLISH SSH CONNECTION FOR USER:
>> root
>> <virthyp04.virt.in.bmrc.ox.ac.uk> SSH: EXEC ssh -C -o
>> ControlMaster=auto -o ControlPersist=60s -o KbdInteractiveAuthentication=no
>> -o
>> PreferredAuthentications=gssapi-with-mic,gssapi-keyex,hostbased,publickey
>> -o PasswordAuthentication=no -o User=root -o ConnectTimeout=10 -o
>> ControlPath=/etc/ansible/.ansible/cp/2c1e73363c
>> virthyp04.virt.in.bmrc.ox.ac.uk '/bin/sh -c '"'"'rm -f
-r
>> /root/.ansible/tmp/ansible-tmp-1553937525.89-247224387363848/ > /dev/null
>> 2>&1 && sleep 0'"'"''
>> <virthyp04.virt.in.bmrc.ox.ac.uk> (0, '', '')
>> The full traceback is:
>> Traceback (most recent call last):
>> File "/tmp/ansible_ovirt_storage_domain_payload_w8oO0Y/__main__.py",
>> line 664, in main
>> storage_domains_module.post_create_check(sd_id)
>> File "/tmp/ansible_ovirt_storage_domain_payload_w8oO0Y/__main__.py",
>> line 526, in post_create_check
>> id=storage_domain.id,
>> File "/usr/lib64/python2.7/site-packages/ovirtsdk4/services.py",
line
>> 3053, in add
>> return self._internal_add(storage_domain, headers, query, wait)
>> File "/usr/lib64/python2.7/site-packages/ovirtsdk4/service.py", line
>> 232, in _internal_add
>> return future.wait() if wait else future
>> File "/usr/lib64/python2.7/site-packages/ovirtsdk4/service.py", line
>> 55, in wait
>> return self._code(response)
>> File "/usr/lib64/python2.7/site-packages/ovirtsdk4/service.py", line
>> 229, in callback
>> self._check_fault(response)
>> File "/usr/lib64/python2.7/site-packages/ovirtsdk4/service.py", line
>> 132, in _check_fault
>> self._raise_error(response, body)
>> File "/usr/lib64/python2.7/site-packages/ovirtsdk4/service.py", line
>> 118, in _raise_error
>> raise error
>> Error: Fault reason is "Operation Failed". Fault detail is
"[]". HTTP
>> response code is 400.
>> fatal: [virthyp04.virt.in.bmrc.ox.ac.uk]: FAILED! => {
>> "changed": false,
>> "invocation": {
>> "module_args": {
>> "backup": null,
>> "comment": null,
>> "critical_space_action_blocker": null,
>> "data_center": "Default",
>> "description": null,
>> "destroy": null,
>> "discard_after_delete": null,
>> "domain_function": "data",
>> "fcp": null,
>> "fetch_nested": false,
>> "format": null,
>> "glusterfs": null,
>> "host": "virthyp04.virt.in.bmrc.ox.ac.uk",
>> "id": null,
>> "iscsi": null,
>> "localfs": null,
>> "name": "hosted_storage",
>> "nested_attributes": [],
>> "nfs": null,
>> "poll_interval": 3,
>> "posixfs": null,
>> "state": "present",
>> "timeout": 180,
>> "wait": true,
>> "warning_low_space": null,
>> "wipe_after_delete": null
>> }
>> },
>> "msg": "Fault reason is \"Operation Failed\". Fault
detail is
>> \"[]\". HTTP response code is 400."
>> }
>> to retry, use: --limit @/etc/ansible/playbook/ovirt.retry
>>
>> PLAY RECAP
>>
******************************************************************************************
>> virthyp04.virt.in.bmrc.ox.ac.uk : ok=216 changed=69 unreachable=0
>> failed=1
>>
>> Regards,
>> Callum
>>
>> --
>>
>> Callum Smith
>> Research Computing Core
>> Wellcome Trust Centre for Human Genetics
>> University of Oxford
>> e. callum(a)well.ox.ac.uk
>>
>> On 29 Mar 2019, at 17:48, Simone Tiraboschi <stirabos(a)redhat.com> wrote:
>>
>>
>>
>> On Fri, Mar 29, 2019 at 6:14 PM Callum Smith <callum(a)well.ox.ac.uk>
>> wrote:
>>
>>> So close now:
>>>
>>> fatal: [virthyp04.virt.in.bmrc.ox.ac.uk]: FAILED! => {"changed":
true,
>>> "cmd": ["dd", "bs=20480", "count=1",
"oflag=direct",
>>> "if=/var/tmp/localvmDBMVgn/e208e0f9-0f4d-4d0d-9104-10d8a26bfab6",
>>>
"of=/rhev/data-center/mnt/10.141.15.248:_export_virtman_hosted__storage/b4f93b28-1497-44b0-9eaf-5e5e2b71bce8/images/645c4286-71e4-4cce-9049-345903929e1b/e208e0f9-0f4d-4d0d-9104-10d8a26bfab6"],
>>> "delta": "0:00:00.005134", "end":
"2019-03-29 17:04:27.952367", "msg":
>>> "non-zero return code", "rc": 1, "start":
"2019-03-29 17:04:27.947233",
>>> "stderr": "dd: failed to open
‘/rhev/data-center/mnt/10.141.15.248:_export_virtman_hosted__storage/b4f93b28-1497-44b0-9eaf-5e5e2b71bce8/images/645c4286-71e4-4cce-9049-345903929e1b/e208e0f9-0f4d-4d0d-9104-10d8a26bfab6’:
>>> Permission denied", "stderr_lines": ["dd: failed to open
>>>
‘/rhev/data-center/mnt/10.141.15.248:_export_virtman_hosted__storage/b4f93b28-1497-44b0-9eaf-5e5e2b71bce8/images/645c4286-71e4-4cce-9049-345903929e1b/e208e0f9-0f4d-4d0d-9104-10d8a26bfab6’:
>>> Permission denied"], "stdout": "",
"stdout_lines": []}
>>> to retry, use: --limit @/etc/ansible/playbook/ovirt.retry
>>>
>>> ls -laZ /rhev/data-center/mnt/10.141.15.248:
>>>
_export_virtman_hosted__storage/b4f93b28-1497-44b0-9eaf-5e5e2b71bce8/images/645c4286-71e4-4cce-9049-345903929e1b/e208e0f9-0f4d-4d0d-9104-10d8a26bfab6
>>> -rw-rw----. vdsm kvm system_u:object_r:nfs_t:s0
>>> /rhev/data-center/mnt/10.141.15.248:
>>>
_export_virtman_hosted__storage/b4f93b28-1497-44b0-9eaf-5e5e2b71bce8/images/645c4286-71e4-4cce-9049-345903929e1b/e208e0f9-0f4d-4d0d-9104-10d8a26bfab6
>>>
>>> Any ideas on this one? I can almost touch this deployment now...
>>> Looking at the command it should run as `vdsm` so should work fine, could
>>> this be SELinux?
>>>
>>
>> Yes, it should be executed as vdsm user:
>>
>>
https://github.com/oVirt/ovirt-ansible-hosted-engine-setup/blob/master/ta...
>>
>> Did you tried executing it with -vvv ?
>> Can you please share the playbook you are using to trigger that role?
>>
>>
>>
>>>
>>> Regards,
>>> Callum
>>>
>>> --
>>>
>>> Callum Smith
>>> Research Computing Core
>>> Wellcome Trust Centre for Human Genetics
>>> University of Oxford
>>> e. callum(a)well.ox.ac.uk
>>>
>>> On 29 Mar 2019, at 15:50, Callum Smith <callum(a)well.ox.ac.uk> wrote:
>>>
>>> Guilty, will roll back and try again!
>>>
>>> Regards,
>>> Callum
>>>
>>> --
>>>
>>> Callum Smith
>>> Research Computing Core
>>> Wellcome Trust Centre for Human Genetics
>>> University of Oxford
>>> e. callum(a)well.ox.ac.uk
>>>
>>> On 29 Mar 2019, at 15:35, Simone Tiraboschi <stirabos(a)redhat.com>
>>> wrote:
>>>
>>> The error comes from here:
>>>
>>> TASK [ovirt.hosted_engine_setup : Parse OVF]
>>>
***************************************************************************************************************************************
>>> fatal: [virthyp04.virt.in.bmrc.ox.ac.uk]: FAILED! =>
{"changed":
>>> false, "msg": "missing parameter(s) required by
'attribute': value"}
>>>
>>> but are you really using it with ansible 2.8 alpha 1?
>>> I'd strongly suggest to switch back to a stable release of ansible
>>> which is currently 2.7.9.
>>>
>>> That one was due to:
https://github.com/ansible/ansible/issues/53459
>>> In the next ansible build it will be just a warning as for:
>>>
https://github.com/ansible/ansible/pull/54336
>>>
>>>
>>>
https://github.com/oVirt/ovirt-ansible-hosted-engine-setup/pull/150/files
>>> already address this on ovirt-ansible-hosted-engine-setup to be
>>> compatible with future ansible releases.
>>>
>>>
>>> On Fri, Mar 29, 2019 at 3:53 PM Callum Smith <callum(a)well.ox.ac.uk>
>>> wrote:
>>>
>>>> The OVF in question is here:
>>>>
>>>> <ovf:Envelope ovf:version="0.9" xmlns:ovf="
>>>>
http://schemas.dmtf.org/ovf/envelope/1/" xmlns:rasd="
>>>>
http://schemas.dmtf.org/wbem/wscim/1/cim-schema/2/CIM_ResourceAllocationS...
>>>> xmlns:vssd="
>>>>
http://schemas.dmtf.org/wbem/wscim/1/cim-schema/2/CIM_VirtualSystemSettin...
>>>>
xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance"><Re...
>>>> ovf:description="074a62d4-44f9-4ffe-a172-2702a9fe96df"
>>>>
ovf:href="6f76686b-199c-4cb3-bbbe-86fc34365745/72bc3948-5d8d-4877-bac8-7db4995045b5"
>>>> ovf:id="72bc3948-5d8d-4877-bac8-7db4995045b5"
ovf:size="54760833024"
>>>> /></References><Section
xsi:type="ovf:NetworkSection_Type"><Info>List of
>>>> Networks</Info></Section><Section
xsi:type="ovf:DiskSection_Type"><Disk
>>>> ovf:actual_size="51" ovf:boot="true"
ovf:disk-interface="VirtIO"
>>>> ovf:disk-type="System"
ovf:diskId="72bc3948-5d8d-4877-bac8-7db4995045b5"
>>>>
ovf:fileRef="6f76686b-199c-4cb3-bbbe-86fc34365745/72bc3948-5d8d-4877-bac8-7db4995045b5"
>>>>
ovf:format="http://www.vmware.com/specifications/vmdk.html#sparse"
>>>> ovf:parentRef="" ovf:size="51"
>>>> ovf:vm_snapshot_id="5f2be758-82d7-4c07-a220-9060e782dc7a"
>>>> ovf:volume-format="COW" ovf:volume-type="Sparse"
>>>> ovf:wipe-after-delete="false" /></Section><Content
ovf:id="out"
>>>>
xsi:type="ovf:VirtualSystem_Type"><Name>074a62d4-44f9-4ffe-a172-2702a9fe96df</Name><TemplateId>074a62d4-44f9-4ffe-a172-2702a9fe96df</TemplateId><Description>Created
>>>> by OVABuilder</Description><Domain
/><CreationDate>2019/03/19
>>>> 08:35:09</CreationDate><TimeZone
>>>>
/><IsAutoSuspend>false</IsAutoSuspend><VmType>1</VmType><default_display_type>0</default_display_type><default_boot_sequence>1</default_boot_sequence><Section
>>>> ovf:id="074a62d4-44f9-4ffe-a172-2702a9fe96df"
ovf:required="false"
>>>>
xsi:type="ovf:OperatingSystemSection_Type"><Info>Guest
>>>>
OS</Info><Description>OtherLinux</Description></Section><Section
>>>> xsi:type="ovf:VirtualHardwareSection_Type"><Info>4
CPU, 16384
>>>> Memory</Info><System><vssd:VirtualSystemType>RHEVM
>>>>
4.6.0.163</vssd:VirtualSystemType></System><Item><rasd:Caption>4
virtual
>>>> CPU</rasd:Caption><rasd:Description>Number of virtual
>>>>
CPU</rasd:Description><rasd:InstanceId>1</rasd:InstanceId><rasd:ResourceType>3</rasd:ResourceType><rasd:num_of_sockets>1</rasd:num_of_sockets><rasd:cpu_per_socket>4</rasd:cpu_per_socket></Item><Item><rasd:Caption>16384
>>>> MB of memory</rasd:Caption><rasd:Description>Memory
>>>>
Size</rasd:Description><rasd:InstanceId>2</rasd:InstanceId><rasd:ResourceType>4</rasd:ResourceType><rasd:AllocationUnits>MegaBytes</rasd:AllocationUnits><rasd:VirtualQuantity>16384</rasd:VirtualQuantity></Item><Item><rasd:Caption>Drive
>>>>
1</rasd:Caption><rasd:InstanceId>72bc3948-5d8d-4877-bac8-7db4995045b5</rasd:InstanceId><rasd:ResourceType>17</rasd:ResourceType><rasd:HostResource>6f76686b-199c-4cb3-bbbe-86fc34365745/72bc3948-5d8d-4877-bac8-7db4995045b5</rasd:HostResource><rasd:Parent>00000000-0000-0000-0000-000000000000</rasd:Parent><rasd:Template>00000000-0000-0000-0000-000000000000</rasd:Template><rasd:ApplicationList
>>>>
/><rasd:StorageId>00000000-0000-0000-0000-000000000000</rasd:StorageId><rasd:StoragePoolId>00000000-0000-0000-0000-000000000000</rasd:StoragePoolId><rasd:CreationDate>2019/03/19
>>>> 08:35:09</rasd:CreationDate><rasd:LastModified>2019/03/19
>>>>
08:35:09</rasd:LastModified></Item><Item><rasd:Caption>Ethernet 0
>>>>
rhevm</rasd:Caption><rasd:InstanceId>3</rasd:InstanceId><rasd:ResourceType>10</rasd:ResourceType><rasd:ResourceSubType>3</rasd:ResourceSubType><rasd:Connection>rhevm</rasd:Connection><rasd:Name>eth0</rasd:Name><rasd:speed>1000</rasd:speed></Item><Item><rasd:Caption>Graphics</rasd:Caption><rasd:InstanceId>5</rasd:InstanceId><rasd:ResourceType>20</rasd:ResourceType><rasd:VirtualQuantity>1</rasd:VirtualQuantity></Item></Section></Content></ovf:Envelope>
>>>>
>>>> Regards,
>>>> Callum
>>>>
>>>> --
>>>>
>>>> Callum Smith
>>>> Research Computing Core
>>>> Wellcome Trust Centre for Human Genetics
>>>> University of Oxford
>>>> e. callum(a)well.ox.ac.uk
>>>>
>>>> On 29 Mar 2019, at 14:42, Callum Smith <callum(a)well.ox.ac.uk>
wrote:
>>>>
>>>> Ok so we're getting very close now, weird OVF error:
>>>>
>>>> Full ansible log attached
>>>>
>>>> Only error in the engine.log looks normal/expected to me:
>>>> 2019-03-29 14:32:44,370Z ERROR
>>>> [org.ovirt.engine.core.bll.pm.FenceProxyLocator]
>>>> (EE-ManagedThreadFactory-engineScheduled-Thread-71) [4405d6db] Can not
run
>>>> fence action on host 'vir
>>>> thyp04.virt.in.bmrc.ox.ac.uk', no suitable proxy host was found.
>>>>
>>>> <ovirt.20190329133000.ansible.log>
>>>>
>>>> Feeling damn close to success here, but have managed to replicate this
>>>> issue twice re-running the installer.
>>>>
>>>> Regards,
>>>> Callum
>>>>
>>>> --
>>>>
>>>> Callum Smith
>>>> Research Computing Core
>>>> Wellcome Trust Centre for Human Genetics
>>>> University of Oxford
>>>> e. callum(a)well.ox.ac.uk
>>>>
>>>> On 29 Mar 2019, at 11:50, Simone Tiraboschi <stirabos(a)redhat.com>
>>>> wrote:
>>>>
>>>>
>>>>
>>>> On Fri, Mar 29, 2019 at 12:36 PM Callum Smith
<callum(a)well.ox.ac.uk>
>>>> wrote:
>>>>
>>>>> ip link del ovirtmgmt has done the job....
>>>>>
>>>>> Another issue, but this is likely due to randomised MAC addresses:
>>>>> fatal: [virthyp04.virt.in.bmrc.ox.ac.uk]: FAILED! =>
{"changed":
>>>>> true, "cmd": ["virt-install", "-n",
"HostedEngineLocal", "--os-variant",
>>>>> "rhel7", "--virt-type", "kvm",
"--memory", "4096", "--vcpus", "64",
>>>>> "--network",
"network=default,mac=fe:58:6c:da:1e:cc,model=virtio",
>>>>> "--disk",
>>>>>
"/var/tmp/localvmOCYiyF/images/6f76686b-199c-4cb3-bbbe-86fc34365745/72bc3948-5d8d-4877-bac8-7db4995045b5",
>>>>> "--import", "--disk",
"path=/var/tmp/localvmOCYiyF/seed.iso,device=cdrom",
>>>>> "--noautoconsole", "--rng",
"/dev/random", "--graphics", "vnc", "--video",
>>>>> "vga", "--sound", "none",
"--controller", "usb,model=none", "--memballoon",
>>>>> "none", "--boot", "hd,menu=off",
"--clock", "kvmclock_present=yes"],
>>>>> "delta": "0:00:01.355834", "end":
"2019-03-29 11:31:02.100143", "msg":
>>>>> "non-zero return code", "rc": 1,
"start": "2019-03-29 11:31:00.744309",
>>>>> "stderr": "ERROR unsupported configuration: Unable
to use MAC address
>>>>> starting with reserved value 0xFE - 'fe:58:6c:da:1e:cc' -
\nDomain
>>>>> installation does not appear to have been successful.\nIf it was, you
can
>>>>> restart your domain by running:\n virsh --connect qemu:///system
start
>>>>> HostedEngineLocal\notherwise, please restart your
installation.",
>>>>> "stderr_lines": ["ERROR unsupported configuration:
Unable to use MAC
>>>>> address starting with reserved value 0xFE -
'fe:58:6c:da:1e:cc' - ",
>>>>> "Domain installation does not appear to have been
successful.", "If it was,
>>>>> you can restart your domain by running:", " virsh
--connect
>>>>> qemu:///system start HostedEngineLocal", "otherwise, please
restart
>>>>> your installation."], "stdout": "\nStarting
install...", "stdout_lines":
>>>>> ["", "Starting install..."]}
>>>>>
>>>>> Seems it doesn't take into account reserved values when
generating.
>>>>>
>>>>
>>>> If not specified by the user,
>>>> a random unicast MAC address is randomly generated here:
>>>>
>>>>
https://github.com/oVirt/ovirt-ansible-hosted-engine-setup/blob/master/ta...
>>>>
>>>> for sure it's unicast but maybe we should make it more robust on
>>>> reserved values.
>>>> Simply try again for now; I'll open an issue to track it, thanks!
>>>>
>>>>
>>>>>
>>>>> I hope this feedback is valuable - I have a good feeling about the
>>>>> current deploy otherwise though.
>>>>>
>>>>> Regards,
>>>>> Callum
>>>>>
>>>>> --
>>>>>
>>>>> Callum Smith
>>>>> Research Computing Core
>>>>> Wellcome Trust Centre for Human Genetics
>>>>> University of Oxford
>>>>> e. callum(a)well.ox.ac.uk
>>>>>
>>>>> On 29 Mar 2019, at 11:01, Simone Tiraboschi
<stirabos(a)redhat.com>
>>>>> wrote:
>>>>>
>>>>>
>>>>>
>>>>> On Fri, Mar 29, 2019 at 11:56 AM Callum Smith
<callum(a)well.ox.ac.uk>
>>>>> wrote:
>>>>>
>>>>>> Dear Simone,
>>>>>>
>>>>>> It doesn't seem to want to work:
>>>>>>
>>>>>> # Settings
>>>>>> he_fqdn: "he.virt.in.bmrc.ox.ac.uk"
>>>>>> he_ansible_host_name:
"virthyp04.virt.in.bmrc.ox.ac.uk"
>>>>>> he_admin_password: <snip>
>>>>>> he_appliance_password: <snip>
>>>>>>
>>>>>> # Resources
>>>>>> he_mem_size_MB: "4096"
>>>>>>
>>>>>> # Storage
>>>>>> he_domain_type: "nfs"
>>>>>> he_storage_domain_addr: <snip>
>>>>>> he_storage_domain_path: <snip>
>>>>>>
>>>>>> # Network
>>>>>> he_vm_ip_addr: "10.141.31.240"
>>>>>> he_vm_ip_prefix: "20"
>>>>>> he_dns_addr:
["10.141.31.251","10.141.31.252","10.141.31.253"]
>>>>>> he_default_gateway_4: "10.141.31.254"
>>>>>> he_gateway: he_default_gateway_4
>>>>>> he_force_ip4: true
>>>>>> he_bridge_if: bond0.910
>>>>>> #he_just_collect_network_interfaces: true
>>>>>>
>>>>>> # Email
>>>>>> he_smtp_port: 25
>>>>>> he_smtp_server: smtp.ox.ac.uk
>>>>>> he_dest_email: rescomp-ops(a)well.ox.ac.uk
>>>>>> he_source_email: ovirt(a)bmrc.ox.ac.uk
>>>>>>
>>>>>> # Ansible Stuff
>>>>>> ansible_ssh_user: root
>>>>>> ansible_become: false
>>>>>>
>>>>>> host_key_checking: false
>>>>>>
>>>>>> I've attached the output of the ansible command as a log
file, this
>>>>>> is what happens when the IF bond0.910 is assigned the IP and
`ovirtmgmt` is
>>>>>> not defined on the host.
>>>>>>
>>>>>
>>>>>
>>>>> TASK [ovirt.hosted_engine_setup : debug]
>>>>>
*******************************************************************************************************************************************
>>>>> ok: [virthyp04.virt.in.bmrc.ox.ac.uk] => {
>>>>> "target_address_v4": {
>>>>> "changed": true,
>>>>> "cmd": "ip addr show ovirtmgmt | grep
'inet ' | cut -d' ' -f6
>>>>> | cut -d'/' -f1",
>>>>> "delta": "0:00:00.008744",
>>>>> "end": "2019-03-29 10:26:07.510481",
>>>>> "failed": false,
>>>>> "rc": 0,
>>>>> "start": "2019-03-29 10:26:07.501737",
>>>>> "stderr": "",
>>>>> "stderr_lines": [],
>>>>> "stdout": "",
>>>>> "stdout_lines": []
>>>>> }
>>>>> }
>>>>>
>>>>> according to the logs ovirtmgmt is still there.
>>>>>
>>>>> can you please share the output of 'ip a'?
>>>>>
>>>>>
>>>>>>
>>>>>> Regards,
>>>>>> Callum
>>>>>>
>>>>>> --
>>>>>>
>>>>>> Callum Smith
>>>>>> Research Computing Core
>>>>>> Wellcome Trust Centre for Human Genetics
>>>>>> University of Oxford
>>>>>> e. callum(a)well.ox.ac.uk
>>>>>>
>>>>>> On 28 Mar 2019, at 16:23, Simone Tiraboschi
<stirabos(a)redhat.com>
>>>>>> wrote:
>>>>>>
>>>>>>
>>>>>>
>>>>>> On Thu, Mar 28, 2019 at 1:44 PM Callum Smith
<callum(a)well.ox.ac.uk>
>>>>>> wrote:
>>>>>>
>>>>>> Dear Simone,
>>>>>>
>>>>>> This is my experience too, but I'm now hitting this error on
the
>>>>>> hosted-engine install at the part where it registers the
hypervisor as the
>>>>>> first host in the engine:
>>>>>>
>>>>>> 2019-03-28 12:40:50,025Z INFO
>>>>>> [org.ovirt.engine.core.bll.host.HostConnectivityChecker]
>>>>>> (EE-ManagedThreadFactory-engine-Thread-1) [49f371c1] Engine
managed to
>>>>>> communicate with VDSM
>>>>>> agent on host 'virthyp04.virt.in.bmrc.ox.ac.uk' with
address '
>>>>>> virthyp04.virt.in.bmrc.ox.ac.uk'
>>>>>> ('db571f8a-fc85-40d3-b86f-c0038e3cd7e7')
>>>>>> 2019-03-28 12:40:53,111Z WARN
>>>>>> [org.ovirt.engine.core.bll.network.NetworkConfigurator]
>>>>>> (EE-ManagedThreadFactory-engine-Thread-1) [49f371c1] Failed to
find a valid
>>>>>> interface for the
>>>>>> management network of host virthyp04.virt.in.bmrc.ox.ac.uk. If
the
>>>>>> interface ovirtmgmt is a bridge, it should be torn-down
manually.
>>>>>> 2019-03-28 12:40:53,111Z ERROR
>>>>>> [org.ovirt.engine.core.bll.hostdeploy.InstallVdsInternalCommand]
>>>>>> (EE-ManagedThreadFactory-engine-Thread-1) [49f371c1] Exception:
>>>>>> org.ovirt.engine.cor
>>>>>> e.bll.network.NetworkConfigurator$NetworkConfiguratorException:
>>>>>> Interface ovirtmgmt is invalid for management network
>>>>>>
>>>>>> The host's ovirtmgmt network connection is a statically
assigned IP
>>>>>> on a VLAN on a bond, how should I be configuring this if not
manually?
>>>>>>
>>>>>>
>>>>>> If you need to deploy over vlan 123 over bond0 simply configure
a
>>>>>> device exactly called bond0.123 and statically configure your IP
address
>>>>>> there.
>>>>>> Choose it for hosted-engine deployment, nothing more: ovirtmgmt
will
>>>>>> be automatically created over that device and the vlan ID will be
set at
>>>>>> engine level for the whole management network.
>>>>>>
>>>>>>
>>>>>>
>>>>>> Regards,
>>>>>> Callum
>>>>>>
>>>>>> --
>>>>>>
>>>>>> Callum Smith
>>>>>> Research Computing Core
>>>>>> Wellcome Trust Centre for Human Genetics
>>>>>> University of Oxford
>>>>>> e. callum(a)well.ox.ac.uk
>>>>>>
>>>>>> On 27 Mar 2019, at 17:09, Simone Tiraboschi
<stirabos(a)redhat.com>
>>>>>> wrote:
>>>>>>
>>>>>>
>>>>>>
>>>>>> On Wed, Mar 27, 2019 at 4:27 PM Callum Smith
<callum(a)well.ox.ac.uk>
>>>>>> wrote:
>>>>>>
>>>>>> It's ok, migrating to 4.3.2 on the oVirt node (from 4.3.0)
did the
>>>>>> job of fixing it.
>>>>>>
>>>>>> It is a bug if you intend on using the ovirtmgmt network to
deploy
>>>>>> your ansible from
>>>>>>
>>>>>>
>>>>>> This is a bit tricky: when the engine brings up the host it also
>>>>>> creates the management bridge and this could lead to a temporary
network
>>>>>> down on the selected interface for the bridge creation time (a
couple of
>>>>>> seconds?)
>>>>>> I tried it on a LAN and ansible ssh connection always survived
but I
>>>>>> 'm not sure it's always true.
>>>>>>
>>>>>>
>>>>>> , and you need it to have an IP address already on that range!
But -
>>>>>> it works as expected with the ovirtmgmt bridge setup so nothing
to worry
>>>>>> about.
>>>>>>
>>>>>> Regards,
>>>>>> Callum
>>>>>>
>>>>>> --
>>>>>>
>>>>>> Callum Smith
>>>>>> Research Computing Core
>>>>>> Wellcome Trust Centre for Human Genetics
>>>>>> University of Oxford
>>>>>> e. callum(a)well.ox.ac.uk
>>>>>>
>>>>>> On 27 Mar 2019, at 14:57, Simone Tiraboschi
<stirabos(a)redhat.com>
>>>>>> wrote:
>>>>>>
>>>>>>
>>>>>>
>>>>>> On Wed, Mar 27, 2019 at 3:24 PM Callum Smith
<callum(a)well.ox.ac.uk>
>>>>>> wrote:
>>>>>>
>>>>>> Dear All,
>>>>>>
>>>>>> We're trying to deploy our hosted engine remotely using the
ansible
>>>>>> hosted engine playbook, which has been a rocky road but we're
now at the
>>>>>> point where it's installing, and failing. We've got a
pre-defined bond/VLAN
>>>>>> setup for our interface which has the correct bond0 bond0.123 and
ovirtmgmt
>>>>>> bridge on top but we're hitting the classic error:
>>>>>>
>>>>>> Failed to find a valid in
>>>>>> terface for the management network of host
>>>>>> virthyp04.virt.in.bmrc.ox.ac.uk. If the interface ovirtmgmt is a
>>>>>> bridge, it should be torn-down manually.
>>>>>>
>>>>>> Does this bug still exist in the latest (4.3) version, and is
>>>>>> installing using ansible with this network configuration
impossible?
>>>>>>
>>>>>>
>>>>>> I don't think it's a bug; please avoid manually creating
ovirtmgmt
>>>>>> and simply set:
>>>>>> he_bridge_if: "bond0.123", in ansible variable file and
the
>>>>>> management bridge will be created for you at host-deploy time.
>>>>>>
>>>>>>
>>>>>>
>>>>>> Regards,
>>>>>> Callum
>>>>>>
>>>>>> --
>>>>>>
>>>>>> Callum Smith
>>>>>> Research Computing Core
>>>>>> Wellcome Trust Centre for Human Genetics
>>>>>> University of Oxford
>>>>>> e. callum(a)well.ox.ac.uk
>>>>>>
>>>>>> _______________________________________________
>>>>>> Users mailing list -- users(a)ovirt.org
>>>>>> To unsubscribe send an email to users-leave(a)ovirt.org
>>>>>> Privacy Statement:
https://www.ovirt.org/site/privacy-policy/
>>>>>> oVirt Code of Conduct:
>>>>>>
https://www.ovirt.org/community/about/community-guidelines/
>>>>>> List Archives:
>>>>>>
https://lists.ovirt.org/archives/list/users@ovirt.org/message/SBOZ6FRBRQK...
>>>>>>
>>>>>>
>>>>
>>>>
>>> _______________________________________________
>>> Users mailing list -- users(a)ovirt.org
>>> To unsubscribe send an email to users-leave(a)ovirt.org
>>> Privacy Statement:
https://www.ovirt.org/site/privacy-policy/
>>> oVirt Code of Conduct:
>>>
https://www.ovirt.org/community/about/community-guidelines/
>>> List Archives:
>>>
https://lists.ovirt.org/archives/list/users@ovirt.org/message/QOGFRHXDVOI...
>>>
>>>
>>
>