1) I am running in a Nested env, but under libvirt/kvm on remote Centos 7.4
Hosts
Please advise how to connect with VNC to the local HE vm. I see it's
running, but this is on a remote host, not my local machine :
qemu 13376 100 3.7 17679424 845216 ? Sl 12:46 85:08
/usr/libexec/qemu-kvm -name guest=HostedEngineLocal,debug-threads=on -S
-object
secret,id=masterKey0,format=raw,file=/var/lib/libvirt/qemu/domain-1-HostedEngineLocal/master-key.aes
-machine pc-i440fx-rhel7.6.0,accel=kvm,usb=off,dump-guest-core=off -cpu
Haswell-noTSX,+kvmclock -m 16384 -realtime mlock=off -smp
4,sockets=4,cores=1,threads=1 -uuid 6fe7c1c3-ea93-4343-a385-0d9e14bb563a
-no-user-config -nodefaults -chardev
socket,id=charmonitor,fd=27,server,nowait -mon
chardev=charmonitor,id=monitor,mode=control -rtc base=utc -no-shutdown
-global PIIX4_PM.disable_s3=1 -global PIIX4_PM.disable_s4=1 -boot
menu=off,strict=on -device
virtio-serial-pci,id=virtio-serial0,bus=pci.0,addr=0x4 -drive
file=/var/tmp/localvmgmyYik/images/15023c8a-e3a7-4851-a97d-3b90996b423b/07fdcff3-11ce-4f7c-af05-0a878593e78e,format=qcow2,if=none,id=drive-virtio-disk0
-device
virtio-blk-pci,scsi=off,bus=pci.0,addr=0x5,drive=drive-virtio-disk0,id=virtio-disk0,bootindex=1
-drive
file=/var/tmp/localvmgmyYik/seed.iso,format=raw,if=none,id=drive-ide0-0-0,readonly=on
-device ide-cd,bus=ide.0,unit=0,drive=drive-ide0-0-0,id=ide0-0-0 -netdev
tap,fd=29,id=hostnet0,vhost=on,vhostfd=30 -device
virtio-net-pci,netdev=hostnet0,id=net0,mac=00:16:3e:3e:fe:28,bus=pci.0,addr=0x3
-chardev pty,id=charserial0 -device
isa-serial,chardev=charserial0,id=serial0 -chardev
socket,id=charchannel0,fd=31,server,nowait -device
virtserialport,bus=virtio-serial0.0,nr=1,chardev=charchannel0,id=channel0,name=org.qemu.guest_agent.0
*-vnc 127.0.0.1:0 <
http://127.0.0.1:0> -device
VGA*,id=video0,vgamem_mb=16,bus=pci.0,addr=0x2
-object rng-random,id=objrng0,filename=/dev/random -device
virtio-rng-pci,rng=objrng0,id=rng0,bus=pci.0,addr=0x6 -sandbox
on,obsolete=deny,elevateprivileges=deny,spawn=deny,resourcecontrol=deny
-msg timestamp=on
2) [root@vs-inf-int-kvm-fr-301-210 ~]# cat
/etc/libvirt/qemu/networks/default.xml
<!--
WARNING: THIS IS AN AUTO-GENERATED FILE. CHANGES TO IT ARE LIKELY TO BE
OVERWRITTEN AND LOST. Changes to this xml configuration should be made
using:
virsh net-edit default
or other application using the libvirt API.
-->
<network>
<name>default</name>
<uuid>ba7bbfc8-28b8-459e-a42d-c2d6218e2cb6</uuid>
<forward mode='nat'/>
<bridge name='virbr0' stp='on' delay='0'/>
<mac address='52:54:00:e5:fe:3b'/>
<ip address='192.168.122.1' netmask='255.255.255.0'>
<dhcp>
<range start='192.168.122.2' end='192.168.122.254'/>
</dhcp>
</ip>
</network>
You have new mail in /var/spool/mail/root
[root@vs-inf-int-kvm-fr-301-210 ~]
Guillaume Pavese
Ingénieur Système et Réseau
Interactiv-Group
On Mon, Feb 25, 2019 at 9:57 PM Simone Tiraboschi <stirabos(a)redhat.com>
wrote:
On Mon, Feb 25, 2019 at 1:14 PM Guillaume Pavese <
guillaume.pavese(a)interactiv-group.com> wrote:
> He deployment with "hosted-engine --deploy" fails at TASK
> [ovirt.hosted_engine_setup : Get local VM IP]
>
> See following Error :
>
> 2019-02-25 12:46:50,154+0100 INFO
> otopi.ovirt_hosted_engine_setup.ansible_utils
> ansible_utils._process_output:109 TASK [ovirt.hosted_engine_setup : Get
> local VM IP]
> 2019-02-25 12:55:26,823+0100 DEBUG
> otopi.ovirt_hosted_engine_setup.ansible_utils
> ansible_utils._process_output:103 {u'_ansible_parsed': True,
> u'stderr_lines': [], u'cmd': u"virsh -r net-dhcp-leases default
| grep -i 00
> :16:3e:3e:fe:28 | awk '{ print $5 }' | cut -f1 -d'/'",
u'end':
> u'2019-02-25 12:55:26.666925', u'_ansible_no_log': False,
u'stdout': u'',
> u'changed': True, u'invocation': {u'module_args':
{u'warn': True,
> u'executable':
> None, u'_uses_shell': True, u'_raw_params': u"virsh -r
net-dhcp-leases
> default | grep -i 00:16:3e:3e:fe:28 | awk '{ print $5 }' | cut -f1
-d'/'",
> u'removes': None, u'argv': None, u'creates': None,
u'chdir': None, u'std
> in': None}}, u'start': u'2019-02-25 12:55:26.584686',
u'attempts': 50,
> u'stderr': u'', u'rc': 0, u'delta':
u'0:00:00.082239', u'stdout_lines': []}
> 2019-02-25 12:55:26,924+0100 ERROR
> otopi.ovirt_hosted_engine_setup.ansible_utils
> ansible_utils._process_output:107 fatal: [localhost]: FAILED! =>
> {"attempts": 50, "changed": true, "cmd": "virsh -r
net-dhcp-leases default
> | grep -i 00:16:3e:3e:fe:28 | awk '{ print $5 }' | cut -f1
-d'/'", "delta":
> "0:00:00.082239", "end": "2019-02-25 12:55:26.666925",
"rc": 0, "start":
> "2019-02-25 12:55:26.584686", "stderr": "",
"stderr_lines": [], "stdout":
> "", "stdout_lines": []}
>
Here we are just waiting for the bootstrap engine VM to fetch an IP
address from default libvirt network over DHCP but it your case it never
happened.
Possible issues: something went wrong in the bootstrap process for the
engine VM or the default libvirt network is not correctly configured.
1. can you try to reach the engine VM via VNC and check what's happening
there? (another question, are you running it nested? AFAIK it will not work
if nested over ESXi)
2. can you please share the output of
cat /etc/libvirt/qemu/networks/default.xml
>
> Guillaume Pavese
> Ingénieur Système et Réseau
> Interactiv-Group
> _______________________________________________
> Users mailing list -- users(a)ovirt.org
> To unsubscribe send an email to users-leave(a)ovirt.org
> Privacy Statement:
https://www.ovirt.org/site/privacy-policy/
> oVirt Code of Conduct:
>
https://www.ovirt.org/community/about/community-guidelines/
> List Archives:
>
https://lists.ovirt.org/archives/list/users@ovirt.org/message/VXRMU3SQWTM...
>