Hi All!

Just did more research on this and it seems as if the reason was related to my interface configuration. Disclaimer upfront: I've a public IP configured on this server (since it's a hosted root server), but changed the IP addr here to 192.168.0.99

I started with the output from ´vdsm-tool restore-nets':
    ipv4addr, prefix = addr['address'].split('/')
ValueError: need more than 1 value to unpack

So I dumped the addr dictionary:
{'address': '192.168.0.99',
 'family': 'inet',
 'flags': frozenset(['permanent']),
 'index': 2,
 'label': 'eth0',
 'prefixlen': 32,
 'scope': 'universe'}

I've no clue why there's no "/32" at the end, but that's what my netmask actually is due to the special configuration I got from my hosting provider:

[root@vhost1 ~]# cat /etc/sysconfig/network-scripts/ifcfg-eth0
DEVICE=eth0
BOOTPROTO=none
ONBOOT=yes
HWADDR=00:52:9F:A8:AA:BB
IPADDR=192.168.0.99
NETMASK=255.255.255.255
SCOPE="peer 192.168.0.1"

(again, public IPs changed to private one, if that matters. And I skipped the IPv6 config above...)

So what I did next, was to patch the netinfo.py:
[root@vhost1 vdsm]# diff -u netinfo_orig.py netinfo.py
--- netinfo_orig.py 2014-12-24 22:16:23.362198715 +0100
+++ netinfo.py 2014-12-24 22:16:02.567625247 +0100
@@ -368,7 +368,12 @@
         if addr['family'] == 'inet':
             ipv4addrs.append(addr['address'])
             if 'secondary' not in addr['flags']:
-                ipv4addr, prefix = addr['address'].split('/')
+                """Assume /32 if no prefix was found"""
+                if addr['address'].find('/') == -1:
+                    ipv4addr = addr['address']
+                    prefix = "32"
+                else:
+                    ipv4addr, prefix = addr['address'].split('/')
                 ipv4netmask = prefix2netmask(addr['prefixlen'])
         else:
             ipv6addrs.append(addr['address'])


and recompiled it:
[root@vhost1 vdsm]# python -m py_compile netinfo.py



Et voilà: 
vdsm-tool ran fine:
`hosted-engine --deploy' passed the previous failing stage!

Hope this helps to find the root cause....

Regards, 

Andreas


On Tue, Dec 23, 2014 at 10:58 PM, Andreas Mather <andreas@allaboutapps.at> wrote:
Hi All!

Just tried to install the hosted engine on a fresh CentOS 6.6: beside setting up a gluster cluster, I just added the repo and installed hosted-engine-setup package. Otherwise it's just a very minimal installation. hosted-engine deploy failed immediately. The issue seems to be the same as described here http://lists.ovirt.org/pipermail/users/2014-October/028461.html but that conversation didn't continue after the reporting user was asked for additional details.

Failed installation:

[root@vhost1 ~]# hosted-engine --deploy
[ INFO  ] Stage: Initializing
          Continuing will configure this host for serving as hypervisor and create a VM where you have to install oVirt Engine afterwards.
          Are you sure you want to continue? (Yes, No)[Yes]:
[ INFO  ] Generating a temporary VNC password.
[ INFO  ] Stage: Environment setup
          Configuration files: []
          Log file: /var/log/ovirt-hosted-engine-setup/ovirt-hosted-engine-setup-20141223224236-y9ttk4.log
          Version: otopi-1.3.0 (otopi-1.3.0-1.el6)
[ INFO  ] Hardware supports virtualization
[ INFO  ] Stage: Environment packages setup
[ INFO  ] Stage: Programs detection
[ INFO  ] Stage: Environment setup
[ ERROR ] Failed to execute stage 'Environment setup': Command '/sbin/service' failed to execute
[ INFO  ] Stage: Clean up
[ INFO  ] Generating answer file '/etc/ovirt-hosted-engine/answers.conf'
[ INFO  ] Answer file '/etc/ovirt-hosted-engine/answers.conf' has been updated
[ INFO  ] Stage: Pre-termination
[ INFO  ] Stage: Termination

Packages:

[root@vhost1 ~]# rpm -qa|grep hosted-engine
ovirt-hosted-engine-setup-1.2.1-1.el6.noarch
ovirt-hosted-engine-ha-1.2.4-1.el6.noarch

[root@vhost1 ~]# rpm -qa|grep vdsm
vdsm-yajsonrpc-4.16.7-1.gitdb83943.el6.noarch
vdsm-python-zombiereaper-4.16.7-1.gitdb83943.el6.noarch
vdsm-python-4.16.7-1.gitdb83943.el6.noarch
vdsm-xmlrpc-4.16.7-1.gitdb83943.el6.noarch
vdsm-4.16.7-1.gitdb83943.el6.x86_64
vdsm-cli-4.16.7-1.gitdb83943.el6.noarch
vdsm-jsonrpc-4.16.7-1.gitdb83943.el6.noarch


So here's the output that was requested in the other thread. Hope someone can help me here. Thanks!

[root@vhost1 ~]#    find /var/lib/vdsm/persistence
/var/lib/vdsm/persistence
[root@vhost1 ~]#     find /var/run/vdsm/netconf
find: `/var/run/vdsm/netconf': No such file or directory
[root@vhost1 ~]# ip l 1: lo: <LOOPBACK,UP,LOWER_UP> mtu 65536 qdisc noqueue state UNKNOWN link/loopback 00:00:00:00:00:00 brd 00:00:00:00:00:00 2: eth0: <BROADCAST,MULTICAST,UP,LOWER_UP> mtu 1500 qdisc mq state UP qlen 1000 link/ether 00:25:90:d8:0a:b0 brd ff:ff:ff:ff:ff:ff 4: ;vdsmdummy;: <BROADCAST,MULTICAST> mtu 1500 qdisc noop state DOWN link/ether 62:e5:28:13:9d:ba brd ff:ff:ff:ff:ff:ff 5: bond0: <BROADCAST,MULTICAST,MASTER> mtu 1500 qdisc noop state DOWN link/ether 00:00:00:00:00:00 brd ff:ff:ff:ff:ff:ff virsh -r net 

[root@vhost1 ~]#  virsh -r net
error: unknown command: 'net'

[root@vhost1 ~]#  virsh -r net-list
Name                 State      Autostart     Persistent
--------------------------------------------------
;vdsmdummy;          active     no            no

[root@vhost1 ~]# vdsm-tool restore-nets
Traceback (most recent call last):
  File "/usr/share/vdsm/vdsm-restore-net-config", line 137, in <module>
    restore()
  File "/usr/share/vdsm/vdsm-restore-net-config", line 123, in restore
    unified_restoration()
  File "/usr/share/vdsm/vdsm-restore-net-config", line 57, in unified_restoration
    _inRollback=True)
  File "/usr/share/vdsm/network/api.py", line 616, in setupNetworks
    netinfo._libvirtNets2vdsm(libvirt_nets)))
  File "/usr/lib/python2.6/site-packages/vdsm/netinfo.py", line 822, in get
    d['nics'][dev.name] = _nicinfo(dev, paddr, ipaddrs)
  File "/usr/lib/python2.6/site-packages/vdsm/netinfo.py", line 653, in _nicinfo
    info = _devinfo(link, ipaddrs)
  File "/usr/lib/python2.6/site-packages/vdsm/netinfo.py", line 681, in _devinfo
    ipv4addr, ipv4netmask, ipv4addrs, ipv6addrs = getIpInfo(link.name, ipaddrs)
  File "/usr/lib/python2.6/site-packages/vdsm/netinfo.py", line 370, in getIpInfo
    ipv4addr, prefix = addr['address'].split('/')
ValueError: need more than 1 value to unpack
Traceback (most recent call last):
  File "/usr/bin/vdsm-tool", line 209, in main
    return tool_command[cmd]["command"](*args)
  File "/usr/lib/python2.6/site-packages/vdsm/tool/restore_nets.py", line 36, in restore_command
    restore()
  File "/usr/lib/python2.6/site-packages/vdsm/tool/restore_nets.py", line 45, in restore
    raise EnvironmentError('Failed to restore the persisted networks')
EnvironmentError: Failed to restore the persisted networks



following is mentioned in the original thread, but should help other googlers:


ovirt-hosted-engine-setup.log (first attempt):

2014-12-23 22:04:38 DEBUG otopi.plugins.otopi.services.rhel plugin.execute:861 execute-output: ('/sbin/service', 'vdsmd', 'start') stdout:
Starting multipathd daemon: [  OK  ]
Starting rpcbind: [  OK  ]
Starting ntpd: [  OK  ]
Loading the softdog kernel module: [  OK  ]
Starting wdmd: [  OK  ]
Starting sanlock: [  OK  ]
supervdsm start[  OK  ]
Starting iscsid: [  OK  ]
[  OK  ]
vdsm: Running mkdirs
vdsm: Running configure_coredump
vdsm: Running configure_vdsm_logs
vdsm: Running run_init_hooks
vdsm: Running check_is_configured
libvirt is already configured for vdsm
vdsm: Running validate_configuration
SUCCESS: ssl configured to true. No conflicts
vdsm: Running prepare_transient_repository
vdsm: Running syslog_available
vdsm: Running nwfilter
vdsm: Running dummybr
vdsm: Running load_needed_modules
vdsm: Running tune_system
vdsm: Running test_space
vdsm: Running test_lo
vdsm: Running unified_network_persistence_upgrade
vdsm: stopped during execute unified_network_persistence_upgrade task (task returned with error code 1).
vdsm start[FAILED]

2014-12-23 22:04:38 DEBUG otopi.plugins.otopi.services.rhel plugin.execute:866 execute-output: ('/sbin/service', 'vdsmd', 'start') stderr:
initctl: Job is already running: libvirtd
libvirt: Network Filter Driver error : Network filter not found: no nwfilter with matching name 'vdsm-no-mac-spoofing'

2014-12-23 22:04:38 DEBUG otopi.context context._executeMethod:152 method exception
Traceback (most recent call last):
  File "/usr/lib/python2.6/site-packages/otopi/context.py", line 142, in _executeMethod
    method['method']()
  File "/usr/share/ovirt-hosted-engine-setup/scripts/../plugins/ovirt-hosted-engine-setup/system/vdsmenv.py", line 155, in _late_setup
    state=True
  File "/usr/share/otopi/plugins/otopi/services/rhel.py", line 188, in state
    'start' if state else 'stop'
  File "/usr/share/otopi/plugins/otopi/services/rhel.py", line 96, in _executeServiceCommand
    raiseOnError=raiseOnError
  File "/usr/lib/python2.6/site-packages/otopi/plugin.py", line 871, in execute
    command=args[0],
RuntimeError: Command '/sbin/service' failed to execute
2014-12-23 22:04:38 ERROR otopi.context context._executeMethod:161 Failed to execute stage 'Environment setup': Command '/sbin/service' failed to execute


ovirt-hosted-engine-setup.log (further attempts):

2014-12-23 22:42:40 DEBUG otopi.plugins.otopi.services.rhel plugin.execute:861 execute-output: ('/sbin/service', 'vdsmd', 'start') stdout:
vdsm: Running mkdirs
vdsm: Running configure_coredump
vdsm: Running configure_vdsm_logs
vdsm: Running run_init_hooks
vdsm: Running check_is_configured
libvirt is already configured for vdsm
vdsm: Running validate_configuration
SUCCESS: ssl configured to true. No conflicts
vdsm: Running prepare_transient_repository
vdsm: Running syslog_available
vdsm: Running nwfilter
vdsm: Running dummybr
vdsm: Running load_needed_modules
vdsm: Running tune_system
vdsm: Running test_space
vdsm: Running test_lo
vdsm: Running unified_network_persistence_upgrade
vdsm: stopped during execute unified_network_persistence_upgrade task (task returned with error code 1).
vdsm start[FAILED]

2014-12-23 22:42:40 DEBUG otopi.plugins.otopi.services.rhel plugin.execute:866 execute-output: ('/sbin/service', 'vdsmd', 'start') stderr:
initctl: Job is already running: libvirtd

2014-12-23 22:42:40 DEBUG otopi.context context._executeMethod:152 method exception
Traceback (most recent call last):
  File "/usr/lib/python2.6/site-packages/otopi/context.py", line 142, in _executeMethod
    method['method']()
  File "/usr/share/ovirt-hosted-engine-setup/scripts/../plugins/ovirt-hosted-engine-setup/system/vdsmenv.py", line 155, in _late_setup
    state=True
  File "/usr/share/otopi/plugins/otopi/services/rhel.py", line 188, in state
    'start' if state else 'stop'
  File "/usr/share/otopi/plugins/otopi/services/rhel.py", line 96, in _executeServiceCommand
    raiseOnError=raiseOnError
  File "/usr/lib/python2.6/site-packages/otopi/plugin.py", line 871, in execute
    command=args[0],
RuntimeError: Command '/sbin/service' failed to execute
2014-12-23 22:42:40 ERROR otopi.context context._executeMethod:161 Failed to execute stage 'Environment setup': Command '/sbin/service' failed to execute

----

Andreas