Hi Joop,

1. I did disabled firewalld/iptbles on both nodes and engines.
2. Disabling selinux: when setting to permissive via /etc/selinux/config (and of course unpersist/persist them -> /config/etc........) doesn't work ( as write this I've just reboot one node with the above modification and when I try to ssh node I get "connection closed by..." - of course selinux is enabled; after setenforce 0 ssh is OK.The only way I manage to diasble selinux was

mount -o.rw,remount /run/initramfs/live
edit grub.cfg and add selinux=0 to kernel line

3. one strange issue: after I reboot nodes, trying to ssh raise ssh warning WARNING: REMOTE HOST IDENTIFICATION HAS CHANGED! " so I have to remove host entry from .ssh/.known_host in order to connect
4. I am about to check name resolution
5.I have glusterfs also on both nodes. After I put nodes on maintenace, stop gluster volumes  and rebooting nodes caused gluster volumes to dissapear from engine web interface

Thanks




On Fri, Jan 17, 2014 at 10:35 AM, noc <noc@nieuwland.nl> wrote:
On 17-1-2014 8:53, Gabi C wrote:
've been there! :-D

I mean exactly same issuse you had on Centos, I had on Fedora 19.
Did you disable selinux on nodes? 'cause that's what is causing SSh connection closing

My setup:

1 engine on vmware  - fedora 19, up-to-date


2 nodes on IBM x series 3650  - fedora 19 based -oVirt Node - 3.0.3 - 1.1.fc19 with nodes beig in glusterfs cluster also.....


Right now, I'm banging my head against "Operation Add-Disk failed to complete." , message I have got after adding a new virtual machine and try to addd its disk


On Fri, Jan 17, 2014 at 6:08 AM, Will Dennis (Live.com) <willarddennis@live.com> wrote:
Hi all, ready for a story? (well, more of a rant, but hopefully it will be a
good UX tale, and may even be entertaining.)

Had one of the groups come to me at work this week and request a OpenStack
setup. When I sat down and discussed their needs, it turns out that they
really only need a multi-hypervisor setup where they can spin up VMs for
their research projects. The VMs should be fairly long-lived, and will have
persistent storage. Their other request is that the storage should be local
on the hypervisor nodes (they plan to use Intel servers with 8-10 2TB drives
for VM storage on each node.) They desire this in order to keep the VM I/O
local - they do not have a SAN of any sort anyhow, and they do not care
about live migration, etc
@Will
If the installation ends, either with or without error, it will give you a log location. Upload the log to a paste.bin and mail the link.

@Gabi,
There should be more info in either the vdsm.log on the SPM server or in the engine.log on the engine server, see above for lettings us know what the error is.

Having installed oVirt, probably dozens of times, I have some guidelines:
- temporarily disable firewalld/iptables (if all works, enable should still work, scripts with rules are generated and location is given)
- make selinux permissive either via setenforce 0 (until next boot) or via /etc/selinux/config ( survives reboots), dont disable it!
- make sure fqdn work in both directions between engine and host(s) (either using /etc/hosts or DNS)
- make sure NetworkManager is disabled and network enabled

Joop


_______________________________________________
Users mailing list
Users@ovirt.org
http://lists.ovirt.org/mailman/listinfo/users