Hi Joop,
1. I did disabled firewalld/iptbles on both nodes and engines.
2. Disabling selinux: when setting to permissive via /etc/selinux/config
(and of course unpersist/persist them -> /config/etc........) doesn't work
( as write this I've just reboot one node with the above modification and
when I try to ssh node I get "connection closed by..." - of course selinux
is enabled; after setenforce 0 ssh is OK.The only way I manage to diasble
selinux was
mount -o.rw,remount /run/initramfs/live
edit grub.cfg and add selinux=0 to kernel line
3. one strange issue: after I reboot nodes, trying to ssh raise ssh warning
WARNING: REMOTE HOST IDENTIFICATION HAS CHANGED! " so I have to remove host
entry from .ssh/.known_host in order to connect
4. I am about to check name resolution
5.I have glusterfs also on both nodes. After I put nodes on maintenace,
stop gluster volumes and rebooting nodes caused gluster volumes to
dissapear from engine web interface
Thanks
On Fri, Jan 17, 2014 at 10:35 AM, noc <noc(a)nieuwland.nl> wrote:
On 17-1-2014 8:53, Gabi C wrote:
've been there! :-D
I mean exactly same issuse you had on Centos, I had on Fedora 19.
Did you disable selinux on nodes? 'cause that's what is causing SSh
connection closing
My setup:
1 engine on vmware - fedora 19, up-to-date
2 nodes on IBM x series 3650 - fedora 19 based -oVirt Node - 3.0.3 -
1.1.fc19 with nodes beig in glusterfs cluster also.....
Right now, I'm banging my head against "Operation Add-Disk failed to
complete." , message I have got after adding a new virtual machine and try
to addd its disk
On Fri, Jan 17, 2014 at 6:08 AM, Will Dennis (
Live.com) <
willarddennis(a)live.com> wrote:
> Hi all, ready for a story? (well, more of a rant, but hopefully it will
> be a
> good UX tale, and may even be entertaining.)
>
> Had one of the groups come to me at work this week and request a OpenStack
> setup. When I sat down and discussed their needs, it turns out that they
> really only need a multi-hypervisor setup where they can spin up VMs for
> their research projects. The VMs should be fairly long-lived, and will
> have
> persistent storage. Their other request is that the storage should be
> local
> on the hypervisor nodes (they plan to use Intel servers with 8-10 2TB
> drives
> for VM storage on each node.) They desire this in order to keep the VM I/O
> local - they do not have a SAN of any sort anyhow, and they do not care
> about live migration, etc
>
@Will
If the installation ends, either with or without error, it will give you a
log location. Upload the log to a paste.bin and mail the link.
@Gabi,
There should be more info in either the vdsm.log on the SPM server or in
the engine.log on the engine server, see above for lettings us know what
the error is.
Having installed oVirt, probably dozens of times, I have some guidelines:
- temporarily disable firewalld/iptables (if all works, enable should
still work, scripts with rules are generated and location is given)
- make selinux permissive either via setenforce 0 (until next boot) or via
/etc/selinux/config ( survives reboots), dont disable it!
- make sure fqdn work in both directions between engine and host(s)
(either using /etc/hosts or DNS)
- make sure NetworkManager is disabled and network enabled
Joop
_______________________________________________
Users mailing list
Users(a)ovirt.org
http://lists.ovirt.org/mailman/listinfo/users