<div dir="ltr"><div><div><div>'ve been there! :-D<br><br></div>I mean exactly same issuse you had on Centos, I had on Fedora 19.<br></div>Did you disable selinux on nodes? 'cause that's what is causing SSh connection closing<br>
<br></div><div>My setup:<br><br></div><div>1 engine on vmware - fedora 19, up-to-date<br></div><div><br><div><div><br></div><div>2 nodes on IBM x series 3650 - fedora 19 based -oVirt Node - 3.0.3 - 1.1.fc19 with nodes beig in glusterfs cluster also.....<br>
<br><br></div><div>Right now, I'm banging my head against "Operation Add-Disk failed to complete." , message I have got after adding a new virtual machine and try to addd its disk<br></div></div></div></div>
<div class="gmail_extra"><br><br><div class="gmail_quote">On Fri, Jan 17, 2014 at 6:08 AM, Will Dennis (Live.com) <span dir="ltr"><<a href="mailto:willarddennis@live.com" target="_blank">willarddennis@live.com</a>></span> wrote:<br>
<blockquote class="gmail_quote" style="margin:0 0 0 .8ex;border-left:1px #ccc solid;padding-left:1ex">Hi all, ready for a story? (well, more of a rant, but hopefully it will be a<br>
good UX tale, and may even be entertaining.)<br>
<br>
Had one of the groups come to me at work this week and request a OpenStack<br>
setup. When I sat down and discussed their needs, it turns out that they<br>
really only need a multi-hypervisor setup where they can spin up VMs for<br>
their research projects. The VMs should be fairly long-lived, and will have<br>
persistent storage. Their other request is that the storage should be local<br>
on the hypervisor nodes (they plan to use Intel servers with 8-10 2TB drives<br>
for VM storage on each node.) They desire this in order to keep the VM I/O<br>
local - they do not have a SAN of any sort anyhow, and they do not care<br>
about live migration, etc.<br>
<br>
In any case, knowing that they did not want to afford a VMware setup (which<br>
is what I'm used to using), I proposed using oVirt to fill their needs,<br>
having heard and read up on it a bit (It's "open-source VMware", right?)<br>
even though I had not used it before (I have however made single-node KVM<br>
hypervisors for their group before, utilizing Open vSwitch, libvirt,<br>
virt-manager etc., so I'm not completely ignorant of KVM/libvirt etc.)<br>
<br>
In any case, I took one of their older servers which was already running<br>
CentOS 6.5, installed the requisite packages on it, and in short order had<br>
an engine server up and running (oVirt 3.3.2). That seems to have been the<br>
easy part :-/ Now came the installation of a hypervisor node. I downloaded<br>
and burned an ISO of the latest oVirt node installer<br>
(ovirt-node-iso-3.0.3-1.1.vdsm.fc19.iso) and tried to install it on one of<br>
their target Intel servers. On the 1st try I got to the end of the setup<br>
TUI, invoked the Install link, and was promptly thrown an error (sorry, but<br>
forgot what it was, something like "press X for a command prompt, or<br>
Reboot".) No problem, I rebooted, selected booting off the CD again, waited<br>
until the TUI came up, and when I tried to move past the first screen, it<br>
threw me out to a login prompt. OK, enough of that (the server takes a long<br>
time to reboot, and then boot off the CD) - I then thought I would try it on<br>
a VMware Workstation VM (yes, I get the irony, but VMware wkstn can handle<br>
nested virt, so it's a great testbed platform for OpenStack, etc.) because<br>
that would install a heck of a lot faster. That went a lot better - got the<br>
oVirt node 3.0.3 installed on the first try.<br>
<br>
More pain was soon to follow, however. I logged in and started configuring<br>
the node. The TUI was easy enough - much like an ESXi node ;) I set the NIC<br>
to IPv4 static, entered in the correct IP info, registered a DNS name for<br>
the IP I had assigned, and then tested pinging the engine, all was good. I<br>
then moved on to the section where you define the engine. I entered in the<br>
FQDN of the engine, verified the key fingerprint, and clicked the "Save and<br>
Register" link at the bottom. That seemed to work, so I completed the rest<br>
of the TUI, and then looked at the oVirt engine web UI. There was my new<br>
node, ready for authorization. I clicked the link to authorize it, and after<br>
a while, the UI came back with "Install Failed" status. Hmmm. So I went back<br>
to the node's TUI, and now some of the screens said that the IP addr was<br>
unconfigured? I went then to the Network screen, and sure enough, the NIC at<br>
the bottom showed "Unconfigured". WTF? So I went and entered in the correct<br>
info back in the IPv4 section, and then arrowed down to the Save link and<br>
clicked it - and the next screen said something like "No info needing<br>
changes, nothing to do." Whaaaa? Went back to the network setup screen, NIC<br>
still showing "Unconfigured" even though the IPv4 info still was there. I<br>
did a ping test at this point from the Ping link on the network setup page,<br>
and what do you know - I could still ping IP's (the engine, the default gw,<br>
etc.) But as I moved around the TUI, other screens still said that the<br>
network was unconfigured. Went back to the Web UI of the engine, put the<br>
host in Maint, then tried to Activate it, still no go - Install Failed. Even<br>
though I had configured the node to allow remote access and set a password,<br>
and also verified via nmap that TCP port 22 on the node was indeed<br>
listening, when I tried to SSH into the node as admin, I immediately got a<br>
"connection closed" message, so that failed as well. Went back to the node's<br>
network setup page, set the IPv4 to "Disabled", saved it, then went back and<br>
set it back to "Static" then re-entered the IPv4 info. Clicked the Save<br>
link, it went thru the setup again, came back with a success, verified with<br>
ping etc. that networking was working on the node. The engine web UI still<br>
said that it could not connect to the node however. So I put the node in<br>
Maint, and then removed it. I went back to the node, went to the Engine<br>
setup page, and re-did the screen to define the engine on the node. I notice<br>
that after I did this, however, that the node screens went back to saying<br>
that the network was unconfigured. Grrrrrr. But the node was back in the<br>
engine's Web UI, however no joy this time either - "Install failed" again.<br>
Well, the hell with this, said I - I removed the node again from the engine,<br>
and went and installed Fedora 19 minimal install on the VM, so I could use<br>
the directions found in<br>
<a href="http://www.ovirt.org/Quick_Start_Guide#Install_Fedora_Host" target="_blank">http://www.ovirt.org/Quick_Start_Guide#Install_Fedora_Host</a> and give that a<br>
try. (At least I can see what's going on with the node's OS using F19.)<br>
<br>
Installing F19 on the VM was a breeze, then logged in as root and did the<br>
"yum localinstall <a href="http://ovirt.org/releases/ovirt-release-fedora.noarch.rpm" target="_blank">http://ovirt.org/releases/ovirt-release-fedora.noarch.rpm</a>"<br>
which ran fine. I then stopped firewalld, and set it not to run at boot<br>
(which is what we typically do anyhow for internal research servers.) Then<br>
went over to the engine UI, and manually added the node. Oh happy day - the<br>
node seemed to install OK - it had the status of "Installing" for quite some<br>
time, and I looked at the processes on the F19 node, and could see python<br>
installer programs running via an SSH session from the engine. HOWEVER, at<br>
the end of the process, the Web UI reported a status of "Non Responsive",<br>
even though the F19 node looks OK (it had sanlock, supervdsmServer, vdsm<br>
processes running.) So thinking that it may take an after-install reboot, I<br>
did that, waited until the node came back up again, then clicked in the<br>
engine web UI and executed the "Confirm 'Host has been rebooted'" command,<br>
but still no good - the node remains in "Non Responsive" status.<br>
<br>
So I have no idea on how to proceed now, and what methods I can use to try<br>
and debug the connectivity problem between the engine and node. And there's<br>
many miles to go in setting up the whole environment. Maaaaaybe OpenStack<br>
would be easier ;-P No, I will press on and try to get this thing<br>
working... It seems it works for others, and looks like the right fit for<br>
the job. Just wish it was easier to get up and running.<br>
<br>
Oh yes, I tried reading the docs - go to<br>
<a href="http://www.ovirt.org/Quick_Start_Guide#Install_Hosts" target="_blank">http://www.ovirt.org/Quick_Start_Guide#Install_Hosts</a> and click on the link<br>
for "see the oVirt Node deployment documentation." Not too helpful... (Bug<br>
report has been opened.)<br>
<br>
Thanks for reading, and any clues on getting a node up and running<br>
gratefully accepted...<br>
<br>
- Will<br>
<br>
_______________________________________________<br>
Users mailing list<br>
<a href="mailto:Users@ovirt.org">Users@ovirt.org</a><br>
<a href="http://lists.ovirt.org/mailman/listinfo/users" target="_blank">http://lists.ovirt.org/mailman/listinfo/users</a><br>
</blockquote></div><br></div>