[ovirt-users] Installation of oVirt 4.1, Gluster Storage and Hosted Engine
knarra
knarra at redhat.com
Thu Jul 13 12:23:40 UTC 2017
On 07/13/2017 04:30 PM, Simone Marchioni wrote:
> Il 12/07/2017 10:59, knarra ha scritto:
>> On 07/12/2017 01:43 PM, Simone Marchioni wrote:
>>> Il 11/07/2017 11:23, knarra ha scritto:
>>>
>>> Hi,
>>>
>>> reply here to both Gianluca and Kasturi.
>>>
>>> Gianluca: I had ovirt-4.1-dependencies.repo enabled, and gluster 3.8
>>> packages, but glusterfs-server was missing in my "yum install"
>>> command, so added glusterfs-server to my installation.
>>>
>>> Kasturi: packages ovirt-hosted-engine-setup, gdeploy and
>>> cockpit-ovirt-dashboard already installed and updated. vdsm-gluster
>>> was missing, so added to my installation.
>> okay, cool.
>
> :-)
>
>>>
>>> Rerun deployment and IT WORKED! I can read the message "Succesfully
>>> deployed Gluster" with the blue button "Continue to Hosted Engine
>>> Deployment". There's a minor glitch in the window: the green "V" in
>>> the circle is missing, like there's a missing image (or a wrong
>>> path, as I had to remove "ansible" from the grafton-sanity-check.sh
>>> path...)
>> There is a bug for this and it will be fixed soon. Here is the bug id
>> for your reference. https://bugzilla.redhat.com/show_bug.cgi?id=1462082
>
> Ok, thank you!
>
>>>
>>> Although the deployment worked, and the firewalld and gluterfs
>>> errors are gone, a couple of errors remains:
>>>
>>>
>>> AFTER VG/LV CREATION, START/STOP/RELOAD/GLUSTER AND FIREWALLD HANDLING:
>>>
>>> PLAY [gluster_servers]
>>> *********************************************************
>>>
>>> TASK [Run a shell script]
>>> ******************************************************
>>> fatal: [ha1.domain.it]: FAILED! => {"failed": true, "msg": "The
>>> conditional check 'result.rc != 0' failed. The error was: error
>>> while evaluating conditional (result.rc != 0): 'dict object' has no
>>> attribute 'rc'"}
>>> fatal: [ha2.domain.it]: FAILED! => {"failed": true, "msg": "The
>>> conditional check 'result.rc != 0' failed. The error was: error
>>> while evaluating conditional (result.rc != 0): 'dict object' has no
>>> attribute 'rc'"}
>>> fatal: [ha3.domain.it]: FAILED! => {"failed": true, "msg": "The
>>> conditional check 'result.rc != 0' failed. The error was: error
>>> while evaluating conditional (result.rc != 0): 'dict object' has no
>>> attribute 'rc'"}
>>> to retry, use: --limit @/tmp/tmpJnz4g3/run-script.retry
>> May be you missed to change the path of the script
>> "/usr/share/ansible/gdeploy/scripts/disable-gluster-hooks.sh" . That
>> is why this failure.
>
> You're right: changed the path and now it's ok.
>
>>>
>>> PLAY RECAP
>>> *********************************************************************
>>> ha1.domain.it : ok=0 changed=0 unreachable=0 failed=1
>>> ha2.domain.it : ok=0 changed=0 unreachable=0 failed=1
>>> ha3.domain.it : ok=0 changed=0 unreachable=0 failed=1
>>>
>>>
>>> PLAY [gluster_servers]
>>> *********************************************************
>>>
>>> TASK [Run a command in the shell]
>>> **********************************************
>>> failed: [ha1.domain.it] (item=usermod -a -G gluster qemu) =>
>>> {"changed": true, "cmd": "usermod -a -G gluster qemu", "delta":
>>> "0:00:00.003144", "end": "2017-07-12 00:22:46.836832", "failed":
>>> true, "item": "usermod -a -G gluster qemu", "rc": 6, "start":
>>> "2017-07-12 00:22:46.833688", "stderr": "usermod: group 'gluster'
>>> does not exist", "stderr_lines": ["usermod: group 'gluster' does not
>>> exist"], "stdout": "", "stdout_lines": []}
>>> failed: [ha2.domain.it] (item=usermod -a -G gluster qemu) =>
>>> {"changed": true, "cmd": "usermod -a -G gluster qemu", "delta":
>>> "0:00:00.003647", "end": "2017-07-12 00:22:46.895964", "failed":
>>> true, "item": "usermod -a -G gluster qemu", "rc": 6, "start":
>>> "2017-07-12 00:22:46.892317", "stderr": "usermod: group 'gluster'
>>> does not exist", "stderr_lines": ["usermod: group 'gluster' does not
>>> exist"], "stdout": "", "stdout_lines": []}
>>> failed: [ha3.domain.it] (item=usermod -a -G gluster qemu) =>
>>> {"changed": true, "cmd": "usermod -a -G gluster qemu", "delta":
>>> "0:00:00.007008", "end": "2017-07-12 00:22:47.016600", "failed":
>>> true, "item": "usermod -a -G gluster qemu", "rc": 6, "start":
>>> "2017-07-12 00:22:47.009592", "stderr": "usermod: group 'gluster'
>>> does not exist", "stderr_lines": ["usermod: group 'gluster' does not
>>> exist"], "stdout": "", "stdout_lines": []}
>>> to retry, use: --limit @/tmp/tmpJnz4g3/shell_cmd.retry
>>>
>>> PLAY RECAP
>>> *********************************************************************
>>> ha1.domain.it : ok=0 changed=0 unreachable=0 failed=1
>>> ha2.domain.it : ok=0 changed=0 unreachable=0 failed=1
>>> ha3.domain.it : ok=0 changed=0 unreachable=0 failed=1
>> This error can be safely ignored.
>
> Ok
>
>>>
>>>
>>> These are a problem for my installation or can I ignore them?
>> You can just manually run the script to disable hooks on all the
>> nodes. Other error you can ignore.
>
> Done it
>
>>>
>>> By the way, I'm writing and documenting this process and can prepare
>>> a tutorial if someone is interested.
>>>
>>> Thank you again for your support: now I'll proceed with the Hosted
>>> Engine Deployment.
>> Good to know that you can now start with Hosted Engine Deployment.
>
> Started the Hosted Engine Deployment, but I have a different problem now.
>
> As the installer asked, I specified some parameters, in particular a
> pingable gateway address. Specified the host1 gateway.
> Proceeding with the installer, it requires The Engine VM IP address
> (DHCP or Static). I selected static and specified an IP Address, but
> the IP *IS NOT* in the same subnet as the host1. The VMs IP addresses
> are all on a different subnet.
> The installer shows a red message:
>
> The Engine VM (aa.bb.cc.dd/SM) and the default gateway (ww.xx.yy.zz)
> will not be in the same IP subnet. Static routing configuration are
> not supported on automatic VM configuration.
>
> I'm starting to think that BOTH the hosts IPs and the VM IPs MUST BE
> ON THE SAME SUBNET.
>
> Is this a requirement or there's a way to deal with this configuration?
> It's related only to "automatic VM configuration" or to oVirt in
> general? Once installed oVirt Engine can I have VMs on different subnet?
Here are some Requirements :
1) Engine VM and the managed hosts should be on management subnet
2) Managed host should be able to resolve EngineVM hostname and
vice-versa. We configure the engine appliance with DHCP or static
addressing.
3) Before proceeding with hosted-engine deployment we recommend the user
to have an FQDN assigned to a mac address to go via DHCP way.
4) If you are using static then IP address/ netmask /gateway based on
host ones , and add entries to /etc/hosts if you do not have local DNS.
Hope this helps !!!
Thanks
kasturi
>
> Hi
> Simone
> _______________________________________________
> Users mailing list
> Users at ovirt.org
> http://lists.ovirt.org/mailman/listinfo/users
More information about the Users
mailing list