Thank you all for your help, I have managed to deploy the engine
successfully. It was a quiet a lesson.
On Wed, Jul 11, 2018 at 11:55 AM, Sakhi Hadebe <sakhi(a)sanren.ac.za> wrote:
Hi Sahina,
Yes the glusterd daemon was not running. I have started it and is able to
add a glusterfs storage domain. Thank you so much for your help.
Oops! I allocated 50GiB for this storage domain and it requires 60GiB.
On Wed, Jul 11, 2018 at 11:47 AM, Sahina Bose <sabose(a)redhat.com> wrote:
> Is glusterd running on the server: goku.sanren.**
> There's an error
> Failed to get volume info: Command execution failed
> error: Connection failed. Please check if gluster daemon is operational
>
> Please check the volume status using "gluster volume status engine"
>
> and if all looks ok, attach the mount logs from /var/log/glusterfs
>
> On Wed, Jul 11, 2018 at 1:57 PM, Sakhi Hadebe <sakhi(a)sanren.ac.za> wrote:
>
>> Hi,
>>
>> I have managed to fix the error by enabling the DMA Virtualisation in
>> BIOS. I am now hit with a new error: It's failing to add a glusterfs
>> storage domain:
>>
>> [ INFO ] TASK [Add glusterfs storage domain]
>> [ ERROR ] Error: Fault reason is "Operation Failed". Fault detail is
>> "[Problem while trying to mount target]". HTTP response code is 400.
>> [ ERROR ] fatal: [localhost]: FAILED! => {"changed": false,
"msg":
>> "Fault reason is \"Operation Failed\". Fault detail is
\"[Problem while
>> trying to mount target]\". HTTP response code is 400."}
>> Please specify the storage you would like to use (glusterfs,
>> iscsi, fc, nfs)[nfs]:
>>
>> Attached are vdsm and engine log files.
>>
>>
>>
>>
>>
>> On Wed, Jul 11, 2018 at 9:57 AM, Sakhi Hadebe <sakhi(a)sanren.ac.za>
>> wrote:
>>
>>>
>>>
>>> On Wed, Jul 11, 2018 at 9:33 AM, Sakhi Hadebe <sakhi(a)sanren.ac.za>
>>> wrote:
>>>
>>>> Hi,
>>>>
>>>> Below are the versions of packages installed. Please find the logs
>>>> attached.
>>>> Qemu:
>>>> ipxe-roms-qemu-20170123-1.git4e85b27.el7_4.1.noarch
>>>> libvirt-daemon-driver-qemu-3.9.0-14.el7_5.6.x86_64
>>>> qemu-img-ev-2.10.0-21.el7_5.4.1.x86_64
>>>> qemu-kvm-ev-2.10.0-21.el7_5.4.1.x86_64
>>>> qemu-kvm-common-ev-2.10.0-21.el7_5.4.1.x86_64
>>>>
>>>> Libvirt installed packages:
>>>> libvirt-daemon-driver-storage-disk-3.9.0-14.el7_5.6.x86_64
>>>> libvirt-daemon-config-nwfilter-3.9.0-14.el7_5.6.x86_64
>>>> libvirt-daemon-driver-storage-iscsi-3.9.0-14.el7_5.6.x86_64
>>>> libvirt-daemon-driver-network-3.9.0-14.el7_5.6.x86_64
>>>> libvirt-libs-3.9.0-14.el7_5.6.x86_64
>>>> libvirt-daemon-driver-secret-3.9.0-14.el7_5.6.x86_64
>>>> libvirt-daemon-driver-storage-core-3.9.0-14.el7_5.6.x86_64
>>>> libvirt-daemon-driver-storage-gluster-3.9.0-14.el7_5.6.x86_64
>>>> libvirt-daemon-driver-storage-3.9.0-14.el7_5.6.x86_64
>>>> libvirt-daemon-driver-qemu-3.9.0-14.el7_5.6.x86_64
>>>> libvirt-3.9.0-14.el7_5.6.x86_64
>>>> libvirt-python-3.9.0-1.el7.x86_64
>>>> libvirt-daemon-driver-nodedev-3.9.0-14.el7_5.6.x86_64
>>>> libvirt-daemon-driver-storage-rbd-3.9.0-14.el7_5.6.x86_64
>>>> libvirt-daemon-driver-storage-scsi-3.9.0-14.el7_5.6.x86_64
>>>> libvirt-daemon-config-network-3.9.0-14.el7_5.6.x86_64
>>>> libvirt-client-3.9.0-14.el7_5.6.x86_64
>>>> libvirt-daemon-kvm-3.9.0-14.el7_5.6.x86_64
>>>> libvirt-daemon-driver-storage-logical-3.9.0-14.el7_5.6.x86_64
>>>> libvirt-daemon-3.9.0-14.el7_5.6.x86_64
>>>> libvirt-daemon-driver-interface-3.9.0-14.el7_5.6.x86_64
>>>> libvirt-lock-sanlock-3.9.0-14.el7_5.6.x86_64
>>>> libvirt-daemon-driver-storage-mpath-3.9.0-14.el7_5.6.x86_64
>>>> libvirt-daemon-driver-lxc-3.9.0-14.el7_5.6.x86_64
>>>> libvirt-daemon-driver-nwfilter-3.9.0-14.el7_5.6.x86_64
>>>>
>>>> Virt-manager:
>>>> virt-manager-common-1.4.3-3.el7.noarch
>>>>
>>>> oVirt:
>>>> [root@localhost network-scripts]# rpm -qa | grep ovirt
>>>> ovirt-setup-lib-1.1.4-1.el7.centos.noarch
>>>> cockpit-ovirt-dashboard-0.11.28-1.el7.noarch
>>>> ovirt-imageio-common-1.3.1.2-0.el7.centos.noarch
>>>> ovirt-vmconsole-host-1.0.5-4.el7.centos.noarch
>>>> ovirt-host-dependencies-4.2.3-1.el7.x86_64
>>>> ovirt-engine-sdk-python-3.6.9.1-1.el7.noarch
>>>> ovirt-imageio-daemon-1.3.1.2-0.el7.centos.noarch
>>>> ovirt-host-4.2.3-1.el7.x86_64
>>>> python-ovirt-engine-sdk4-4.2.7-2.el7.x86_64
>>>> ovirt-host-deploy-1.7.4-1.el7.noarch
>>>> cockpit-machines-ovirt-169-1.el7.noarch
>>>> ovirt-hosted-engine-ha-2.2.14-1.el7.noarch
>>>> ovirt-vmconsole-1.0.5-4.el7.centos.noarch
>>>> ovirt-provider-ovn-driver-1.2.11-1.el7.noarch
>>>> ovirt-engine-appliance-4.2-20180626.1.el7.noarch
>>>> ovirt-release42-4.2.4-1.el7.noarch
>>>> ovirt-hosted-engine-setup-2.2.22.1-1.el7.noarch
>>>>
>>>>
>>>>
>>>>
>>>>
>>>>
>>>> On Wed, Jul 11, 2018 at 6:48 AM, Yedidyah Bar David
<didi(a)redhat.com>
>>>> wrote:
>>>>
>>>>> On Tue, Jul 10, 2018 at 11:32 PM, Sakhi Hadebe
<sakhi(a)sanren.ac.za>
>>>>> wrote:
>>>>>
>>>>>> Hi,
>>>>>>
>>>>>> I did not select any CPU architecture. It doenst gove me the
option
>>>>>> to select one. It only states the number of virtual CPUs and the
memory for
>>>>>> the engine VM.
>>>>>>
>>>>>> Looking at the documentation of installing
ovirt-release36.rpm....it
>>>>>> does allow you to select te CPU, but not when installing
ovirt-release42.rpm
>>>>>>
>>>>>> On Tuesday, July 10, 2018, Alastair Neil
<ajneil.tech(a)gmail.com>
>>>>>> wrote:
>>>>>>
>>>>>>> what did you select as your CPU architecture when you created
the
>>>>>>> cluster? It looks like the VM is trying to use a CPU type of
"Custom", how
>>>>>>> many nodes in your cluster? I suggest you specify the lowest
common
>>>>>>> denominator of CPU architecture (e.g. Sandybridge) of the
nodes as the CPU
>>>>>>> architecture of the cluster..
>>>>>>>
>>>>>>> On Tue, 10 Jul 2018 at 12:01, Sakhi Hadebe
<sakhi(a)sanren.ac.za>
>>>>>>> wrote:
>>>>>>>
>>>>>>>> Hi,
>>>>>>>>
>>>>>>>> I have just re-installed centOS 7 in 3 servers and have
configured
>>>>>>>> gluster volumes following this documentation:
>>>>>>>>
https://www.ovirt.org/blog/2016/03/up-and-running-with-ovirt-3-6/,
>>>>>>>> But I have installed
>>>>>>>>
>>>>>>>>
http://resources.ovirt.org/pub/yum-repo/ovirt-release42.rpm
>>>>>>>>
>>>>>>>> package.
>>>>>>>> Hosted-engine --deploy is failing with this error:
>>>>>>>>
>>>>>>>> "rhel7", "--virt-type",
"kvm", "--memory", "16384", "--vcpus",
>>>>>>>> "4", "--network",
"network=default,mac=00:16:3e:09:5e:5d,model=virtio",
>>>>>>>> "--disk", "/var/tmp/localvm0nnJH9/images
>>>>>>>>
/eacac30d-0304-4c77-8753-6965e4b8c2e7/d494577e-027a-4209-895b-6132e6fc6b9a",
>>>>>>>> "--import", "--disk",
"path=/var/tmp/localvm0nnJH9/seed.iso,device=cdrom",
>>>>>>>> "--noautoconsole", "--rng",
"/dev/random", "--graphics", "vnc", "--video",
>>>>>>>> "vga", "--sound", "none",
"--controller", "usb,model=none", "--memballoon",
>>>>>>>> "none", "--boot",
"hd,menu=off", "--clock", "kvmclock_present=yes"],
>>>>>>>> "delta": "0:00:00.979003",
"end": "2018-07-10 17:55:11.308555", "msg":
>>>>>>>> "non-zero return code", "rc": 1,
"start": "2018-07-10 17:55:10.329552",
>>>>>>>> "stderr": "ERROR unsupported
configuration: CPU mode 'custom' for x86_64
>>>>>>>> kvm domain on x86_64 host is not supported by
hypervisor\nDomain
>>>>>>>> installation does not appear to have been successful.\nIf
it was, you can
>>>>>>>> restart your domain by running:\n virsh --connect
qemu:///system start
>>>>>>>> HostedEngineLocal\notherwise, please restart your
installation.",
>>>>>>>> "stderr_lines": ["ERROR unsupported
configuration: CPU mode 'custom' for
>>>>>>>> x86_64 kvm domain on x86_64 host is not supported by
hypervisor", "Domain
>>>>>>>> installation does not appear to have been
successful.", "If it was, you can
>>>>>>>> restart your domain by running:", " virsh
--connect qemu:///system start
>>>>>>>> HostedEngineLocal", "otherwise, please restart
your installation."],
>>>>>>>> "stdout": "\nStarting install...",
"stdout_lines": ["", "Starting
>>>>>>>> install..."]}
>>>>>>>>
>>>>>>>
>>>>> This seems to be in the phase where we create a local vm for the
>>>>> engine. We do this with plain virt-install, nothing fancy. Searching
the
>>>>> net for "unsupported configuration: CPU mode
'custom'" finds other relevant
>>>>> reports, you might want to check them. You can see the command in
>>>>> bootstrap_local_vm.yml .
>>>>>
>>>>> Please check/share versions of relevant packages (libvirt*, qemu*,
>>>>> etc) and relevant logs (libvirt).
>>>>>
>>>>> Also updating the subject line and adding Simone.
>>>>>
>>>>> Best regards,
>>>>> --
>>>>> Didi
>>>>>
>>>>
>>>>
>>>>
>>>> --
>>>> Regards,
>>>> Sakhi Hadebe
>>>>
>>>> Engineer: South African National Research Network (SANReN)Competency
Area, Meraka, CSIR
>>>>
>>>> Tel: +27 12 841 2308 <+27128414213>
>>>> Fax: +27 12 841 4223 <+27128414223>
>>>> Cell: +27 71 331 9622 <+27823034657>
>>>> Email: sakhi(a)sanren.ac.za <shadebe(a)csir.co.za>
>>>>
>>>>
>>>
>>>
>>> --
>>> Regards,
>>> Sakhi Hadebe
>>>
>>> Engineer: South African National Research Network (SANReN)Competency Area,
Meraka, CSIR
>>>
>>> Tel: +27 12 841 2308 <+27128414213>
>>> Fax: +27 12 841 4223 <+27128414223>
>>> Cell: +27 71 331 9622 <+27823034657>
>>> Email: sakhi(a)sanren.ac.za <shadebe(a)csir.co.za>
>>>
>>>
>>
>>
>> --
>> Regards,
>> Sakhi Hadebe
>>
>> Engineer: South African National Research Network (SANReN)Competency Area,
Meraka, CSIR
>>
>> Tel: +27 12 841 2308 <+27128414213>
>> Fax: +27 12 841 4223 <+27128414223>
>> Cell: +27 71 331 9622 <+27823034657>
>> Email: sakhi(a)sanren.ac.za <shadebe(a)csir.co.za>
>>
>>
>> _______________________________________________
>> Users mailing list -- users(a)ovirt.org
>> To unsubscribe send an email to users-leave(a)ovirt.org
>> Privacy Statement:
https://www.ovirt.org/site/privacy-policy/
>> oVirt Code of Conduct:
https://www.ovirt.org/communit
>> y/about/community-guidelines/
>> List Archives:
https://lists.ovirt.org/archiv
>> es/list/users(a)ovirt.org/message/YHKUKW22QLRVS56XZBXEWOGORFWFEGIA/
>>
>>
>
--
Regards,
Sakhi Hadebe
Engineer: South African National Research Network (SANReN)Competency Area, Meraka, CSIR
Tel: +27 12 841 2308 <+27128414213>
Fax: +27 12 841 4223 <+27128414223>
Cell: +27 71 331 9622 <+27823034657>
Email: sakhi(a)sanren.ac.za <shadebe(a)csir.co.za>
--
Regards,
Sakhi Hadebe
Engineer: South African National Research Network (SANReN)Competency
Area, Meraka, CSIR
Tel: +27 12 841 2308 <+27128414213>
Fax: +27 12 841 4223 <+27128414223>
Cell: +27 71 331 9622 <+27823034657>
Email: sakhi(a)sanren.ac.za <shadebe(a)csir.co.za>