Re: Node not starting | blk_cloned_rq_check_limits: over max size limit
by Strahil
You can check https://access.redhat.com/solutions/2437991 & https://access.redhat.com/solutions/3014361
You have 2 options:
1. Set a udev rule like this one (replace NETAPP with your storage)
ACTION!="add|change", GOTO="rule_end" ENV{ID_VENDOR}=="NETAPP*", RUN+="/bin/sh -c 'echo 4096 > /sys%p/queue/max_sectors_kb'" LABEL="rule_end"
2. Set max_sectors_kb in devices section of multipath.conf
You will need to stop LVM and then flush the device map , so the new option to take effect (faster is to reboot)
Good Luck & Happy New Year.
Best Regards,
Strahil Nikolov
On Dec 31, 2019 17:53, Stefan Wolf <shb256(a)gmail.com> wrote:
>
> hi all,
>
> i ve 4 nodes running with current ovirt.
> I ve only a problem on one host even after a fresh installation .
> I ve installed the latest image.
> Than I add the node to the cluster
> Everything is working good.
> After this I configure the network.
> BUT, after a restart the host does not come up again.
> I got this error: blk_cloned_rq_check_limits: over max size limit
> every 5 seconds
>
> I can continue with control-D
> or I can login with root password to fix the problem. but i dont know what is the problem and where does it came from
>
> I ve also changed the sas disk to nvme storage, but I ve changed this on every host. And this problem exists only on one host
>
> i found this https://lists.centos.org/pipermail/centos/2017-December/167727.html
> the output is
> [root@kvm380 ~]# ./test.sh
> Sys Block Node : Device max_sectors_kb max_hw_sectors_kb
> /sys/block/dm-0 : onn_kvm380-pool00_tmeta 256 4096
> /sys/block/dm-1 : onn_kvm380-pool00_tdata 256 4096
> /sys/block/dm-10 : onn_kvm380-var 256 4096
> /sys/block/dm-11 : onn_kvm380-tmp 256 4096
> /sys/block/dm-12 : onn_kvm380-home 256 4096
> /sys/block/dm-13 : onn_kvm380-var_crash 256 4096
> /sys/block/dm-2 : onn_kvm380-pool00-tpool 256 4096
> /sys/block/dm-3 : onn_kvm380-ovirt--node--ng--4.3.7--0.20191121.0+1 256 4096
> /sys/block/dm-4 : onn_kvm380-swap 256 4096
> /sys/block/dm-5 : eui.0025385991b1e27a 512 2048
> /sys/block/dm-6 : eui.0025385991b1e27a1 512 2048
> /sys/block/dm-7 : onn_kvm380-pool00 256 4096
> /sys/block/dm-8 : onn_kvm380-var_log_audit 256 4096
> /sys/block/dm-9 : onn_kvm380-var_log 256 4096
> cat: /sys/block/nvme0n1/device/vendor: Datei oder Verzeichnis nicht gefunden
> /sys/block/nvme0n1: Samsung SSD 970 EVO 1TB 512 2048
> /sys/block/sda : HP LOGICAL VOLUME 256 4096
>
> is the nvme not starting correct
> [root@kvm380 ~]# systemctl status multipathd
> ● multipathd.service - Device-Mapper Multipath Device Controller
> Loaded: loaded (/usr/lib/systemd/system/multipathd.service; enabled; vendor preset: enabled)
> Active: active (running) since Di 2019-12-31 16:16:32 CET; 31min ago
> Process: 1919 ExecStart=/sbin/multipathd (code=exited, status=0/SUCCESS)
> Process: 1916 ExecStartPre=/sbin/multipath -A (code=exited, status=0/SUCCESS)
> Process: 1911 ExecStartPre=/sbin/modprobe dm-multipath (code=exited, status=0/SUCCESS)
> Main PID: 1921 (multipathd)
> Tasks: 7
> CGroup: /system.slice/multipathd.service
> └─1921 /sbin/multipathd
>
> Dez 31 16:47:58 kvm380.durchhalten.intern multipathd[1921]: nvme0n1: mark as failed
> Dez 31 16:47:58 kvm380.durchhalten.intern multipathd[1921]: eui.0025385991b1e27a: Entering recovery mode: max_retries=4
> Dez 31 16:47:58 kvm380.durchhalten.intern multipathd[1921]: eui.0025385991b1e27a: remaining active paths: 0
> Dez 31 16:48:02 kvm380.durchhalten.intern multipathd[1921]: 259:0: reinstated
> Dez 31 16:48:02 kvm380.durchhalten.intern multipathd[1921]: eui.0025385991b1e27a: queue_if_no_path enabled
> Dez 31 16:48:02 kvm380.durchhalten.intern multipathd[1921]: eui.0025385991b1e27a: Recovered to normal mode
> Dez 31 16:48:02 kvm380.durchhalten.intern multipathd[1921]: eui.0025385991b1e27a: remaining active paths: 1
> Dez 31 16:48:03 kvm380.durchhalten.intern multipathd[1921]: nvme0n1: mark as failed
> Dez 31 16:48:03 kvm380.durchhalten.intern multipathd[1921]: eui.0025385991b1e27a: Entering recovery mode: max_retries=4
> Dez 31 16:48:03 kvm380.durchhalten.intern multipathd[1921]: eui.0025385991b1e27a: remaining active paths: 0
>
> why is it marked as failed?
>
> if i create a new volume with cockpit and use it for bricks for gluster, every thing is fine. until reboot
>
>
> maybe some one can point me the direction
> _______________________________________________
> Users mailing list -- users(a)ovirt.org
> To unsubscribe send an email to users-leave(a)ovirt.org
> Privacy Statement: https://www.ovirt.org/site/privacy-policy/
> oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/
> List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/MHHFFWAY5T5...
4 years, 12 months
Node not starting | blk_cloned_rq_check_limits: over max size limit
by Stefan Wolf
hi all,
i ve 4 nodes running with current ovirt.
I ve only a problem on one host even after a fresh installation .
I ve installed the latest image.
Than I add the node to the cluster
Everything is working good.
After this I configure the network.
BUT, after a restart the host does not come up again.
I got this error: blk_cloned_rq_check_limits: over max size limit
every 5 seconds
I can continue with control-D
or I can login with root password to fix the problem. but i dont know what is the problem and where does it came from
I ve also changed the sas disk to nvme storage, but I ve changed this on every host. And this problem exists only on one host
i found this https://lists.centos.org/pipermail/centos/2017-December/167727.html
the output is
[root@kvm380 ~]# ./test.sh
Sys Block Node : Device max_sectors_kb max_hw_sectors_kb
/sys/block/dm-0 : onn_kvm380-pool00_tmeta 256 4096
/sys/block/dm-1 : onn_kvm380-pool00_tdata 256 4096
/sys/block/dm-10 : onn_kvm380-var 256 4096
/sys/block/dm-11 : onn_kvm380-tmp 256 4096
/sys/block/dm-12 : onn_kvm380-home 256 4096
/sys/block/dm-13 : onn_kvm380-var_crash 256 4096
/sys/block/dm-2 : onn_kvm380-pool00-tpool 256 4096
/sys/block/dm-3 : onn_kvm380-ovirt--node--ng--4.3.7--0.20191121.0+1 256 4096
/sys/block/dm-4 : onn_kvm380-swap 256 4096
/sys/block/dm-5 : eui.0025385991b1e27a 512 2048
/sys/block/dm-6 : eui.0025385991b1e27a1 512 2048
/sys/block/dm-7 : onn_kvm380-pool00 256 4096
/sys/block/dm-8 : onn_kvm380-var_log_audit 256 4096
/sys/block/dm-9 : onn_kvm380-var_log 256 4096
cat: /sys/block/nvme0n1/device/vendor: Datei oder Verzeichnis nicht gefunden
/sys/block/nvme0n1: Samsung SSD 970 EVO 1TB 512 2048
/sys/block/sda : HP LOGICAL VOLUME 256 4096
is the nvme not starting correct
[root@kvm380 ~]# systemctl status multipathd
● multipathd.service - Device-Mapper Multipath Device Controller
Loaded: loaded (/usr/lib/systemd/system/multipathd.service; enabled; vendor preset: enabled)
Active: active (running) since Di 2019-12-31 16:16:32 CET; 31min ago
Process: 1919 ExecStart=/sbin/multipathd (code=exited, status=0/SUCCESS)
Process: 1916 ExecStartPre=/sbin/multipath -A (code=exited, status=0/SUCCESS)
Process: 1911 ExecStartPre=/sbin/modprobe dm-multipath (code=exited, status=0/SUCCESS)
Main PID: 1921 (multipathd)
Tasks: 7
CGroup: /system.slice/multipathd.service
└─1921 /sbin/multipathd
Dez 31 16:47:58 kvm380.durchhalten.intern multipathd[1921]: nvme0n1: mark as failed
Dez 31 16:47:58 kvm380.durchhalten.intern multipathd[1921]: eui.0025385991b1e27a: Entering recovery mode: max_retries=4
Dez 31 16:47:58 kvm380.durchhalten.intern multipathd[1921]: eui.0025385991b1e27a: remaining active paths: 0
Dez 31 16:48:02 kvm380.durchhalten.intern multipathd[1921]: 259:0: reinstated
Dez 31 16:48:02 kvm380.durchhalten.intern multipathd[1921]: eui.0025385991b1e27a: queue_if_no_path enabled
Dez 31 16:48:02 kvm380.durchhalten.intern multipathd[1921]: eui.0025385991b1e27a: Recovered to normal mode
Dez 31 16:48:02 kvm380.durchhalten.intern multipathd[1921]: eui.0025385991b1e27a: remaining active paths: 1
Dez 31 16:48:03 kvm380.durchhalten.intern multipathd[1921]: nvme0n1: mark as failed
Dez 31 16:48:03 kvm380.durchhalten.intern multipathd[1921]: eui.0025385991b1e27a: Entering recovery mode: max_retries=4
Dez 31 16:48:03 kvm380.durchhalten.intern multipathd[1921]: eui.0025385991b1e27a: remaining active paths: 0
why is it marked as failed?
if i create a new volume with cockpit and use it for bricks for gluster, every thing is fine. until reboot
maybe some one can point me the direction
4 years, 12 months
4.2.8 to 4.3.7 > Management slow
by Demeter Tibor
Dear Users,
I've successfully upgraded my 4 node hyperconverged system from 4.2.8 to 4.3.7.
After upgrade everything seems to working fine, but the whole management system seems very slow.
Spends many seconds when I clicking on "virtual machines" or I want to edit a virtual machines.
The speed of vms and the IO is fine.
It is running on a glusterfs (distributed replicate, on 3 node, 9 bricks). There are no errors, everything fine. but terrible slow:(
The engine vm has 0.2-0.3 load.
What can I do?
Thanks in advance and I wish Happy New Year!
Regards,
Tibor
4 years, 12 months
Re: Issue deploying self hosted engine on new install
by Strahil
Hi,
Most probably it is a bug.
Can you add 'debugger: on_failed' to the task ?
Maybe it will provide a hint where is the issue.
Best Regards,
Strahil NikolovOn Dec 31, 2019 06:36, realapyo(a)gmail.com wrote:
>
> Hi,
>
> I have encountered the same error that results in the failure of hosted engine deployment. I have tried to replace "ovirt_host_facts" by "ovirt_host_info" in /usr/share/ansible/roles/ovirt.hosted_engine_setup/tasks/bootstrap_local_vm/05_add_host.yml however it did not help with the following error message:
>
> 2019-12-31 11:23:27,595+0900 INFO otopi.ovirt_hosted_engine_setup.ansible_utils ansible_utils._process_output:109 TASK [ovirt.hosted_engine_setup : Wait for the host to be up]
> 2019-12-31 11:23:28,997+0900 DEBUG otopi.ovirt_hosted_engine_setup.ansible_utils ansible_utils._process_output:103 {u'msg': u"The conditional check 'host_result_up_check is succeeded and host_result_up_check.ansible_facts.ovirt_hosts|length >= 1 and ( host_result_up_check.ansible_facts.ovirt_hosts[0].status == 'up' or host_result_up_check.ansible_facts.ovirt_hosts[0].status == 'non_operational' )' failed. The error was: error while evaluating conditional (host_result_up_check is succeeded and host_result_up_check.ansible_facts.ovirt_hosts|length >= 1 and ( host_result_up_check.ansible_facts.ovirt_hosts[0].status == 'up' or host_result_up_check.ansible_facts.ovirt_hosts[0].status == 'non_operational' )): 'dict object' has no attribute 'ansible_facts'", u'_ansible_no_log': False}
> 2019-12-31 11:23:29,098+0900 ERROR otopi.ovirt_hosted_engine_setup.ansible_utils ansible_utils._process_output:107 fatal: [localhost]: FAILED! => {"msg": "The conditional check 'host_result_up_check is succeeded and host_result_up_check.ansible_facts.ovirt_hosts|length >= 1 and ( host_result_up_check.ansible_facts.ovirt_hosts[0].status == 'up' or host_result_up_check.ansible_facts.ovirt_hosts[0].status == 'non_operational' )' failed. The error was: error while evaluating conditional (host_result_up_check is succeeded and host_result_up_check.ansible_facts.ovirt_hosts|length >= 1 and ( host_result_up_check.ansible_facts.ovirt_hosts[0].status == 'up' or host_result_up_check.ansible_facts.ovirt_hosts[0].status == 'non_operational' )): 'dict object' has no attribute 'ansible_facts'"}
>
> According to the latest doc of Ansible (2.9) regarding "ovirt_host_facts", it is deprecated and one should use "ovirt_host_info" instead as you can refer here: https://docs.ansible.com/ansible/latest/modules/ovirt_host_facts_module.html
>
> And one of the requirements to use "ovirt_host_info" is "ovirt-engine-sdk-python >= 4.3.0" however in my installation, the installed package is "python-ovirt-engine-sdk4 (4.3.2-2.el7)" instead of "ovirt-engine-sdk-python". The "ovirt-engine-sdk-python" package is available however the version is 3.6.9.1-1.el7.
>
> I have been trying to install the latest oVirt (based in 4.3 repository) and hyperconverged mode using 3 nodes. Tried already with cockpit installation and now debugging it by using CLI installation.
>
> Is this a bug or is there already a fix?
>
> Best regards,
> Sang-Un
> _______________________________________________
> Users mailing list -- users(a)ovirt.org
> To unsubscribe send an email to users-leave(a)ovirt.org
> Privacy Statement: https://www.ovirt.org/site/privacy-policy/
> oVirt Code of Conduct: https://www.ovirt.org/community/about/community-guidelines/
> List Archives: https://lists.ovirt.org/archives/list/users@ovirt.org/message/V2OYCPZAHUV...
4 years, 12 months
Re: Cannot add Foreman as a provider
by Strahil
Maybe 6.1 means Foreman with Red Hat Satellite 6.1 ?
Best Regards,
Strahil NikolovOn Dec 30, 2019 16:41, Robert Webb <rwebb(a)ropeguru.com> wrote:
>
> oVirt – 4.3.7
>
> Foreman - foreman-1.24.0-15.el7.noarch
>
>
>
> I have not tried any downgrades at this point.
>
>
>
> Looking at this page, https://www.ovirt.org/documentation/admin-guide/chap-External_Providers.html , it would seem that 4.3.x of oVirt is not compatible with any version of Foreman, as the page shows oVirt 4.2 is supported with Foreman 6.1. Which also confuses me as I cannot find any Foreman version 6.1.
>
>
>
>
>
> From: Strahil Nikolov <hunter86_bg(a)yahoo.com>
> Sent: Sunday, December 29, 2019 5:00 PM
> To: users <users(a)ovirt.org>; Robert Webb <rwebb(a)ropeguru.com>
> Subject: Re: [ovirt-users] Cannot add Foreman as a provider
>
>
4 years, 12 months
Cannot add Foreman as a provider
by Robert Webb
So I just setup a brand new Foreman instance and have been trying to get it setup with oVirt. I was able to install the oVirt plugins under Foreman and have the oVirt engine setup as a computer resource.
Where I am running into an issue is adding Foreman as an additional provider under oVirt.
In the log I get:
Command 'org.ovirt.engine.core.bll.provider.TestProviderConnectivityCommand' failed: EngineException: https://foreman.saroza-webb.internal:443/api/v2/discovered_hosts?format=j... (Failed with error PROVIDER_FAILURE and code 5050)
I have done some searching, but everything I have found seems to be from years ago and deals with oVirt 3.5.x
Any thoughts?
4 years, 12 months