(although the engine fix is enough to get the ost to pass).
I'm not sure it passes OST without the engine fix, verifying it now.
Thanks,
Alona.
On Sun, Jul 1, 2018 at 11:21 AM, Ehud Yonasi <eyonasi(a)redhat.com> wrote:
Hey Alona,
what is the current status of the fix?
On Thu, Jun 28, 2018 at 5:57 PM Dafna Ron <dron(a)redhat.com> wrote:
> vdsm also failing on the same issue:
>
>
https://jenkins.ovirt.org/job/ovirt-master_change-queue-tester/8407/
>
> Thanks,
> Dafna
>
>
> On Thu, Jun 28, 2018 at 11:11 AM, Dafna Ron <dron(a)redhat.com> wrote:
>
>> Thanks Alona,
>> Can you please update me once you have a fix?
>>
>> Thanks,
>> Dafna
>>
>>
>> On Thu, Jun 28, 2018 at 10:28 AM, Alona Kaplan <alkaplan(a)redhat.com>
>> wrote:
>>
>>> Hi,
>>> I'm aware to the error. Francesco and me are working on it.
>>>
>>> Thank,
>>> Alona.
>>>
>>> On Thu, Jun 28, 2018, 12:23 Dafna Ron <dron(a)redhat.com> wrote:
>>>
>>>> ovirt-hosted-engine-ha failed on the same issue as well.
>>>>
>>>> On Thu, Jun 28, 2018 at 10:07 AM, Dafna Ron <dron(a)redhat.com>
wrote:
>>>>
>>>>> Hi,
>>>>>
>>>>> We had a failure in test 098_ovirt_provider_ovn.use_ovn_provider.
>>>>>
>>>>> Although CQ is pointing to this change:
https://gerrit.ovirt.org/#/c/
>>>>> 92567/ - packaging: Add python-netaddr requirement I actually think
>>>>> from the error its because of changes made to multiqueues
>>>>>
>>>>>
https://gerrit.ovirt.org/#/c/92009/ - engine: Update libvirtVmXml to
>>>>> consider vmBase.multiQueuesEnabled attribute
>>>>>
https://gerrit.ovirt.org/#/c/92008/ - engine: Introduce algorithm
>>>>> for calculating how many queues asign per vnic
>>>>>
https://gerrit.ovirt.org/#/c/92007/ - engine: Add multiQueuesEnabled
>>>>> to VmBase
>>>>>
https://gerrit.ovirt.org/#/c/92318/ - restapi: Add 'Multi Queues
>>>>> Enabled' to the relevant mappers
>>>>>
https://gerrit.ovirt.org/#/c/92149/ - webadmin: Add 'Multi
Queues
>>>>> Enabled' to vm dialog
>>>>>
>>>>> Alona, can you please take a look?
>>>>>
>>>>>
>>>>> *Link to Job:*
>>>>>
>>>>>
>>>>>
>>>>>
>>>>> *http://jenkins.ovirt.org/job/ovirt-master_change-queue-tester/8375/
>>>>>
<
http://jenkins.ovirt.org/job/ovirt-master_change-queue-tester/8375/>Link
>>>>> to all
>>>>>
logs:https://jenkins.ovirt.org/job/ovirt-master_change-queue-tester/8375/...
>>>>>
<
https://jenkins.ovirt.org/job/ovirt-master_change-queue-tester/8375/artif...
>>>>> error snippet from the log: <error>*
>>>>> *eng**ine: *
>>>>>
>>>>> 2018-06-27 13:59:25,976-04 ERROR [org.ovirt.engine.core.
>>>>> vdsbroker.vdsbroker.GetAllVmStatsVDSCommand]
>>>>> (EE-ManagedThreadFactory-engineScheduled-Thread-80) [] Command
>>>>> 'GetAllVmStatsVDSCommand(HostName =
lago-basic-suite-master-host-1,
>>>>>
VdsIdVDSCommandParametersBase:{hostId='d9094c95-3275-4616-b4c2-815e753bcfed'})'
>>>>> execution failed: VDSGenericException: VDSNetworkException: Broken
pipe
>>>>> 2018-06-27 13:59:25,977-04 DEBUG
[org.ovirt.engine.core.utils.threadpool.ThreadPoolUtil]
>>>>> (EE-ManagedThreadFactory-engine-Thread-442) [] Executing task:
>>>>> EE-ManagedThreadFactory-engine-Thread-442
>>>>> 2018-06-27 13:59:25,977-04 DEBUG [org.ovirt.engine.core.common.
>>>>> di.interceptor.DebugLoggingInterceptor]
(EE-ManagedThreadFactory-engine-Thread-442)
>>>>> [] method: getVdsManager, params:
[d9094c95-3275-4616-b4c2-815e753bcfed],
>>>>> timeElapsed: 0ms
>>>>> 2018-06-27 13:59:25,977-04 WARN
[org.ovirt.engine.core.vdsbroker.VdsManager]
>>>>> (EE-ManagedThreadFactory-engine-Thread-442) [] Host
>>>>> 'lago-basic-suite-master-host-1' is not responding.
>>>>> 2018-06-27 13:59:25,979-04 ERROR [org.ovirt.engine.core.dal.
>>>>> dbbroker.auditloghandling.AuditLogDirector]
(EE-ManagedThreadFactory-engineScheduled-Thread-63)
>>>>> [] EVENT_ID: VDS_BROKER_COMMAND_FAILURE(10,802), VDSM
>>>>> lago-basic-suite-master-host-1 command GetStatsAsyncVDS failed:
Broken pipe
>>>>> 2018-06-27 13:59:25,976-04 DEBUG [org.ovirt.engine.core.
>>>>> vdsbroker.vdsbroker.GetAllVmStatsVDSCommand]
>>>>> (EE-ManagedThreadFactory-engineScheduled-Thread-80) [] Exception:
>>>>> org.ovirt.engine.core.vdsbroker.vdsbroker.VDSNetworkException:
>>>>> VDSGenericException: VDSNetworkException: Broken pipe
>>>>> at org.ovirt.engine.core.vdsbroker.vdsbroker.
>>>>>
BrokerCommandBase.proceedProxyReturnValue(BrokerCommandBase.java:189)
>>>>> [vdsbroker.jar:]
>>>>> at org.ovirt.engine.core.vdsbroker.vdsbroker.
>>>>> GetAllVmStatsVDSCommand.executeVdsBrokerCommand(
>>>>> GetAllVmStatsVDSCommand.java:23) [vdsbroker.jar:]
>>>>> at org.ovirt.engine.core.vdsbroker.vdsbroker.
>>>>>
VdsBrokerCommand.executeVdsCommandWithNetworkEvent(VdsBrokerCommand.java:123)
>>>>> [vdsbroker.jar:]
>>>>> at org.ovirt.engine.core.vdsbroker.vdsbroker.
>>>>> VdsBrokerCommand.executeVDSCommand(VdsBrokerCommand.java:111)
>>>>> [vdsbroker.jar:]
>>>>> at org.ovirt.engine.core.vdsbroker.VDSCommandBase.
>>>>> executeCommand(VDSCommandBase.java:65) [vdsbroker.jar:]
>>>>> at
org.ovirt.engine.core.dal.VdcCommandBase.execute(VdcCommandBase.java:31)
>>>>> [dal.jar:]
>>>>> at org.ovirt.engine.core.vdsbroker.vdsbroker.
>>>>> DefaultVdsCommandExecutor.execute(DefaultVdsCommandExecutor.java:14)
>>>>> [vdsbroker.jar:]
>>>>> at org.ovirt.engine.core.vdsbroker.ResourceManager.
>>>>> runVdsCommand(ResourceManager.java:399) [vdsbroker.jar:]
>>>>> at org.ovirt.engine.core.vdsbroker.ResourceManager$
>>>>> Proxy$_$$_WeldSubclass.runVdsCommand$$super(Unknown Source)
>>>>> [vdsbroker.jar:]
>>>>> at sun.reflect.GeneratedMethodAccessor270.invoke(Unknown
>>>>> Source) [:1.8.0_171]
>>>>> at sun.reflect.DelegatingMethodAccessorImpl.invoke(
>>>>> DelegatingMethodAccessorImpl.java:43) [rt.jar:1.8.0_171]
>>>>> at java.lang.reflect.Method.invoke(Method.java:498)
>>>>> [rt.jar:1.8.0_171]
>>>>> at org.jboss.weld.interceptor.proxy.
>>>>> TerminalAroundInvokeInvocationContext.proceedInternal(
>>>>> TerminalAroundInvokeInvocationContext.java:49)
>>>>> [weld-core-impl-2.4.3.Final.jar:2.4.3.Final]
>>>>> at org.jboss.weld.interceptor.proxy.
>>>>>
AroundInvokeInvocationContext.proceed(AroundInvokeInvocationContext.java:77)
>>>>> [weld-core-impl-2.4.3.Final.jar:2.4.3.Final]
>>>>> at org.ovirt.engine.core.common.di.interceptor.
>>>>> LoggingInterceptor.apply(LoggingInterceptor.java:12) [common.jar:]
>>>>> at sun.reflect.GeneratedMethodAccessor68.invoke(Unknown
>>>>> Source) [:1.8.0_171]
>>>>> at sun.reflect.DelegatingMethodAccessorImpl.invoke(
>>>>> DelegatingMethodAccessorImpl.java:43) [rt.jar:1.8.0_171]
>>>>> at java.lang.reflect.Method.invoke(Method.java:498)
>>>>> [rt.jar:1.8.0_171]
>>>>> at org.jboss.weld.interceptor.reader.
>>>>> SimpleInterceptorInvocation$SimpleMethodInvocation.invoke(
>>>>> SimpleInterceptorInvocation.java:73) [weld-core-impl-2.4.3.Final.
>>>>> jar:2.4.3.Final]
>>>>> at
org.jboss.weld.interceptor.proxy.InterceptorMethodHandler.
>>>>> executeAroundInvoke(InterceptorMethodHandler.java:84)
>>>>> [weld-core-impl-2.4.3.Final.jar:2.4.3.Final]
>>>>> at
org.jboss.weld.interceptor.proxy.InterceptorMethodHandler.
>>>>> executeInterception(InterceptorMethodHandler.java:72)
>>>>> [weld-core-impl-2.4.3.Final.jar:2.4.3.Final]
>>>>> at org.jboss.weld.interceptor.pro
>>>>> xy.InterceptorMethodHandler.invoke(InterceptorMethodHandler.java:56)
>>>>> [weld-core-impl-2.4.3.Final.jar:2.4.3.Final]
>>>>> at org.jboss.weld.bean.proxy.CombinedInterceptorAndDecorato
>>>>>
rStackMethodHandler.invoke(CombinedInterceptorAndDecoratorStackMethodHandler.java:79)
>>>>> [weld-core-impl-2.4.3.Final.jar:2.4.3.Final]
>>>>> at org.jboss.weld.bean.proxy.CombinedInterceptorAndDecorato
>>>>>
rStackMethodHandler.invoke(CombinedInterceptorAndDecoratorStackMethodHandler.java:68)
>>>>> [weld-core-impl-2.4.3.Final.jar:2.4.3.Final]
>>>>> at org.ovirt.engine.core.vdsbroker.ResourceManager$
>>>>> Proxy$_$$_WeldSubclass.runVdsCommand(Unknown Source)
[vdsbroker.jar:]
>>>>> at org.ovirt.engine.core.vdsbroker.monitoring.
>>>>> VmsStatisticsFetcher.poll(VmsStatisticsFetcher.java:29)
>>>>> [vdsbroker.jar:]
>>>>> at org.ovirt.engine.core.vdsbroker.monitoring.
>>>>> VmsListFetcher.fetch(VmsListFetcher.java:49) [vdsbroker.jar:]
>>>>> at org.ovirt.engine.core.vdsbroker.monitoring.
>>>>> PollVmStatsRefresher.poll(PollVmStatsRefresher.java:44)
>>>>> [vdsbroker.jar:]
>>>>> at
java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
>>>>> [rt.jar:1.8.0_171]
>>>>> at
java.util.concurrent.FutureTask.runAndReset(FutureTask.java:308)
>>>>> [rt.jar:1.8.0_171]
>>>>> at org.glassfish.enterprise.concurrent.internal.
>>>>> ManagedScheduledThreadPoolExecutor$ManagedScheduledFutureTask.
>>>>> access$201(ManagedScheduledThreadPoolExecutor.java:383)
>>>>> [javax.enterprise.concurrent-1.0.jar:]
>>>>> at org.glassfish.enterprise.concurrent.internal.
>>>>> ManagedScheduledThreadPoolExecutor$ManagedScheduledFutureTask.run(
>>>>> ManagedScheduledThreadPoolExecutor.java:534)
>>>>> [javax.enterprise.concurrent-1.0.jar:]
>>>>> at
java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
>>>>> [rt.jar:1.8.0_171]
>>>>> at
java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
>>>>> [rt.jar:1.8.0_171]
>>>>> at java.lang.Thread.run(Thread.java:748) [rt.jar:1.8.0_171]
>>>>> at org.glassfish.enterprise.concurrent.
>>>>> ManagedThreadFactoryImpl$ManagedThread.run(
>>>>> ManagedThreadFactoryImpl.java:250) [javax.enterprise.concurrent-
>>>>> 1.0.jar:]
>>>>> at org.jboss.as.ee.concurrent.service.
>>>>> ElytronManagedThreadFactory$ElytronManagedThread.run(
>>>>> ElytronManagedThreadFactory.java:78)
>>>>>
>>>>> 2018-06-27 13:59:25,984-04 DEBUG [org.ovirt.engine.core.
>>>>> vdsbroker.vdsbroker.GetAllVmStatsVDSCommand]
>>>>> (EE-ManagedThreadFactory-engineScheduled-Thread-80) [] FINISH,
>>>>> GetAllVmStatsVDSCommand, return: , log id: 56d99e77
>>>>> 2018-06-27 13:59:25,984-04 DEBUG [org.ovirt.engine.core.common.
>>>>> di.interceptor.DebugLoggingInterceptor]
(EE-ManagedThreadFactory-engineScheduled-Thread-80)
>>>>> [] method: runVdsCommand, params: [GetAllVmStats,
>>>>>
VdsIdVDSCommandParametersBase:{hostId='d9094c95-3275-4616-b4c2-815e753bcfed'}],
>>>>> timeElapsed: 1497ms
>>>>> 2018-06-27 13:59:25,984-04 INFO [org.ovirt.engine.core.
>>>>> vdsbroker.monitoring.PollVmStatsRefresher]
(EE-ManagedThreadFactory-engineScheduled-Thread-80)
>>>>> [] Failed to fetch vms info for host
'lago-basic-suite-master-host-1'
>>>>> - skipping VMs monitoring.
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>>
>>>>> *vdsm: 2018-06-27 14:10:17,314-0400 INFO (jsonrpc/7) [virt.vm]
>>>>> (vmId='b8a11304-07e3-4e64-af35-7421be780d5b') Hotunplug NIC
xml: <?xml
>>>>> version='1.0' encoding='utf-8'?><interface
type="bridge"> <address
>>>>> bus="0x00" domain="0x0000"
function="0x0" slot="0x0b" type="pci" /> <mac
>>>>> address="00:1a:4a:16:01:0e" /> <model
type="virtio" /> <source
>>>>> bridge="network_1" /> <link state="up"
/> <driver name="vhost"
>>>>> queues="" /> <alias
name="ua-3c77476f-f194-476a-8412-d76a9e58d1f9"
>>>>> /></interface> (vm:3321)2018-06-27 14:10:17,328-0400 ERROR
(jsonrpc/7)
>>>>> [virt.vm] (vmId='b8a11304-07e3-4e64-af35-7421be780d5b')
Hotunplug failed
>>>>> (vm:3353)Traceback (most recent call last): File
>>>>> "/usr/lib/python2.7/site-packages/vdsm/virt/vm.py", line
3343, in
>>>>> hotunplugNic self._dom.detachDevice(nicXml) File
>>>>> "/usr/lib/python2.7/site-packages/vdsm/virt/virdomain.py",
line 99, in f
>>>>> ret = attr(*args, **kwargs) File
>>>>>
"/usr/lib/python2.7/site-packages/vdsm/common/libvirtconnection.py", line
>>>>> 131, in wrapper ret = f(*args, **kwargs) File
>>>>> "/usr/lib/python2.7/site-packages/vdsm/common/function.py",
line 93, in
>>>>> wrapper return func(inst, *args, **kwargs) File
>>>>> "/usr/lib64/python2.7/site-packages/libvirt.py", line 1177,
in
>>>>> detachDevice if ret == -1: raise libvirtError
('virDomainDetachDevice()
>>>>> failed', dom=self)libvirtError: 'queues' attribute must
be positive number:
>>>>> 2018-06-27 14:10:17,345-0400 DEBUG (jsonrpc/7) [api] FINISH
hotunplugNic
>>>>> response={'status': {'message':
"'queues' attribute must be positive
>>>>> number: ", 'code': 50}} (api:136)2018-06-27
14:10:17,346-0400 INFO
>>>>> (jsonrpc/7) [api.virt] FINISH hotunplugNic return={'status':
{'message':
>>>>> "'queues' attribute must be positive number: ",
'code': 50}}
>>>>> from=::ffff:192.168.201.4,32976, flow_id=ecb6652,
>>>>> vmId=b8a11304-07e3-4e64-af35-7421be780d5b (api:53)2018-06-27
>>>>> 14:10:17,346-0400 INFO (jsonrpc/7) [jsonrpc.JsonRpcServer] RPC call
>>>>> VM.hotunplugNic failed (error 50) in 0.07 seconds
(__init__:311)2018-06-27
>>>>> 14:10:19,244-0400 DEBUG (qgapoller/2) [vds] Not sending QEMU-GA
command
>>>>> 'guest-get-users' to
vm_id='b8a11304-07e3-4e64-af35-7421be780d5b', command
>>>>> is not supported (qemuguestagent:192)2018-06-27 14:10:20,038-0400
DEBUG
>>>>> (jsonrpc/1) [jsonrpc.JsonRpcServer] Calling
'Host.getAllVmStats' in bridge
>>>>> with {} (__init__:328)2018-06-27 14:10:20,038-0400 INFO (jsonrpc/1)
>>>>> [api.host] START getAllVmStats() from=::1,48032 (api:47)2018-06-27
>>>>> 14:10:20,041-0400 INFO (jsonrpc/1) [api.host] FINISH getAllVmStats
>>>>> return={'status': {'message': 'Done',
'code': 0}, 'statsList':
>>>>> (suppressed)} from=::1,48032 (api:53)2018-06-27 14:10:20,043-0400
DEBUG
>>>>> (jsonrpc/1) [jsonrpc.JsonRpcServer] Return
'Host.getAllVmStats' in bridge
>>>>> with (suppressed) (__init__:355)2018-06-27 14:10:20,043-0400 INFO
>>>>> (jsonrpc/1) [jsonrpc.JsonRpcServer] RPC call Host.getAllVmStats
succeeded
>>>>> in 0.00 seconds (__init__:311)2018-06-27 14:10:20,057-0400 DEBUG
>>>>> (jsonrpc/6) [jsonrpc.JsonRpcServer] Calling
'Host.getAllVmIoTunePolicies'
>>>>> in bridge with {} (__init__:328)2018-06-27 14:10:20,058-0400 INFO
>>>>> (jsonrpc/6) [api.host] START getAllVmIoTunePolicies() from=::1,48032
>>>>> (api:47)2018-06-27 14:10:20,058-0400 INFO (jsonrpc/6) [api.host]
FINISH
>>>>> getAllVmIoTunePolicies return={'status': {'message':
'Done', 'code': 0},
>>>>> 'io_tune_policies_dict':
{'b8a11304-07e3-4e64-af35-7421be780d5b':
>>>>> {'policy': [], 'current_values': [{'ioTune':
{'write_bytes_sec': 0L,
>>>>> 'total_iops_sec': 0L, 'read_iops_sec': 0L,
'read_bytes_sec': 0L,
>>>>> 'write_iops_sec': 0L, 'total_bytes_sec': 0L},
'path':
>>>>>
'/rhev/data-center/mnt/blockSD/cf23ceeb-81a3-4714-85a0-c6ddd1e024da/images/650fe4ae-47a1-4f2d-9cba-1617a8c868c3/03e75c3c-24e7-4e68-a6f1-21728aaaa73e',
>>>>> 'name': 'vda'}]}}} from=::1,48032 (api:53)2018-06-27
14:10:20,059-0400
>>>>> DEBUG (jsonrpc/6) [jsonrpc.JsonRpcServer] Return
>>>>> 'Host.getAllVmIoTunePolicies' in bridge with
>>>>> {'b8a11304-07e3-4e64-af35-7421be780d5b': {'policy':
[], 'current_values':
>>>>> [{'ioTune': {'write_bytes_sec': 0L,
'total_iops_sec': 0L, 'read_iops_sec':
>>>>> 0L, 'read_bytes_sec': 0L, 'write_iops_sec': 0L,
'total_bytes_sec': 0L},
>>>>> 'path':
>>>>>
'/rhev/data-center/mnt/blockSD/cf23ceeb-81a3-4714-85a0-c6ddd1e024da/images/650fe4ae-47a1-4f2d-9cba-1617a8c868c3/03e75c3c-24e7-4e68-a6f1-21728aaaa73e',
>>>>> 'name': 'vda'}]}} (__init__:355)</error>*
>>>>>
>>>>
>>>>
>>
> _______________________________________________
> Infra mailing list -- infra(a)ovirt.org
> To unsubscribe send an email to infra-leave(a)ovirt.org
> Privacy Statement:
https://www.ovirt.org/site/privacy-policy/
> oVirt Code of Conduct:
https://www.ovirt.org/community/about/community-
> guidelines/
> List Archives:
https://lists.ovirt.org/archives/list/infra@ovirt.org/
> message/52RUJGRYJVGAYTEVHD2PUCVINQRHE5QQ/
>