[ovirt-devel] [ OST Failure Report ] [ oVirt master ] [ 9/4/2017 ] [ add_secondary_storage_domains ]
Piotr Kliczewski
pkliczew at redhat.com
Mon Apr 10 14:01:42 UTC 2017
On Mon, Apr 10, 2017 at 3:39 PM, Martin Betak <mbetak at redhat.com> wrote:
> I have now enqueued another run of OST for the next VDSM patch in line
> (that was merged around the time of the first failure)
> https://gerrit.ovirt.org/#/c/75299/2 to see if it was the culprit.
>
This patch is xmlrpc related. Can you point me to the issue it caused?
> Not sure why we allow only one (*globally for all developers) *concurrent
> run of the manual OST job... I've created a jira issue for this
> https://ovirt-jira.atlassian.net/browse/OST-61
>
> Martin
>
> On Mon, Apr 10, 2017 at 3:29 PM, Martin Betak <mbetak at redhat.com> wrote:
>
>> Actually it appears that this failure is not engine related (tried
>> reverting my patches from saturday but still received similar errors with
>> lago in the [ add_secondary_storage_domains] test).
>> But when running the current engine master with VDSM cutoff before the
>> time of failure, it passed http://jenkins.ovirt.or
>> g/job/ovirt-system-tests_manual/229/
>>
>> This was the VDSM commit that *worked*
>> *Commit 5c8aff6177bdaa81ee11c20d417c9bb10e651fb8 by Francesco Romani
>> <http://jenkins.ovirt.org/user/fromani@redhat.com/>*
>>
>> On Mon, Apr 10, 2017 at 11:08 AM, Michal Skrivanek <
>> michal.skrivanek at redhat.com> wrote:
>>
>>>
>>> > On 9 Apr 2017, at 09:42, Barak Korren <bkorren at redhat.com> wrote:
>>> >
>>> > On 9 April 2017 at 10:39, Yaniv Kaul <ykaul at redhat.com> wrote:
>>> >> 1. The error:
>>> >> 2017-04-08 14:12:11,376-04 ERROR
>>> >> [org.ovirt.vdsm.jsonrpc.client.JsonRpcClient] (ResponseWorker) []
>>> Not able
>>> >> to update response for "fc0cef3f-d8fa-4074-b315-e36cc2f63fa1"
>>> >>
>>> >> Is still seen, not sure why.
>>> >>
>>> >> 2. MOM is not available (unrelated, but still a bug)
>>> >>
>>> >> 3. New error I have not seen in the past on host1
>>> >> (http://jenkins.ovirt.org/job/test-repo_ovirt_experimental_m
>>> aster/6257/artifact/exported-artifacts/basic-suit-master-el7
>>> /test_logs/basic-suite-master/post-002_bootstrap.py/lago-bas
>>> ic-suite-master-host1/_var_log/vdsm/vdsm.log
>>> >> ) :
>>> >>
>>> >> 2017-04-08 14:12:22,208-0400 WARN (jsonrpc/6) [storage.LVM] lvm pvs
>>> failed:
>>> >> 5 [] [' Failed to find physical volume
>>> >> "/dev/mapper/360014050eacbb3c8b21428fb6683f074".'] (lvm:325)
>>> >>
>>> >> 4. New warning:
>>> >> 2017-04-08 14:13:17,327-0400 WARN (check/loop) [storage.asyncutils]
>>> Call
>>> >> <bound method DirectioChecker._check of <DirectioChecker
>>> >> /dev/0eb0f05c-0507-4f02-ba13-b8d865538d7a/metadata running
>>> >> next_check=4295183.36 at 0x2fb9190>> delayed by 1.43 seconds
>>> >> (asyncutils:138)
>>> >>
>>> >>
>>> >> Do we know when it began?
>>> >
>>> > Yes. I linked to the patch that seems to have started this.
>>>
>>> This should be easy to revert
>>> Working on it
>>> Then we’ll see what exactly is the culprit
>>>
>>> >
>>> >
>>> >
>>> > --
>>> > Barak Korren
>>> > bkorren at redhat.com
>>> > RHCE, RHCi, RHV-DevOps Team
>>> > https://ifireball.wordpress.com/
>>> > _______________________________________________
>>> > Devel mailing list
>>> > Devel at ovirt.org
>>> > http://lists.ovirt.org/mailman/listinfo/devel
>>> >
>>> >
>>>
>>> _______________________________________________
>>> Devel mailing list
>>> Devel at ovirt.org
>>> http://lists.ovirt.org/mailman/listinfo/devel
>>>
>>
>>
>
-------------- next part --------------
An HTML attachment was scrubbed...
URL: <http://lists.ovirt.org/pipermail/devel/attachments/20170410/caf84076/attachment-0001.html>
More information about the Devel
mailing list