I have now enqueued another run of OST for the next VDSM patch in
line
(that was merged around the time of the first failure)
https://gerrit.ovirt.org/#/c/75299/2 to see if it was the culprit.
This patch is xmlrpc related. Can you point me to the issue it caused?
Not sure why we allow only one (*globally for all developers)
*concurrent
run of the manual OST job... I've created a jira issue for this
https://ovirt-jira.atlassian.net/browse/OST-61
Martin
On Mon, Apr 10, 2017 at 3:29 PM, Martin Betak <mbetak(a)redhat.com> wrote:
> Actually it appears that this failure is not engine related (tried
> reverting my patches from saturday but still received similar errors with
> lago in the [ add_secondary_storage_domains] test).
> But when running the current engine master with VDSM cutoff before the
> time of failure, it passed
http://jenkins.ovirt.or
> g/job/ovirt-system-tests_manual/229/
>
> This was the VDSM commit that *worked*
> *Commit 5c8aff6177bdaa81ee11c20d417c9bb10e651fb8 by Francesco Romani
> <
http://jenkins.ovirt.org/user/fromani@redhat.com/>*
>
> On Mon, Apr 10, 2017 at 11:08 AM, Michal Skrivanek <
> michal.skrivanek(a)redhat.com> wrote:
>
>>
>> > On 9 Apr 2017, at 09:42, Barak Korren <bkorren(a)redhat.com> wrote:
>> >
>> > On 9 April 2017 at 10:39, Yaniv Kaul <ykaul(a)redhat.com> wrote:
>> >> 1. The error:
>> >> 2017-04-08 14:12:11,376-04 ERROR
>> >> [org.ovirt.vdsm.jsonrpc.client.JsonRpcClient] (ResponseWorker) []
>> Not able
>> >> to update response for "fc0cef3f-d8fa-4074-b315-e36cc2f63fa1"
>> >>
>> >> Is still seen, not sure why.
>> >>
>> >> 2. MOM is not available (unrelated, but still a bug)
>> >>
>> >> 3. New error I have not seen in the past on host1
>> >> (
http://jenkins.ovirt.org/job/test-repo_ovirt_experimental_m
>> aster/6257/artifact/exported-artifacts/basic-suit-master-el7
>> /test_logs/basic-suite-master/post-002_bootstrap.py/lago-bas
>> ic-suite-master-host1/_var_log/vdsm/vdsm.log
>> >> ) :
>> >>
>> >> 2017-04-08 14:12:22,208-0400 WARN (jsonrpc/6) [storage.LVM] lvm pvs
>> failed:
>> >> 5 [] [' Failed to find physical volume
>> >> "/dev/mapper/360014050eacbb3c8b21428fb6683f074".']
(lvm:325)
>> >>
>> >> 4. New warning:
>> >> 2017-04-08 14:13:17,327-0400 WARN (check/loop) [storage.asyncutils]
>> Call
>> >> <bound method DirectioChecker._check of <DirectioChecker
>> >> /dev/0eb0f05c-0507-4f02-ba13-b8d865538d7a/metadata running
>> >> next_check=4295183.36 at 0x2fb9190>> delayed by 1.43 seconds
>> >> (asyncutils:138)
>> >>
>> >>
>> >> Do we know when it began?
>> >
>> > Yes. I linked to the patch that seems to have started this.
>>
>> This should be easy to revert
>> Working on it
>> Then we’ll see what exactly is the culprit
>>
>> >
>> >
>> >
>> > --
>> > Barak Korren
>> > bkorren(a)redhat.com
>> > RHCE, RHCi, RHV-DevOps Team
>> >
https://ifireball.wordpress.com/
>> > _______________________________________________
>> > Devel mailing list
>> > Devel(a)ovirt.org
>> >
http://lists.ovirt.org/mailman/listinfo/devel
>> >
>> >
>>
>> _______________________________________________
>> Devel mailing list
>> Devel(a)ovirt.org
>>
http://lists.ovirt.org/mailman/listinfo/devel
>>
>
>