On Sun, Sep 29, 2019 at 1:13 AM Nir Soffer <nsoffer(a)redhat.com>
wrote:
> Since Friday build-artifacts on s390x get stuck again, so we cannot run
> OST.
> This is not a new issue, we have issues on s390x every few weeks.
>
> I posted this patch to disable this job:
>
https://gerrit.ovirt.org/c/97851
>
The patch does not work, it still runs the s390x job, and it run it
incorrectly.
Maybe the issue is not at the slave, but at vdsm automation scripts?
We have this yam:
19 stages:
20 - build-artifacts:
21 substages:
22 - build-py27:
23 archs:
24 - ppc64le
25 - x86_64
26 - build-py37:
27 distributions:
28 - fc30
And we get these jobs:
- build-artifacts.build-py27.el7.ppc64le
- build-artifacts.build-py27.el7.x86_64
- build-artifacts.build-py27.fc29.x86_64
- build-artifacts.build-py37.fc30.x86_64
- build-artifacts.fc29.s390x
The last job - s390x looks wrong - we should have only
build-py27 and build-py37 jobs, using:
- automation/build-artifacts.build-py27.sh
- automation/build-artifacts.build-py37.sh
But both scripts are using a symlinks:
lrwxrwxrwx. 1 nsoffer nsoffer 18 Sep 29 00:55 automation/
build-artifacts.build-py27.sh -> build-artifacts.sh
lrwxrwxrwx. 1 nsoffer nsoffer 18 Sep 29 00:55 automation/
build-artifacts.build-py37.sh -> build-artifacts.sh
-rwxrwxr-x. 1 nsoffer nsoffer 346 Sep 17 02:54
automation/build-artifacts.sh
Is it possible the the CI find build-artifacts.sh and run it even when no
sub stage is specified?
I'll try to rename this script to avoid this.