[oVirt Jenkins] ovirt-system-tests_basic-suite-4.3_nightly - Build
# 668 - Failure!
by jenkins@jenkins.phx.ovirt.org
Project: https://jenkins.ovirt.org/job/ovirt-system-tests_basic-suite-4.3_nightly/
Build: https://jenkins.ovirt.org/job/ovirt-system-tests_basic-suite-4.3_nightly/...
Build Number: 668
Build Status: Failure
Triggered By: Started by timer
-------------------------------------
Changes Since Last Success:
-------------------------------------
Changes for Build #668
[parthdhanjal] Updating max click version
-----------------
Failed Tests:
-----------------
1 tests failed.
FAILED: 004_basic_sanity.run_vms
Error Message:
Fault reason is "Operation Failed". Fault detail is "[Network error during communication with the Host.]". HTTP response code is 400.
Stack Trace:
Traceback (most recent call last):
File "/usr/lib64/python2.7/unittest/case.py", line 369, in run
testMethod()
File "/usr/lib/python2.7/site-packages/nose/case.py", line 197, in runTest
self.test(*self.arg)
File "/usr/lib/python2.7/site-packages/ovirtlago/testlib.py", line 142, in wrapped_test
test()
File "/usr/lib/python2.7/site-packages/ovirtlago/testlib.py", line 60, in wrapper
return func(get_test_prefix(), *args, **kwargs)
File "/home/jenkins/agent/workspace/ovirt-system-tests_basic-suite-4.3_nightly/ovirt-system-tests/basic-suite-4.3/test-scenarios/004_basic_sanity.py", line 694, in run_vms
vm0_service.start(use_cloud_init=True, vm=vm_params)
File "/usr/lib64/python2.7/site-packages/ovirtsdk4/services.py", line 32520, in start
return self._internal_action(action, 'start', None, headers, query, wait)
File "/usr/lib64/python2.7/site-packages/ovirtsdk4/service.py", line 299, in _internal_action
return future.wait() if wait else future
File "/usr/lib64/python2.7/site-packages/ovirtsdk4/service.py", line 55, in wait
return self._code(response)
File "/usr/lib64/python2.7/site-packages/ovirtsdk4/service.py", line 296, in callback
self._check_fault(response)
File "/usr/lib64/python2.7/site-packages/ovirtsdk4/service.py", line 134, in _check_fault
self._raise_error(response, body.fault)
File "/usr/lib64/python2.7/site-packages/ovirtsdk4/service.py", line 118, in _raise_error
raise error
Error: Fault reason is "Operation Failed". Fault detail is "[Network error during communication with the Host.]". HTTP response code is 400.
3 years, 6 months
Build failed in Jenkins:
system-sync_mirrors-centos-base-el8stream-x86_64 #53
by jenkins@jenkins.phx.ovirt.org
See <https://jenkins.ovirt.org/job/system-sync_mirrors-centos-base-el8stream-x...>
Changes:
------------------------------------------
[...truncated 487.66 KB...]
/var/www/html/repos/yum/centos-base-el8stream/2021-07-04-18-28
/var/www/html/repos/yum/centos-base-el8stream/2021-07-05-02-28
/var/www/html/repos/yum/centos-base-el8stream/2021-07-05-10-28
/var/www/html/repos/yum/centos-base-el8stream/2021-07-05-18-28
/var/www/html/repos/yum/centos-base-el8stream/2021-07-06-02-28
/var/www/html/repos/yum/centos-base-el8stream/2021-07-06-10-28
/var/www/html/repos/yum/centos-base-el8stream/2021-07-06-18-28
/var/www/html/repos/yum/centos-base-el8stream/2021-07-07-02-28
/var/www/html/repos/yum/centos-base-el8stream/2021-07-07-08-13
/var/www/html/repos/yum/centos-base-el8stream/2021-07-07-10-28
/var/www/html/repos/yum/centos-base-el8stream/2021-07-07-18-28
/var/www/html/repos/yum/centos-base-el8stream/2021-07-08-02-28
/var/www/html/repos/yum/centos-base-el8stream/2021-07-08-10-28
/var/www/html/repos/yum/centos-base-el8stream/2021-07-08-18-28
/var/www/html/repos/yum/centos-base-el8stream/2021-07-09-02-28
/var/www/html/repos/yum/centos-base-el8stream/2021-07-09-10-28
/var/www/html/repos/yum/centos-base-el8stream/2021-07-09-18-28
/var/www/html/repos/yum/centos-base-el8stream/2021-07-10-02-28
/var/www/html/repos/yum/centos-base-el8stream/2021-07-10-10-28
/var/www/html/repos/yum/centos-base-el8stream/2021-07-10-18-28
/var/www/html/repos/yum/centos-base-el8stream/2021-07-11-02-28
/var/www/html/repos/yum/centos-base-el8stream/2021-07-11-10-28
/var/www/html/repos/yum/centos-base-el8stream/2021-07-11-18-29
/var/www/html/repos/yum/centos-base-el8stream/2021-07-12-02-29
/var/www/html/repos/yum/centos-base-el8stream/2021-07-12-10-29
/var/www/html/repos/yum/centos-base-el8stream/2021-07-12-18-29
/var/www/html/repos/yum/centos-base-el8stream/2021-07-13-02-29
/var/www/html/repos/yum/centos-base-el8stream/2021-07-13-10-29
/var/www/html/repos/yum/centos-base-el8stream/2021-07-13-18-29
/var/www/html/repos/yum/centos-base-el8stream/2021-07-14-02-29
/var/www/html/repos/yum/centos-base-el8stream/2021-07-14-10-29
/var/www/html/repos/yum/centos-base-el8stream/2021-07-14-18-29
/var/www/html/repos/yum/centos-base-el8stream/2021-07-15-02-29
/var/www/html/repos/yum/centos-base-el8stream/2021-07-15-10-29
/var/www/html/repos/yum/centos-base-el8stream/2021-07-15-18-29
/var/www/html/repos/yum/centos-base-el8stream/2021-07-16-02-29
/var/www/html/repos/yum/centos-base-el8stream/2021-07-16-10-29
/var/www/html/repos/yum/centos-base-el8stream/2021-07-16-18-29 == *\/\v\a\r\/\w\w\w\/\h\t\m\l\/\r\e\p\o\s\/\y\u\m\/\c\e\n\t\o\s\-\b\a\s\e\-\e\l\8\s\t\r\e\a\m\/\2\0\2\1\-\0\7\-\1\6\-\0\2\-\2\9* ]]
+ for snapshot in '"${all_snapshots[@]}"'
+ [[ ! /var/www/html/repos/yum/centos-base-el8stream/2021-07-03-02-28
/var/www/html/repos/yum/centos-base-el8stream/2021-07-03-10-28
/var/www/html/repos/yum/centos-base-el8stream/2021-07-03-18-28
/var/www/html/repos/yum/centos-base-el8stream/2021-07-04-02-28
/var/www/html/repos/yum/centos-base-el8stream/2021-07-04-10-28
/var/www/html/repos/yum/centos-base-el8stream/2021-07-04-18-28
/var/www/html/repos/yum/centos-base-el8stream/2021-07-05-02-28
/var/www/html/repos/yum/centos-base-el8stream/2021-07-05-10-28
/var/www/html/repos/yum/centos-base-el8stream/2021-07-05-18-28
/var/www/html/repos/yum/centos-base-el8stream/2021-07-06-02-28
/var/www/html/repos/yum/centos-base-el8stream/2021-07-06-10-28
/var/www/html/repos/yum/centos-base-el8stream/2021-07-06-18-28
/var/www/html/repos/yum/centos-base-el8stream/2021-07-07-02-28
/var/www/html/repos/yum/centos-base-el8stream/2021-07-07-08-13
/var/www/html/repos/yum/centos-base-el8stream/2021-07-07-10-28
/var/www/html/repos/yum/centos-base-el8stream/2021-07-07-18-28
/var/www/html/repos/yum/centos-base-el8stream/2021-07-08-02-28
/var/www/html/repos/yum/centos-base-el8stream/2021-07-08-10-28
/var/www/html/repos/yum/centos-base-el8stream/2021-07-08-18-28
/var/www/html/repos/yum/centos-base-el8stream/2021-07-09-02-28
/var/www/html/repos/yum/centos-base-el8stream/2021-07-09-10-28
/var/www/html/repos/yum/centos-base-el8stream/2021-07-09-18-28
/var/www/html/repos/yum/centos-base-el8stream/2021-07-10-02-28
/var/www/html/repos/yum/centos-base-el8stream/2021-07-10-10-28
/var/www/html/repos/yum/centos-base-el8stream/2021-07-10-18-28
/var/www/html/repos/yum/centos-base-el8stream/2021-07-11-02-28
/var/www/html/repos/yum/centos-base-el8stream/2021-07-11-10-28
/var/www/html/repos/yum/centos-base-el8stream/2021-07-11-18-29
/var/www/html/repos/yum/centos-base-el8stream/2021-07-12-02-29
/var/www/html/repos/yum/centos-base-el8stream/2021-07-12-10-29
/var/www/html/repos/yum/centos-base-el8stream/2021-07-12-18-29
/var/www/html/repos/yum/centos-base-el8stream/2021-07-13-02-29
/var/www/html/repos/yum/centos-base-el8stream/2021-07-13-10-29
/var/www/html/repos/yum/centos-base-el8stream/2021-07-13-18-29
/var/www/html/repos/yum/centos-base-el8stream/2021-07-14-02-29
/var/www/html/repos/yum/centos-base-el8stream/2021-07-14-10-29
/var/www/html/repos/yum/centos-base-el8stream/2021-07-14-18-29
/var/www/html/repos/yum/centos-base-el8stream/2021-07-15-02-29
/var/www/html/repos/yum/centos-base-el8stream/2021-07-15-10-29
/var/www/html/repos/yum/centos-base-el8stream/2021-07-15-18-29
/var/www/html/repos/yum/centos-base-el8stream/2021-07-16-02-29
/var/www/html/repos/yum/centos-base-el8stream/2021-07-16-10-29
/var/www/html/repos/yum/centos-base-el8stream/2021-07-16-18-29 == *\/\v\a\r\/\w\w\w\/\h\t\m\l\/\r\e\p\o\s\/\y\u\m\/\c\e\n\t\o\s\-\b\a\s\e\-\e\l\8\s\t\r\e\a\m\/\2\0\2\1\-\0\7\-\1\6\-\1\0\-\2\9* ]]
+ for snapshot in '"${all_snapshots[@]}"'
+ [[ ! /var/www/html/repos/yum/centos-base-el8stream/2021-07-03-02-28
/var/www/html/repos/yum/centos-base-el8stream/2021-07-03-10-28
/var/www/html/repos/yum/centos-base-el8stream/2021-07-03-18-28
/var/www/html/repos/yum/centos-base-el8stream/2021-07-04-02-28
/var/www/html/repos/yum/centos-base-el8stream/2021-07-04-10-28
/var/www/html/repos/yum/centos-base-el8stream/2021-07-04-18-28
/var/www/html/repos/yum/centos-base-el8stream/2021-07-05-02-28
/var/www/html/repos/yum/centos-base-el8stream/2021-07-05-10-28
/var/www/html/repos/yum/centos-base-el8stream/2021-07-05-18-28
/var/www/html/repos/yum/centos-base-el8stream/2021-07-06-02-28
/var/www/html/repos/yum/centos-base-el8stream/2021-07-06-10-28
/var/www/html/repos/yum/centos-base-el8stream/2021-07-06-18-28
/var/www/html/repos/yum/centos-base-el8stream/2021-07-07-02-28
/var/www/html/repos/yum/centos-base-el8stream/2021-07-07-08-13
/var/www/html/repos/yum/centos-base-el8stream/2021-07-07-10-28
/var/www/html/repos/yum/centos-base-el8stream/2021-07-07-18-28
/var/www/html/repos/yum/centos-base-el8stream/2021-07-08-02-28
/var/www/html/repos/yum/centos-base-el8stream/2021-07-08-10-28
/var/www/html/repos/yum/centos-base-el8stream/2021-07-08-18-28
/var/www/html/repos/yum/centos-base-el8stream/2021-07-09-02-28
/var/www/html/repos/yum/centos-base-el8stream/2021-07-09-10-28
/var/www/html/repos/yum/centos-base-el8stream/2021-07-09-18-28
/var/www/html/repos/yum/centos-base-el8stream/2021-07-10-02-28
/var/www/html/repos/yum/centos-base-el8stream/2021-07-10-10-28
/var/www/html/repos/yum/centos-base-el8stream/2021-07-10-18-28
/var/www/html/repos/yum/centos-base-el8stream/2021-07-11-02-28
/var/www/html/repos/yum/centos-base-el8stream/2021-07-11-10-28
/var/www/html/repos/yum/centos-base-el8stream/2021-07-11-18-29
/var/www/html/repos/yum/centos-base-el8stream/2021-07-12-02-29
/var/www/html/repos/yum/centos-base-el8stream/2021-07-12-10-29
/var/www/html/repos/yum/centos-base-el8stream/2021-07-12-18-29
/var/www/html/repos/yum/centos-base-el8stream/2021-07-13-02-29
/var/www/html/repos/yum/centos-base-el8stream/2021-07-13-10-29
/var/www/html/repos/yum/centos-base-el8stream/2021-07-13-18-29
/var/www/html/repos/yum/centos-base-el8stream/2021-07-14-02-29
/var/www/html/repos/yum/centos-base-el8stream/2021-07-14-10-29
/var/www/html/repos/yum/centos-base-el8stream/2021-07-14-18-29
/var/www/html/repos/yum/centos-base-el8stream/2021-07-15-02-29
/var/www/html/repos/yum/centos-base-el8stream/2021-07-15-10-29
/var/www/html/repos/yum/centos-base-el8stream/2021-07-15-18-29
/var/www/html/repos/yum/centos-base-el8stream/2021-07-16-02-29
/var/www/html/repos/yum/centos-base-el8stream/2021-07-16-10-29
/var/www/html/repos/yum/centos-base-el8stream/2021-07-16-18-29 == *\/\v\a\r\/\w\w\w\/\h\t\m\l\/\r\e\p\o\s\/\y\u\m\/\c\e\n\t\o\s\-\b\a\s\e\-\e\l\8\s\t\r\e\a\m\/\2\0\2\1\-\0\7\-\1\6\-\1\8\-\2\9* ]]
+ rm referenced_pkgs.txt
+ install_repo_pubkey centos-base-el8stream jenkins/data/mirrors-reposync.conf
+ local repo_name=centos-base-el8stream
+ local reposync_conf=jenkins/data/mirrors-reposync.conf
+ local gpg_key_file
++ sed -nr -e '/\[centos-base-el8stream]/{
:loop;
s#^gpgkey\s*=\s*file://(.*)$#\1#p;
n;
/^\[.*\]$/q ;
b loop
}' jenkins/data/mirrors-reposync.conf
+ gpg_key_file=/usr/share/distribution-gpg-keys/centos/RPM-GPG-KEY-CentOS-Official
+ [[ -n /usr/share/distribution-gpg-keys/centos/RPM-GPG-KEY-CentOS-Official ]]
+ [[ -r /usr/share/distribution-gpg-keys/centos/RPM-GPG-KEY-CentOS-Official ]]
+ sudo /usr/bin/rpmkeys --import /usr/share/distribution-gpg-keys/centos/RPM-GPG-KEY-CentOS-Official
+ echo 'Resyncing repo: centos-base-el8stream'
Resyncing repo: centos-base-el8stream
+ perform_yum_sync centos-base-el8stream x86_64 jenkins/data/mirrors-reposync.conf
+ local repo_name=centos-base-el8stream
+ local repo_archs=x86_64
+ local reposync_conf=jenkins/data/mirrors-reposync.conf
+ local repo_mp
+ local repo_comps
+ local sync_newest_only
+ repo_mp=/var/www/html/repos/yum/centos-base-el8stream/base
++ get_modulesmd_path centos-base-el8stream
++ local repo_name=centos-base-el8stream
+++ sed -e 's/xmlns=".*"//g' /home/jenkins/mirrors_cache/centos-base-el8stream/repomd.xml
+++ xmllint --xpath 'string(/repomd/data[@type="modules"]/location/@href)' -
+++ grep -o '[[:alnum:]_.-]*$'
++ filename=
++ [[ -z '' ]]
++ false
+ '[' -z '' ']'
+ sync_newest_only=--newest-only
++ IFS=,
++ echo x86_64
+ for arch in '$(IFS=,; echo $repo_archs)'
+ echo 'Syncing yum repo centos-base-el8stream for arch: x86_64'
Syncing yum repo centos-base-el8stream for arch: x86_64
+ run_reposync centos-base-el8stream x86_64 jenkins/data/mirrors-reposync.conf --downloadcomps --gpgcheck --download-metadata --newest-only
+ local repo_name=centos-base-el8stream
+ local repo_arch=x86_64
+ local reposync_conf=jenkins/data/mirrors-reposync.conf
+ extra_args=("${@:4}")
+ local extra_args
+ reposync --config=jenkins/data/mirrors-reposync.conf --repoid=centos-base-el8stream --arch=x86_64 --cachedir=/home/jenkins/mirrors_cache --download_path=/var/www/html/repos/yum/centos-base-el8stream/base --norepopath --downloadcomps --gpgcheck --download-metadata --newest-only
centos-base-el8stream/x86_64/group_xz | 55 kB 00:00
NetworkManager-adsl-1.32.3-0.z FAILED
(1/69): NetworkManager-1.3 0% [ ] 0.0 B/s | 0 B --:-- ETA NetworkManager-bluetooth-1.32. FAILED
(1/69): NetworkManager-1.3 0% [ ] 0.0 B/s | 0 B --:-- ETA (1/69): NetworkManager-1.32.3-0.z.2.20e3975fd2.el8.x86_64. | 2.6 MB 00:00
(2/69): NetworkManager-config-connectivity-redhat-1.32.3-0 | 130 kB 00:00
(3/69): NetworkManager-config-server-1.32.3-0.z.2.20e3975f | 130 kB 00:00
(4/69): NetworkManager-dispatcher-routing-rules-1.32.3-0.z | 131 kB 00:00
(5/69): NetworkManager-libnm-1.32.3-0.z.2.20e3975fd2.el8.i | 1.8 MB 00:00
(6/69): NetworkManager-libnm-1.32.3-0.z.2.20e3975fd2.el8.x | 1.8 MB 00:00
(7/69): NetworkManager-ovs-1.32.3-0.z.2.20e3975fd2.el8.x86 | 163 kB 00:00
(8/69): NetworkManager-ppp-1.32.3-0.z.2.20e3975fd2.el8.x86 | 152 kB 00:00
(9/69): NetworkManager-team-1.32.3-0.z.2.20e3975fd2.el8.x8 | 147 kB 00:00
(10/69): NetworkManager-tui-1.32.3-0.z.2.20e3975fd2.el8.x8 | 334 kB 00:00
(11/69): NetworkManager-wifi-1.32.3-0.z.2.20e3975fd2.el8.x | 188 kB 00:00
(12/69): NetworkManager-wwan-1.32.3-0.z.2.20e3975fd2.el8.x | 174 kB 00:00
(13/69): OpenIPMI-2.0.31-1.el8.i686.rpm | 127 kB 00:00
(14/69): OpenIPMI-2.0.31-1.el8.x86_64.rpm | 126 kB 00:00
(15/69): OpenIPMI-lanserv-2.0.31-1.el8.i686.rpm | 164 kB 00:00
(16/69): OpenIPMI-lanserv-2.0.31-1.el8.x86_64.rpm | 156 kB 00:00
(17/69): OpenIPMI-libs-2.0.31-1.el8.i686.rpm | 551 kB 00:00
(18/69): OpenIPMI-libs-2.0.31-1.el8.x86_64.rpm | 509 kB 00:00
(19/69): OpenIPMI-perl-2.0.31-1.el8.x86_64.rpm | 178 kB 00:00
(20/69): accel-config-3.1-1.el8.i686.rpm | 68 kB 00:00
(21/69): accel-config-3.1-1.el8.x86_64.rpm | 65 kB 00:00
(22/69): accel-config-libs-3.1-1.el8.i686.rpm | 40 kB 00:00
(23/69): accel-config-libs-3.1-1.el8.x86_64.rpm | 39 kB 00:00
(24/69): binutils-2.30-108.el8.x86_64.rpm | 5.8 MB 00:00
(25/69): cockpit-248-1.el8.x86_64.rpm | 77 kB 00:00
(26/69): cockpit-bridge-248-1.el8.x86_64.rpm | 539 kB 00:00
(27/69): cockpit-doc-248-1.el8.noarch.rpm | 184 kB 00:00
(28/69): cockpit-system-248-1.el8.noarch.rpm | 3.2 MB 00:00
(29/69): cockpit-ws-248-1.el8.x86_64.rpm | 1.3 MB 00:00
(30/69): coreutils-8.30-12.el8.x86_64.rpm | 1.2 MB 00:00
(31/69): coreutils-common-8.30-12.el8.x86_64.rpm | 2.0 MB 00:00
(32/69): coreutils-single-8.30-12.el8.x86_64.rpm | 629 kB 00:00
(33/69): device-mapper-multipath-0.8.4-14.el8.x86_64.rpm | 197 kB 00:00
(34/69): device-mapper-multipath-libs-0.8.4-14.el8.i686.rp | 338 kB 00:00
(35/69): device-mapper-multipath-libs-0.8.4-14.el8.x86_64. | 322 kB 00:00
(36/69): freeipmi-1.6.8-1.el8.i686.rpm | 2.0 MB 00:00
(37/69): freeipmi-1.6.8-1.el8.x86_64.rpm | 2.1 MB 00:00
(38/69): freeipmi-bmc-watchdog-1.6.8-1.el8.x86_64.rpm | 84 kB 00:00
(39/69): freeipmi-ipmidetectd-1.6.8-1.el8.x86_64.rpm | 59 kB 00:00
(40/69): freeipmi-ipmiseld-1.6.8-1.el8.x86_64.rpm | 125 kB 00:00
(41/69): glib2-2.56.4-156.el8.i686.rpm | 2.6 MB 00:00
(42/69): glib2-2.56.4-156.el8.x86_64.rpm | 2.5 MB 00:00
(43/69): glib2-devel-2.56.4-156.el8.i686.rpm | 426 kB 00:00
(44/69): glib2-devel-2.56.4-156.el8.x86_64.rpm | 424 kB 00:00
(45/69): glib2-fam-2.56.4-156.el8.x86_64.rpm | 14 kB 00:00
(46/69): glib2-tests-2.56.4-156.el8.x86_64.rpm | 1.8 MB 00:00
(47/69): kernel-rt-core-4. 49% [=======- ] 0.0 B/s | 50 MB --:-- ETA (47/69): kernel-rt-core-4. 61% [=========- ] 2.4 MB/s | 62 MB 00:15 ETA (47/69): kernel-rt-core-4.18.0-315.rt7.96.el8.x86_64.rpm | 36 MB 00:01
(48/69): kpartx-0.8.4-14.el8.x86_64.rpm | 112 kB 00:00
(49/69): libdmmp-0.8.4-14.el8.i686.rpm | 106 kB 00:00
(50/69): libdmmp-0.8.4-14.el8.x86_64.rpm | 106 kB 00:00
(51/69): nettle-3.4.1-7.el8.i686.rpm | 321 kB 00:00
(52/69): nettle-3.4.1-7.el8.x86_64.rpm | 301 kB 00:00
(53/69): parted-3.2-39.el8.i686.rpm | 591 kB 00:00
(54/69): parted-3.2-39.el8.x86_64.rpm | 555 kB 00:00
(55/69): platform-python-3.6.8-39.el8.x86_64.rpm | 84 kB 00:00
(56/69): python3-libs-3.6.8-39.el8.i686.rpm | 7.9 MB 00:00
(57/69): python3-libs-3.6.8-39.el8.x86_64.rpm | 7.8 MB 00:00
(58/69): python3-openipmi-2.0.31-1.el8.x86_64.rpm | 165 kB 00:00
(59/69): python3-test-3.6.8-39.el8.x86_64.rpm | 8.6 MB 00:00
(60/69): smc-tools-1.6.0-1.git393dd23.el8.i686.rpm | 81 kB 00:00
(61/69): smc-tools-1.6.0-1.git393dd23.el8.x86_64.rpm | 80 kB 00:00
(62/69): tuned-2.16.0-0.1.rc1.el8.noarch.rpm | 311 kB 00:00
(63/69): tuned-profiles-atomic-2.16.0-0.1.rc1.el8.noarch.r | 37 kB 00:00
(64/69): tuned-profiles-compat-2.16.0-0.1.rc1.el8.noarch.r | 40 kB 00:00
(65/69): tuned-profiles-cpu-partitioning-2.16.0-0.1.rc1.el | 40 kB 00:00
(66/69): tuned-profiles-mssql-2.16.0-0.1.rc1.el8.noarch.rp | 37 kB 00:00
(67/69): tuned-profiles-oracle-2.16.0-0.1.rc1.el8.noarch.r | 37 kB 00:00
1:NetworkManager-adsl-1.32.3-0.z.2.20e3975fd2.el8.x86_64: [Errno 256] No more mirrors to try.
1:NetworkManager-bluetooth-1.32.3-0.z.2.20e3975fd2.el8.x86_64: [Errno 256] No more mirrors to try.
Removing NetworkManager-adsl-1.32.3-0.z.2.20e3975fd2.el8.x86_64.rpm due to failed signature check.
Removing NetworkManager-bluetooth-1.32.3-0.z.2.20e3975fd2.el8.x86_64.rpm due to failed signature check.
Build step 'Execute shell' marked build as failure
3 years, 6 months
[oVirt Jenkins] ovirt-system-tests_basic-suite-4.3_nightly - Build
# 666 - Failure!
by jenkins@jenkins.phx.ovirt.org
Project: https://jenkins.ovirt.org/job/ovirt-system-tests_basic-suite-4.3_nightly/
Build: https://jenkins.ovirt.org/job/ovirt-system-tests_basic-suite-4.3_nightly/...
Build Number: 666
Build Status: Failure
Triggered By: Started by timer
-------------------------------------
Changes Since Last Success:
-------------------------------------
Changes for Build #666
[parthdhanjal] Updating max click version
-----------------
Failed Tests:
-----------------
1 tests failed.
FAILED: 004_basic_sanity.run_vms
Error Message:
Fault reason is "Operation Failed". Fault detail is "[Network error during communication with the Host.]". HTTP response code is 400.
Stack Trace:
Traceback (most recent call last):
File "/usr/lib64/python2.7/unittest/case.py", line 369, in run
testMethod()
File "/usr/lib/python2.7/site-packages/nose/case.py", line 197, in runTest
self.test(*self.arg)
File "/usr/lib/python2.7/site-packages/ovirtlago/testlib.py", line 142, in wrapped_test
test()
File "/usr/lib/python2.7/site-packages/ovirtlago/testlib.py", line 60, in wrapper
return func(get_test_prefix(), *args, **kwargs)
File "/home/jenkins/agent/workspace/ovirt-system-tests_basic-suite-4.3_nightly/ovirt-system-tests/basic-suite-4.3/test-scenarios/004_basic_sanity.py", line 694, in run_vms
vm0_service.start(use_cloud_init=True, vm=vm_params)
File "/usr/lib64/python2.7/site-packages/ovirtsdk4/services.py", line 32520, in start
return self._internal_action(action, 'start', None, headers, query, wait)
File "/usr/lib64/python2.7/site-packages/ovirtsdk4/service.py", line 299, in _internal_action
return future.wait() if wait else future
File "/usr/lib64/python2.7/site-packages/ovirtsdk4/service.py", line 55, in wait
return self._code(response)
File "/usr/lib64/python2.7/site-packages/ovirtsdk4/service.py", line 296, in callback
self._check_fault(response)
File "/usr/lib64/python2.7/site-packages/ovirtsdk4/service.py", line 134, in _check_fault
self._raise_error(response, body.fault)
File "/usr/lib64/python2.7/site-packages/ovirtsdk4/service.py", line 118, in _raise_error
raise error
Error: Fault reason is "Operation Failed". Fault detail is "[Network error during communication with the Host.]". HTTP response code is 400.
3 years, 6 months
[oVirt Jenkins] ovirt-system-tests_basic-suite-master_nightly -
Build # 1326 - Failure!
by jenkins@jenkins.phx.ovirt.org
Project: https://jenkins.ovirt.org/job/ovirt-system-tests_basic-suite-master_nightly/
Build: https://jenkins.ovirt.org/job/ovirt-system-tests_basic-suite-master_night...
Build Number: 1326
Build Status: Failure
Triggered By: Started by timer
-------------------------------------
Changes Since Last Success:
-------------------------------------
Changes for Build #1326
[Michal Skrivanek] ci: create log of stdout so that it gets archived
[Sandro Bonazzola] mock_configs: drop failovermethod
-----------------
Failed Tests:
-----------------
1 tests failed.
FAILED: basic-suite-master.test-scenarios.test_004_basic_sanity.test_cold_incremental_backup_vm2
Error Message:
ovirtsdk4.Error: Fault reason is "Operation Failed". Fault detail is "[Cannot backup VM: Disk is locked. Please try again later.]". HTTP response code is 409.
Stack Trace:
engine_api = <ovirtsdk4.Connection object at 0x7fda4189fcf8>
get_vm_service_for_vm = <function get_vm_service_for_vm.<locals>.service_for at 0x7fda3b675488>
@order_by(_TEST_LIST)
def test_cold_incremental_backup_vm2(engine_api, get_vm_service_for_vm):
_verify_vm_state(engine_api.system_service(), VM2_NAME, types.VmStatus.DOWN)
vm2_backups_service = get_vm_service_for_vm(VM2_NAME).backups_service()
backup.perform_incremental_vm_backup(
> engine_api, vm2_backups_service, DISK2_NAME, "cold_vm_backup")
basic-suite-master/test-scenarios/test_004_basic_sanity.py:1045:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
ost_utils/ost_utils/storage_utils/backup.py:75: in perform_incremental_vm_backup
correlation_id="incremental_" + correlation_id)
ost_utils/ost_utils/storage_utils/backup.py:34: in perform_vm_backup
), query={'correlation_id': correlation_id}
/usr/lib64/python3.6/site-packages/ovirtsdk4/services.py:34139: in add
return self._internal_add(backup, headers, query, wait)
/usr/lib64/python3.6/site-packages/ovirtsdk4/service.py:232: in _internal_add
return future.wait() if wait else future
/usr/lib64/python3.6/site-packages/ovirtsdk4/service.py:55: in wait
return self._code(response)
/usr/lib64/python3.6/site-packages/ovirtsdk4/service.py:229: in callback
self._check_fault(response)
/usr/lib64/python3.6/site-packages/ovirtsdk4/service.py:132: in _check_fault
self._raise_error(response, body)
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
response = <ovirtsdk4.http.Response object at 0x7fda3b5d1da0>
detail = <ovirtsdk4.types.Fault object at 0x7fda3b5d1d30>
@staticmethod
def _raise_error(response, detail=None):
"""
Creates and raises an error containing the details of the given HTTP
response and fault.
This method is intended for internal use by other components of the
SDK. Refrain from using it directly, as backwards compatibility isn't
guaranteed.
"""
fault = detail if isinstance(detail, types.Fault) else None
msg = ''
if fault:
if fault.reason:
if msg:
msg += ' '
msg = msg + 'Fault reason is "%s".' % fault.reason
if fault.detail:
if msg:
msg += ' '
msg = msg + 'Fault detail is "%s".' % fault.detail
if response:
if response.code:
if msg:
msg += ' '
msg = msg + 'HTTP response code is %s.' % response.code
if response.message:
if msg:
msg += ' '
msg = msg + 'HTTP response message is "%s".' % response.message
if isinstance(detail, six.string_types):
if msg:
msg += ' '
msg = msg + detail + '.'
class_ = Error
if response is not None:
if response.code in [401, 403]:
class_ = AuthError
elif response.code == 404:
class_ = NotFoundError
error = class_(msg)
error.code = response.code if response else None
error.fault = fault
> raise error
E ovirtsdk4.Error: Fault reason is "Operation Failed". Fault detail is "[Cannot backup VM: Disk is locked. Please try again later.]". HTTP response code is 409.
/usr/lib64/python3.6/site-packages/ovirtsdk4/service.py:118: Error
3 years, 6 months
[oVirt Jenkins] ovirt-system-tests_basic-suite-master_nightly -
Build # 1324 - Failure!
by jenkins@jenkins.phx.ovirt.org
Project: https://jenkins.ovirt.org/job/ovirt-system-tests_basic-suite-master_nightly/
Build: https://jenkins.ovirt.org/job/ovirt-system-tests_basic-suite-master_night...
Build Number: 1324
Build Status: Failure
Triggered By: Started by timer
-------------------------------------
Changes Since Last Success:
-------------------------------------
Changes for Build #1324
[Michal Skrivanek] ci: create log of stdout so that it gets archived
-----------------
Failed Tests:
-----------------
1 tests failed.
FAILED: basic-suite-master.test-scenarios.test_004_basic_sanity.test_cold_incremental_backup_vm2
Error Message:
ovirtsdk4.Error: Fault reason is "Operation Failed". Fault detail is "[Cannot backup VM. The VM is during a backup operation.]". HTTP response code is 409.
Stack Trace:
engine_api = <ovirtsdk4.Connection object at 0x7fe9f85f5978>
get_vm_service_for_vm = <function get_vm_service_for_vm.<locals>.service_for at 0x7fe9f62fabf8>
@order_by(_TEST_LIST)
def test_cold_incremental_backup_vm2(engine_api, get_vm_service_for_vm):
_verify_vm_state(engine_api.system_service(), VM2_NAME, types.VmStatus.DOWN)
vm2_backups_service = get_vm_service_for_vm(VM2_NAME).backups_service()
backup.perform_incremental_vm_backup(
> engine_api, vm2_backups_service, DISK2_NAME, "cold_vm_backup")
basic-suite-master/test-scenarios/test_004_basic_sanity.py:1045:
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
ost_utils/ost_utils/storage_utils/backup.py:75: in perform_incremental_vm_backup
correlation_id="incremental_" + correlation_id)
ost_utils/ost_utils/storage_utils/backup.py:34: in perform_vm_backup
), query={'correlation_id': correlation_id}
/usr/lib64/python3.6/site-packages/ovirtsdk4/services.py:34139: in add
return self._internal_add(backup, headers, query, wait)
/usr/lib64/python3.6/site-packages/ovirtsdk4/service.py:232: in _internal_add
return future.wait() if wait else future
/usr/lib64/python3.6/site-packages/ovirtsdk4/service.py:55: in wait
return self._code(response)
/usr/lib64/python3.6/site-packages/ovirtsdk4/service.py:229: in callback
self._check_fault(response)
/usr/lib64/python3.6/site-packages/ovirtsdk4/service.py:132: in _check_fault
self._raise_error(response, body)
_ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _
response = <ovirtsdk4.http.Response object at 0x7fe9f62e0dd8>
detail = <ovirtsdk4.types.Fault object at 0x7fe9f61fa550>
@staticmethod
def _raise_error(response, detail=None):
"""
Creates and raises an error containing the details of the given HTTP
response and fault.
This method is intended for internal use by other components of the
SDK. Refrain from using it directly, as backwards compatibility isn't
guaranteed.
"""
fault = detail if isinstance(detail, types.Fault) else None
msg = ''
if fault:
if fault.reason:
if msg:
msg += ' '
msg = msg + 'Fault reason is "%s".' % fault.reason
if fault.detail:
if msg:
msg += ' '
msg = msg + 'Fault detail is "%s".' % fault.detail
if response:
if response.code:
if msg:
msg += ' '
msg = msg + 'HTTP response code is %s.' % response.code
if response.message:
if msg:
msg += ' '
msg = msg + 'HTTP response message is "%s".' % response.message
if isinstance(detail, six.string_types):
if msg:
msg += ' '
msg = msg + detail + '.'
class_ = Error
if response is not None:
if response.code in [401, 403]:
class_ = AuthError
elif response.code == 404:
class_ = NotFoundError
error = class_(msg)
error.code = response.code if response else None
error.fault = fault
> raise error
E ovirtsdk4.Error: Fault reason is "Operation Failed". Fault detail is "[Cannot backup VM. The VM is during a backup operation.]". HTTP response code is 409.
/usr/lib64/python3.6/site-packages/ovirtsdk4/service.py:118: Error
3 years, 6 months