and here is vdsm:The last failure seems to be storage related.@Nir please take a look.Here is engine side error:
2016-11-20 05:54:59,605 DEBUG [org.ovirt.engine.core.vdsbroker.vdsbroker. ConnectStoragePoolVDSCommand] (default task-5) [59fc0074] Exception: org.ovirt.engine.core. vdsbroker.irsbroker. IRSNoMasterDomainException: IRSGenericException: IRSErrorException: IRSNoMasterDomainException: Cannot find master domain: u'spUUID=1ca141f1-b64d-4a52- 8861-05c7de2a72b2, msdUUID=7d4bf750-4fb8-463f- bbb0-92156c47306e'
jsonrpc.Executor/5::ERROR::2016-11-20 05:54:56,331::multipath::95:: Storage.Multipath::(resize_ devices) Could not resize device 360014052749733c7b8248628637b9 90f
Traceback (most recent call last):
File "/usr/share/vdsm/storage/multipath.py", line 93, in resize_devices
_resize_if_needed(guid)
File "/usr/share/vdsm/storage/multipath.py", line 101, in _resize_if_needed
for slave in devicemapper.getSlaves(name)]
File "/usr/share/vdsm/storage/multipath.py", line 158, in getDeviceSize
bs, phyBs = getDeviceBlockSizes(devName)
File "/usr/share/vdsm/storage/multipath.py", line 150, in getDeviceBlockSizes
"queue", "logical_block_size")).read())
IOError: [Errno 2] No such file or directory: '/sys/block/sdb/queue/logical_block_size'
On Sun, Nov 20, 2016 at 12:50 PM, Eyal Edri <eedri@redhat.com> wrote:On Sun, Nov 20, 2016 at 1:42 PM, Yaniv Kaul <ykaul@redhat.com> wrote:On Sun, Nov 20, 2016 at 1:30 PM, Yaniv Kaul <ykaul@redhat.com> wrote:On Sun, Nov 20, 2016 at 1:18 PM, Eyal Edri <eedri@redhat.com> wrote:the test fails to run VM because no hosts are in UP state(?) [1], not sure it is related to the triggering patch[2]status: 400reason: Bad Requestdetail: There are no hosts to use. Check that the cluster contains at least one host in Up state.Thoughts? Shouldn't we fail the test earlier we hosts are not UP?Yes. It's more likely that we are picking the wrong host or so, but who knows - where are the engine and VDSM logs?A simple grep on the engine.log[1] finds serveral unrelated issues I'm not sure are reported, it's despairing to even begin...That being said, I don't see the issue there. We may need better logging on the API level, to see what is being sent. Is it consistent?Just failed now the first time, I didn't see it before.Y.Y.On Sun, Nov 20, 2016 at 1:00 PM, <jenkins@jenkins.phx.ovirt.org> wrote:Build: http://jenkins.ovirt.org/job/test-repo_ovirt_experimental_4. ,0/3015/
Build Number: 3015,
Build Status: FAILURE
_______________________________________________
Infra mailing list
Infra@ovirt.org
http://lists.ovirt.org/mailman/listinfo/infra
--Eyal Edri
Associate ManagerRHV DevOps
EMEA ENG Virtualization R&D
Red Hat Israel
phone: +972-9-7692018
irc: eedri (on #tlv #rhev-dev #rhev-integ)--Eyal Edri
Associate ManagerRHV DevOps
EMEA ENG Virtualization R&D
Red Hat Israel
phone: +972-9-7692018
irc: eedri (on #tlv #rhev-dev #rhev-integ)