Execution is done, 59/65 cases passed. Latest 4.2.4 execution ended with
100% so failures were caused probably due to the changes done in the patch.
Failures are mainly on preview snapshots.
Execution info provided to Martin separately.
On Wed, May 30, 2018 at 5:44 PM, Elad Ben Aharon <ebenahar(a)redhat.com>
wrote:
Triggered a sanity automation execution using [1], which covers all
the
requested areas, on iSCSI, NFS and Gluster.
I'll update with the results.
[1]
*https://gerrit.ovirt.org/#/c/90906/ <
https://gerrit.ovirt.org/#/c/90906/>*
vdsm-4.20.28-6.gitc23aef6.el7.x86_64
On Tue, May 29, 2018 at 4:26 PM, Martin Polednik <mpolednik(a)redhat.com>
wrote:
> On 29/05/18 15:30 +0300, Elad Ben Aharon wrote:
>
>> Hi Martin,
>>
>> Can you please create a cerry pick patch that is based on 4.2?
>>
>
> See
https://gerrit.ovirt.org/#/c/90906/. The CI failure isn unrelated
> (storage needs real env).
>
> mpolednik
>
>
>
>> Thanks
>>
>> On Tue, May 29, 2018 at 1:34 PM, Dan Kenigsberg <danken(a)redhat.com>
>> wrote:
>>
>> On Tue, May 29, 2018 at 1:21 PM, Elad Ben Aharon <ebenahar(a)redhat.com>
>>> wrote:
>>> > Hi Dan,
>>> >
>>> > In the last execution, the success rate was very low due to a large
>>> number
>>> > of failures on start VM caused, according to Michal, by the
>>> > vdsm-hook-allocate_net that was installed on the host.
>>> >
>>> > This is the latest status here, would you like me to re-execute?
>>>
>>> yes, of course. but you should rebase Polednik's code on top of
>>> *current* ovirt-4.2.3 branch.
>>>
>>> > If so, with
>>> > or W/O vdsm-hook-allocate_net installed?
>>>
>>> There was NO reason to have that installed. Please keep it (and any
>>> other needless code) out of the test environment.
>>>
>>> >
>>> > On Tue, May 29, 2018 at 1:14 PM, Dan Kenigsberg
<danken(a)redhat.com>
>>> wrote:
>>> >>
>>> >> On Mon, May 7, 2018 at 3:53 PM, Michal Skrivanek
>>> >> <michal.skrivanek(a)redhat.com> wrote:
>>> >> > Hi Elad,
>>> >> > why did you install vdsm-hook-allocate_net?
>>> >> >
>>> >> > adding Dan as I think the hook is not supposed to fail this
badly
>>> in
>>> any
>>> >> > case
>>> >>
>>> >> yep, this looks bad and deserves a little bug report. Installing
this
>>> >> little hook should not block vm startup.
>>> >>
>>> >> But more importantly - what is the conclusion of this thread? Do we
>>> >> have a green light from QE to take this in?
>>> >>
>>> >>
>>> >> >
>>> >> > Thanks,
>>> >> > michal
>>> >> >
>>> >> > On 5 May 2018, at 19:22, Elad Ben Aharon
<ebenahar(a)redhat.com>
>>> wrote:
>>> >> >
>>> >> > Start VM fails on:
>>> >> >
>>> >> > 2018-05-05 17:53:27,399+0300 INFO (vm/e6ce66ce) [virt.vm]
>>> >> > (vmId='e6ce66ce-852f-48c5-9997-5d2959432a27') drive
'vda' path:
>>> >> >
>>> >> > 'dev=/rhev/data-center/mnt/blockSD/db5a6696-d907-4938-
>>> 9a78-bdd13a843c62/images/6cdabfe5-
>>> >> >
d1ca-40af-ae63-9834f235d1c8/7ef97445-30e6-4435-8425-f35a01928211'
>>> ->
>>> >> >
>>> >> > u'*dev=/rhev/data-center/mnt/blockSD/db5a6696-d907-4938-
>>> 9a78-bdd13a843c62/images/6cdabfe5-d1ca-40af-ae63-
>>> 9834f235d1c8/7ef97445-30e6-4435-8425-
>>> >> > f35a01928211' (storagexml:334)
>>> >> > 2018-05-05 17:53:27,888+0300 INFO (jsonrpc/1) [vdsm.api]
START
>>> >> >
getSpmStatus(spUUID='940fe6f3-b0c6-4d0c-a921-198e7819c1cc',
>>> >> > options=None)
>>> >> > from=::ffff:10.35.161.127,53512,
>>> >> > task_id=c70ace39-dbfe-4f5c-ae49-a1e3a82c
>>> >> > 2758 (api:46)
>>> >> > 2018-05-05 17:53:27,909+0300 INFO (vm/e6ce66ce) [root]
>>> >> > /usr/libexec/vdsm/hooks/before_device_create/10_allocate_net:
rc=2
>>> >> > err=vm
>>> >> > net allocation hook: [unexpected error]: Traceback (most
recent
>>> call
>>> >> > last):
>>> >> > File
"/usr/libexec/vdsm/hooks/before_device_create/10_allocate_ne
>>> t",
>>> >> > line
>>> >> > 105, in <module>
>>> >> > main()
>>> >> > File
"/usr/libexec/vdsm/hooks/before_device_create/10_allocate_ne
>>> t",
>>> >> > line
>>> >> > 93, in main
>>> >> > allocate_random_network(device_xml)
>>> >> > File
"/usr/libexec/vdsm/hooks/before_device_create/10_allocate_ne
>>> t",
>>> >> > line
>>> >> > 62, in allocate_random_network
>>> >> > net = _get_random_network()
>>> >> > File
"/usr/libexec/vdsm/hooks/before_device_create/10_allocate_ne
>>> t",
>>> >> > line
>>> >> > 50, in _get_random_network
>>> >> > available_nets = _parse_nets()
>>> >> > File
"/usr/libexec/vdsm/hooks/before_device_create/10_allocate_ne
>>> t",
>>> >> > line
>>> >> > 46, in _parse_nets
>>> >> > return [net for net in os.environ[AVAIL_NETS_KEY].split()]
>>> >> > File "/usr/lib64/python2.7/UserDict.py", line 23, in
__getitem__
>>> >> > raise KeyError(key)
>>> >> > KeyError: 'equivnets'
>>> >> >
>>> >> >
>>> >> > (hooks:110)
>>> >> > 2018-05-05 17:53:27,915+0300 ERROR (vm/e6ce66ce) [virt.vm]
>>> >> > (vmId='e6ce66ce-852f-48c5-9997-5d2959432a27') The vm
start process
>>> >> > failed
>>> >> > (vm:943)
>>> >> > Traceback (most recent call last):
>>> >> > File
"/usr/lib/python2.7/site-packages/vdsm/virt/vm.py", line
>>> 872,
>>> in
>>> >> > _startUnderlyingVm
>>> >> > self._run()
>>> >> > File
"/usr/lib/python2.7/site-packages/vdsm/virt/vm.py", line
>>> 2861,
>>> in
>>> >> > _run
>>> >> > domxml = hooks.before_vm_start(self._buildDomainXML(),
>>> >> > File
"/usr/lib/python2.7/site-packages/vdsm/virt/vm.py", line
>>> 2254,
>>> in
>>> >> > _buildDomainXML
>>> >> > dom, self.id, self._custom['custom'])
>>> >> > File "/usr/lib/python2.7/site-packages/vdsm/virt/domxml_
>>> preprocess.py",
>>> >> > line 240, in replace_device_xml_with_hooks_xml
>>> >> > dev_custom)
>>> >> > File
"/usr/lib/python2.7/site-packages/vdsm/common/hooks.py",
>>> line
>>> 134,
>>> >> > in
>>> >> > before_device_create
>>> >> > params=customProperties)
>>> >> > File
"/usr/lib/python2.7/site-packages/vdsm/common/hooks.py",
>>> line
>>> 120,
>>> >> > in
>>> >> > _runHooksDir
>>> >> > raise exception.HookError(err)
>>> >> > HookError: Hook Error: ('vm net allocation hook:
[unexpected
>>> error]:
>>> >> > Traceback (most recent call last):\n File
>>> >> >
"/usr/libexec/vdsm/hooks/before_device_create/10_allocate_net",
>>> line
>>> >> > 105, in
>>> >> > <module>\n main()\n
>>> >> > File
"/usr/libexec/vdsm/hooks/before_device_create/10_allocate_ne
>>> t",
>>> >> > line
>>> >> > 93, in main\n allocate_random_network(device_xml)\n File
>>> >> >
"/usr/libexec/vdsm/hooks/before_device_create/10_allocate_net",
>>> line
>>> 62,
>>> >> > i
>>> >> > n allocate_random_network\n net = _get_random_network()\n
File
>>> >> >
"/usr/libexec/vdsm/hooks/before_device_create/10_allocate_net",
>>> line
>>> 50,
>>> >> > in
>>> >> > _get_random_network\n available_nets = _parse_nets()\n
File
>>> "/us
>>> >> >
r/libexec/vdsm/hooks/before_device_create/10_allocate_net", line
>>> 46,
>>> in
>>> >> > _parse_nets\n return [net for net in
>>> >> > os.environ[AVAIL_NETS_KEY].split()]\n File
>>> >> > "/usr/lib64/python2.7/UserDict.py", line 23, in
__getit
>>> >> > em__\n raise KeyError(key)\nKeyError:
\'equivnets\'\n\n\n',)
>>> >> >
>>> >> >
>>> >> >
>>> >> > Hence, the success rate was 28% against 100% running with d/s
>>> (d/s).
>>> If
>>> >> > needed, I'll compare against the latest master, but I think
you get
>>> the
>>> >> > picture with d/s.
>>> >> >
>>> >> > vdsm-4.20.27-3.gitfee7810.el7.centos.x86_64
>>> >> > libvirt-3.9.0-14.el7_5.3.x86_64
>>> >> > qemu-kvm-rhev-2.10.0-21.el7_5.2.x86_64
>>> >> > kernel 3.10.0-862.el7.x86_64
>>> >> > rhel7.5
>>> >> >
>>> >> >
>>> >> > Logs attached
>>> >> >
>>> >> > On Sat, May 5, 2018 at 1:26 PM, Elad Ben Aharon <
>>> ebenahar(a)redhat.com>
>>> >> > wrote:
>>> >> >>
>>> >> >> nvm, found gluster 3.12 repo, managed to install vdsm
>>> >> >>
>>> >> >> On Sat, May 5, 2018 at 1:12 PM, Elad Ben Aharon <
>>> ebenahar(a)redhat.com
>>> >
>>> >> >> wrote:
>>> >> >>>
>>> >> >>> No, vdsm requires it:
>>> >> >>>
>>> >> >>> Error: Package:
vdsm-4.20.27-3.gitfee7810.el7.centos.x86_64
>>> >> >>> (/vdsm-4.20.27-3.gitfee7810.el7.centos.x86_64)
>>> >> >>> Requires: glusterfs-fuse >= 3.12
>>> >> >>> Installed:
glusterfs-fuse-3.8.4-54.8.el7.x86_64
>>> (@rhv-4.2.3)
>>> >> >>>
>>> >> >>> Therefore, vdsm package installation is skipped upon
force
>>> install.
>>> >> >>>
>>> >> >>> On Sat, May 5, 2018 at 11:42 AM, Michal Skrivanek
>>> >> >>> <michal.skrivanek(a)redhat.com> wrote:
>>> >> >>>>
>>> >> >>>>
>>> >> >>>>
>>> >> >>>> On 5 May 2018, at 00:38, Elad Ben Aharon
<ebenahar(a)redhat.com>
>>> wrote:
>>> >> >>>>
>>> >> >>>> Hi guys,
>>> >> >>>>
>>> >> >>>> The vdsm build from the patch requires
glusterfs-fuse > 3.12.
>>> This
>>> is
>>> >> >>>> while the latest 4.2.3-5 d/s build requires 3.8.4
>>> (3.4.0.59rhs-1.el7)
>>> >> >>>>
>>> >> >>>>
>>> >> >>>> because it is still oVirt, not a downstream build.
We can’t
>>> really
>>> do
>>> >> >>>> downstream builds with unmerged changes:/
>>> >> >>>>
>>> >> >>>> Trying to get this gluster-fuse build, so far no
luck.
>>> >> >>>> Is this requirement intentional?
>>> >> >>>>
>>> >> >>>>
>>> >> >>>> it should work regardless, I guess you can force
install it
>>> without
>>> >> >>>> the
>>> >> >>>> dependency
>>> >> >>>>
>>> >> >>>>
>>> >> >>>> On Fri, May 4, 2018 at 2:38 PM, Michal Skrivanek
>>> >> >>>> <michal.skrivanek(a)redhat.com> wrote:
>>> >> >>>>>
>>> >> >>>>> Hi Elad,
>>> >> >>>>> to make it easier to compare, Martin backported
the change to
>>> 4.2
>>> so
>>> >> >>>>> it
>>> >> >>>>> is actually comparable with a run without that
patch. Would you
>>> >> >>>>> please try
>>> >> >>>>> that out?
>>> >> >>>>> It would be best to have 4.2 upstream and
this[1] run to really
>>> >> >>>>> minimize the noise.
>>> >> >>>>>
>>> >> >>>>> Thanks,
>>> >> >>>>> michal
>>> >> >>>>>
>>> >> >>>>> [1]
>>> >> >>>>>
>>> >> >>>>>
http://jenkins.ovirt.org/job/vdsm_4.2_build-artifacts-on-
>>> demand-el7-x86_64/28/
>>> >> >>>>>
>>> >> >>>>> On 27 Apr 2018, at 09:23, Martin Polednik <
>>> mpolednik(a)redhat.com>
>>> >> >>>>> wrote:
>>> >> >>>>>
>>> >> >>>>> On 24/04/18 00:37 +0300, Elad Ben Aharon
wrote:
>>> >> >>>>>
>>> >> >>>>> I will update with the results of the next
tier1 execution on
>>> latest
>>> >> >>>>> 4.2.3
>>> >> >>>>>
>>> >> >>>>>
>>> >> >>>>> That isn't master but old branch though.
Could you run it
>>> against
>>> >> >>>>> *current* VDSM master?
>>> >> >>>>>
>>> >> >>>>> On Mon, Apr 23, 2018 at 3:56 PM, Martin
Polednik
>>> >> >>>>> <mpolednik(a)redhat.com>
>>> >> >>>>> wrote:
>>> >> >>>>>
>>> >> >>>>> On 23/04/18 01:23 +0300, Elad Ben Aharon
wrote:
>>> >> >>>>>
>>> >> >>>>> Hi, I've triggered another execution [1]
due to some issues I
>>> saw
>>> in
>>> >> >>>>> the
>>> >> >>>>> first which are not related to the patch.
>>> >> >>>>>
>>> >> >>>>> The success rate is 78% which is low comparing
to tier1
>>> executions
>>> >> >>>>> with
>>> >> >>>>> code from downstream builds (95-100% success
rates) [2].
>>> >> >>>>>
>>> >> >>>>>
>>> >> >>>>> Could you run the current master (without the
dynamic_ownership
>>> >> >>>>> patch)
>>> >> >>>>> so that we have viable comparision?
>>> >> >>>>>
>>> >> >>>>> From what I could see so far, there is an issue
with move and
>>> copy
>>> >> >>>>>
>>> >> >>>>> operations to and from Gluster domains. For
example [3].
>>> >> >>>>>
>>> >> >>>>> The logs are attached.
>>> >> >>>>>
>>> >> >>>>>
>>> >> >>>>> [1]
>>> >> >>>>>
*https://rhv-jenkins.rhev-ci-vms.eng.rdu2.redhat.com/job/rhv
>>> >> >>>>>
-4.2-ge-runner-tier1-after-upgrade/7/testReport/
>>> >> >>>>>
<
https://rhv-jenkins.rhev-ci-vms.eng.rdu2.redhat.com/job/rhv
>>> >> >>>>>
-4.2-ge-runner-tier1-after-upgrade/7/testReport/>*
>>> >> >>>>>
>>> >> >>>>>
>>> >> >>>>>
>>> >> >>>>> [2]
>>> >> >>>>>
https://rhv-jenkins.rhev-ci-vms.eng.rdu2.redhat.com/job/
>>> >> >>>>>
>>> >> >>>>> rhv-4.2-ge-runner-tier1-after-upgrade/7/
>>> >> >>>>>
>>> >> >>>>>
>>> >> >>>>>
>>> >> >>>>> [3]
>>> >> >>>>> 2018-04-22 13:06:28,316+0300 INFO (jsonrpc/7)
[vdsm.api]
>>> FINISH
>>> >> >>>>> deleteImage error=Image does not exist in
domain:
>>> >> >>>>>
'image=cabb8846-7a4b-4244-9835-5f603e682f33,
>>> >> >>>>>
domain=e5fd29c8-52ba-467e-be09-ca40ff054dd4'
>>> >> >>>>> from=:
>>> >> >>>>> :ffff:10.35.161.182,40936,
>>> >> >>>>> flow_id=disks_syncAction_ba6b2630-5976-4935,
>>> >> >>>>> task_id=3d5f2a8a-881c-409e-93e9-aaa643c10e42
(api:51)
>>> >> >>>>> 2018-04-22 13:06:28,317+0300 ERROR (jsonrpc/7)
>>> >> >>>>> [storage.TaskManager.Task]
>>> >> >>>>>
(Task='3d5f2a8a-881c-409e-93e9-aaa643c10e42') Unexpected error
>>> >> >>>>> (task:875)
>>> >> >>>>> Traceback (most recent call last):
>>> >> >>>>> File
"/usr/lib/python2.7/site-packages/vdsm/storage/task.py",
>>> line
>>> >> >>>>> 882,
>>> >> >>>>> in
>>> >> >>>>> _run
>>> >> >>>>> return fn(*args, **kargs)
>>> >> >>>>> File "<string>", line 2, in
deleteImage
>>> >> >>>>> File
"/usr/lib/python2.7/site-packages/vdsm/common/api.py",
>>> line
>>> 49,
>>> >> >>>>> in
>>> >> >>>>> method
>>> >> >>>>> ret = func(*args, **kwargs)
>>> >> >>>>> File
"/usr/lib/python2.7/site-packages/vdsm/storage/hsm.py",
>>> line
>>> >> >>>>> 1503,
>>> >> >>>>> in
>>> >> >>>>> deleteImage
>>> >> >>>>> raise se.ImageDoesNotExistInSD(imgUUID,
sdUUID)
>>> >> >>>>> ImageDoesNotExistInSD: Image does not exist in
domain:
>>> >> >>>>>
'image=cabb8846-7a4b-4244-9835-5f603e682f33,
>>> >> >>>>>
domain=e5fd29c8-52ba-467e-be09-ca40ff054dd4'
>>> >> >>>>>
>>> >> >>>>> 2018-04-22 13:06:28,317+0300 INFO (jsonrpc/7)
>>> >> >>>>> [storage.TaskManager.Task]
>>> >> >>>>>
(Task='3d5f2a8a-881c-409e-93e9-aaa643c10e42') aborting: Task
>>> is
>>> >> >>>>> aborted:
>>> >> >>>>> "Image does not exist in domain:
'image=cabb8846-7a4b-4244-9835
>>> -
>>> >> >>>>> 5f603e682f33,
domain=e5fd29c8-52ba-467e-be09-ca40ff054dd4'" -
>>> code
>>> >> >>>>> 268
>>> >> >>>>> (task:1181)
>>> >> >>>>> 2018-04-22 13:06:28,318+0300 ERROR (jsonrpc/7)
>>> [storage.Dispatcher]
>>> >> >>>>> FINISH
>>> >> >>>>> deleteImage error=Image does not exist in
domain:
>>> >> >>>>>
'image=cabb8846-7a4b-4244-9835-5f603e682f33,
>>> >> >>>>> domain=e5fd29c8-52ba-467e-be09
>>> >> >>>>> -ca40ff054d
>>> >> >>>>> d4' (dispatcher:82)
>>> >> >>>>>
>>> >> >>>>>
>>> >> >>>>>
>>> >> >>>>> On Thu, Apr 19, 2018 at 5:34 PM, Elad Ben
Aharon
>>> >> >>>>> <ebenahar(a)redhat.com>
>>> >> >>>>> wrote:
>>> >> >>>>>
>>> >> >>>>> Triggered a sanity tier1 execution [1] using
[2], which covers
>>> all
>>> >> >>>>> the
>>> >> >>>>>
>>> >> >>>>> requested areas, on iSCSI, NFS and Gluster.
>>> >> >>>>> I'll update with the results.
>>> >> >>>>>
>>> >> >>>>> [1]
>>> >> >>>>>
https://rhv-jenkins.rhev-ci-vms.eng.rdu2.redhat.com/view/4.2
>>> >> >>>>> _dev/job/rhv-4.2-ge-flow-storage/1161/
>>> >> >>>>>
>>> >> >>>>> [2]
>>> >> >>>>>
https://gerrit.ovirt.org/#/c/89830/
>>> >> >>>>> vdsm-4.30.0-291.git77aef9a.el7.x86_64
>>> >> >>>>>
>>> >> >>>>>
>>> >> >>>>>
>>> >> >>>>> On Thu, Apr 19, 2018 at 3:07 PM, Martin
Polednik
>>> >> >>>>> <mpolednik(a)redhat.com>
>>> >> >>>>> wrote:
>>> >> >>>>>
>>> >> >>>>> On 19/04/18 14:54 +0300, Elad Ben Aharon
wrote:
>>> >> >>>>>
>>> >> >>>>>
>>> >> >>>>> Hi Martin,
>>> >> >>>>>
>>> >> >>>>>
>>> >> >>>>> I see [1] requires a rebase, can you please
take care?
>>> >> >>>>>
>>> >> >>>>>
>>> >> >>>>> Should be rebased.
>>> >> >>>>>
>>> >> >>>>> At the moment, our automation is stable only on
iSCSI, NFS,
>>> Gluster
>>> >> >>>>> and
>>> >> >>>>>
>>> >> >>>>> FC.
>>> >> >>>>> Ceph is not supported and Cinder will be
stabilized soon,
>>> AFAIR,
>>> >> >>>>> it's
>>> >> >>>>> not
>>> >> >>>>> stable enough at the moment.
>>> >> >>>>>
>>> >> >>>>>
>>> >> >>>>> That is still pretty good.
>>> >> >>>>>
>>> >> >>>>>
>>> >> >>>>> [1]
https://gerrit.ovirt.org/#/c/89830/
>>> >> >>>>>
>>> >> >>>>>
>>> >> >>>>>
>>> >> >>>>> Thanks
>>> >> >>>>>
>>> >> >>>>> On Wed, Apr 18, 2018 at 2:17 PM, Martin
Polednik
>>> >> >>>>> <mpolednik(a)redhat.com
>>> >> >>>>> >
>>> >> >>>>> wrote:
>>> >> >>>>>
>>> >> >>>>> On 18/04/18 11:37 +0300, Elad Ben Aharon
wrote:
>>> >> >>>>>
>>> >> >>>>>
>>> >> >>>>> Hi, sorry if I misunderstood, I waited for more
input regarding
>>> what
>>> >> >>>>>
>>> >> >>>>> areas
>>> >> >>>>> have to be tested here.
>>> >> >>>>>
>>> >> >>>>>
>>> >> >>>>> I'd say that you have quite a bit of
freedom in this regard.
>>> >> >>>>>
>>> >> >>>>> GlusterFS
>>> >> >>>>> should be covered by Dennis, so
iSCSI/NFS/ceph/cinder with some
>>> >> >>>>> suite
>>> >> >>>>> that covers basic operations (start & stop
VM, migrate it),
>>> >> >>>>> snapshots
>>> >> >>>>> and merging them, and whatever else would be
important for
>>> storage
>>> >> >>>>> sanity.
>>> >> >>>>>
>>> >> >>>>> mpolednik
>>> >> >>>>>
>>> >> >>>>>
>>> >> >>>>> On Wed, Apr 18, 2018 at 11:16 AM, Martin
Polednik <
>>> >> >>>>> mpolednik(a)redhat.com
>>> >> >>>>> >
>>> >> >>>>>
>>> >> >>>>> wrote:
>>> >> >>>>>
>>> >> >>>>>
>>> >> >>>>> On 11/04/18 16:52 +0300, Elad Ben Aharon
wrote:
>>> >> >>>>>
>>> >> >>>>>
>>> >> >>>>> We can test this on iSCSI, NFS and GlusterFS.
As for ceph and
>>> >> >>>>> cinder,
>>> >> >>>>>
>>> >> >>>>> will
>>> >> >>>>>
>>> >> >>>>> have to check, since usually, we don't
execute our automation
>>> on
>>> >> >>>>> them.
>>> >> >>>>>
>>> >> >>>>>
>>> >> >>>>> Any update on this? I believe the gluster tests
were
>>> successful,
>>> >> >>>>> OST
>>> >> >>>>>
>>> >> >>>>> passes fine and unit tests pass fine, that
makes the storage
>>> >> >>>>> backends
>>> >> >>>>> test the last required piece.
>>> >> >>>>>
>>> >> >>>>>
>>> >> >>>>> On Wed, Apr 11, 2018 at 4:38 PM, Raz Tamir
<ratamir(a)redhat.com
>>> >
>>> >> >>>>> wrote:
>>> >> >>>>>
>>> >> >>>>>
>>> >> >>>>> +Elad
>>> >> >>>>>
>>> >> >>>>>
>>> >> >>>>>
>>> >> >>>>> On Wed, Apr 11, 2018 at 4:28 PM, Dan Kenigsberg
<
>>> danken(a)redhat.com
>>> >> >>>>>
>>> >> >>>>> >
>>> >> >>>>> wrote:
>>> >> >>>>>
>>> >> >>>>> On Wed, Apr 11, 2018 at 12:34 PM, Nir Soffer
<
>>> nsoffer(a)redhat.com>
>>> >> >>>>> wrote:
>>> >> >>>>>
>>> >> >>>>>
>>> >> >>>>> On Wed, Apr 11, 2018 at 12:31 PM Eyal Edri
<eedri(a)redhat.com>
>>> >> >>>>>
>>> >> >>>>> wrote:
>>> >> >>>>>
>>> >> >>>>>
>>> >> >>>>> Please make sure to run as much OST suites on
this patch as
>>> >> >>>>>
>>> >> >>>>> possible
>>> >> >>>>>
>>> >> >>>>> before merging ( using 'ci please
build' )
>>> >> >>>>>
>>> >> >>>>>
>>> >> >>>>>
>>> >> >>>>> But note that OST is not a way to verify the
patch.
>>> >> >>>>>
>>> >> >>>>>
>>> >> >>>>> Such changes require testing with all storage
types we support.
>>> >> >>>>>
>>> >> >>>>> Nir
>>> >> >>>>>
>>> >> >>>>> On Tue, Apr 10, 2018 at 4:09 PM, Martin
Polednik <
>>> >> >>>>> mpolednik(a)redhat.com
>>> >> >>>>> >
>>> >> >>>>>
>>> >> >>>>> wrote:
>>> >> >>>>>
>>> >> >>>>>
>>> >> >>>>> Hey,
>>> >> >>>>>
>>> >> >>>>>
>>> >> >>>>> I've created a patch[0] that is finally
able to activate
>>> >> >>>>>
>>> >> >>>>> libvirt's
>>> >> >>>>> dynamic_ownership for VDSM while not negatively
affecting
>>> >> >>>>> functionality of our storage code.
>>> >> >>>>>
>>> >> >>>>> That of course comes with quite a bit of code
removal, mostly
>>> >> >>>>> in
>>> >> >>>>> the
>>> >> >>>>> area of host devices, hwrng and anything that
touches devices;
>>> >> >>>>> bunch
>>> >> >>>>> of test changes and one XML generation caveat
(storage is
>>> >> >>>>> handled
>>> >> >>>>> by
>>> >> >>>>> VDSM, therefore disk relabelling needs to be
disabled on the
>>> >> >>>>> VDSM
>>> >> >>>>> level).
>>> >> >>>>>
>>> >> >>>>> Because of the scope of the patch, I welcome
>>> >> >>>>> storage/virt/network
>>> >> >>>>> people to review the code and consider the
implication this
>>> >> >>>>> change
>>> >> >>>>> has
>>> >> >>>>> on current/future features.
>>> >> >>>>>
>>> >> >>>>> [0]
https://gerrit.ovirt.org/#/c/89830/
>>> >> >>>>>
>>> >> >>>>>
>>> >> >>>>> In particular: dynamic_ownership was set to 0
prehistorically
>>> >> >>>>> (as
>>> >> >>>>>
>>> >> >>>>>
>>> >> >>>>> part
>>> >> >>>>>
>>> >> >>>>>
>>> >> >>>>> of
https://bugzilla.redhat.com/show_bug.cgi?id=554961 )
>>> because
>>> >> >>>>>
>>> >> >>>>> libvirt,
>>> >> >>>>> running as root, was not able to play properly
with root-squash
>>> >> >>>>> nfs
>>> >> >>>>> mounts.
>>> >> >>>>>
>>> >> >>>>> Have you attempted this use case?
>>> >> >>>>>
>>> >> >>>>> I join to Nir's request to run this with
storage QE.
>>> >> >>>>>
>>> >> >>>>>
>>> >> >>>>>
>>> >> >>>>>
>>> >> >>>>> --
>>> >> >>>>>
>>> >> >>>>>
>>> >> >>>>>
>>> >> >>>>> Raz Tamir
>>> >> >>>>> Manager, RHV QE
>>> >> >>>>>
>>> >> >>>>>
>>> >> >>>>>
>>> >> >>>>>
>>> >> >>>>>
>>> >> >>>>>
>>> >> >>>>>
>>> >> >>>>>
>>> >> >>>>>
_______________________________________________
>>> >> >>>>> Devel mailing list
>>> >> >>>>> Devel(a)ovirt.org
>>> >> >>>>>
http://lists.ovirt.org/mailman/listinfo/devel
>>> >> >>>>>
>>> >> >>>>>
>>> >> >>>>>
>>> >> >>>>
>>> >> >>>>
>>> >> >>>
>>> >> >>
>>> >> >
>>> >> > <logs.tar.gz>
>>> >> >
>>> >> >
>>> >
>>> >
>>>
>>>