Martin, please update, if you think the failures are not related to your
patch I'll test with the master as Nir suggested.

Thanks

On Thu, May 31, 2018 at 1:19 PM, Nir Soffer <nsof...@redhat.com> wrote:

> On Thu, May 31, 2018 at 1:05 PM Martin Polednik <mpoled...@redhat.com>
> wrote:
>
>> On 31/05/18 12:47 +0300, Elad Ben Aharon wrote:
>> >Execution is done, 59/65 cases passed. Latest 4.2.4 execution ended with
>> >100% so failures were caused probably due to the changes done in the
>> patch.
>> >Failures are mainly on preview snapshots.
>>
>
> Can we run the same job on the patch before Martin patch? maybe
> the issue are already in master, caused by other patches?
>
>
>> >
>> >Execution info provided to Martin separately.
>>
>> I'm currently investigating the snapshot breakage, thanks Elad!
>>
>> >On Wed, May 30, 2018 at 5:44 PM, Elad Ben Aharon <ebena...@redhat.com>
>> >wrote:
>> >
>> >> Triggered a sanity automation execution using [1], which covers all the
>> >> requested areas, on iSCSI, NFS and Gluster.
>> >> I'll update with the results.
>> >>
>> >> [1]
>> >> *https://gerrit.ovirt.org/#/c/90906/ <https://gerrit.ovirt.org/#/c/
>> 90906/>*
>> >> vdsm-4.20.28-6.gitc23aef6.el7.x86_64
>> >>
>> >>
>> >> On Tue, May 29, 2018 at 4:26 PM, Martin Polednik <mpoled...@redhat.com
>> >
>> >> wrote:
>> >>
>> >>> On 29/05/18 15:30 +0300, Elad Ben Aharon wrote:
>> >>>
>> >>>> Hi Martin,
>> >>>>
>> >>>> Can you please create a cerry pick patch that is based on 4.2?
>> >>>>
>> >>>
>> >>> See https://gerrit.ovirt.org/#/c/90906/. The CI failure isn unrelated
>> >>> (storage needs real env).
>> >>>
>> >>> mpolednik
>> >>>
>> >>>
>> >>>
>> >>>> Thanks
>> >>>>
>> >>>> On Tue, May 29, 2018 at 1:34 PM, Dan Kenigsberg <dan...@redhat.com>
>> >>>> wrote:
>> >>>>
>> >>>> On Tue, May 29, 2018 at 1:21 PM, Elad Ben Aharon <
>> ebena...@redhat.com>
>> >>>>> wrote:
>> >>>>> > Hi Dan,
>> >>>>> >
>> >>>>> > In the last execution, the success rate was very low due to a
>> large
>> >>>>> number
>> >>>>> > of failures on start VM caused, according to Michal, by the
>> >>>>> > vdsm-hook-allocate_net that was installed on the host.
>> >>>>> >
>> >>>>> > This is the latest status here, would you like me to re-execute?
>> >>>>>
>> >>>>> yes, of course. but you should rebase Polednik's code on top of
>> >>>>> *current* ovirt-4.2.3 branch.
>> >>>>>
>> >>>>> > If so, with
>> >>>>> > or W/O vdsm-hook-allocate_net installed?
>> >>>>>
>> >>>>> There was NO reason to have that installed. Please keep it (and any
>> >>>>> other needless code) out of the test environment.
>> >>>>>
>> >>>>> >
>> >>>>> > On Tue, May 29, 2018 at 1:14 PM, Dan Kenigsberg <
>> dan...@redhat.com>
>> >>>>> wrote:
>> >>>>> >>
>> >>>>> >> On Mon, May 7, 2018 at 3:53 PM, Michal Skrivanek
>> >>>>> >> <michal.skriva...@redhat.com> wrote:
>> >>>>> >> > Hi Elad,
>> >>>>> >> > why did you install vdsm-hook-allocate_net?
>> >>>>> >> >
>> >>>>> >> > adding Dan as I think the hook is not supposed to fail this
>> badly
>> >>>>> in
>> >>>>> any
>> >>>>> >> > case
>> >>>>> >>
>> >>>>> >> yep, this looks bad and deserves a little bug report. Installing
>> this
>> >>>>> >> little hook should not block vm startup.
>> >>>>> >>
>> >>>>> >> But more importantly - what is the conclusion of this thread? Do
>> we
>> >>>>> >> have a green light from QE to take this in?
>> >>>>> >>
>> >>>>> >>
>> >>>>> >> >
>> >>>>> >> > Thanks,
>> >>>>> >> > michal
>> >>>>> >> >
>> >>>>> >> > On 5 May 2018, at 19:22, Elad Ben Aharon <ebena...@redhat.com>
>> >>>>> wrote:
>> >>>>> >> >
>> >>>>> >> > Start VM fails on:
>> >>>>> >> >
>> >>>>> >> > 2018-05-05 17:53:27,399+0300 INFO  (vm/e6ce66ce) [virt.vm]
>> >>>>> >> > (vmId='e6ce66ce-852f-48c5-9997-5d2959432a27') drive 'vda'
>> path:
>> >>>>> >> >
>> >>>>> >> > 'dev=/rhev/data-center/mnt/blockSD/db5a6696-d907-4938-
>> >>>>> 9a78-bdd13a843c62/images/6cdabfe5-
>> >>>>> >> > d1ca-40af-ae63-9834f235d1c8/7ef97445-30e6-4435-8425-f35a0192
>> 8211'
>> >>>>> ->
>> >>>>> >> >
>> >>>>> >> > u'*dev=/rhev/data-center/mnt/blockSD/db5a6696-d907-4938-
>> >>>>> 9a78-bdd13a843c62/images/6cdabfe5-d1ca-40af-ae63-
>> >>>>> 9834f235d1c8/7ef97445-30e6-4435-8425-
>> >>>>> >> > f35a01928211' (storagexml:334)
>> >>>>> >> > 2018-05-05 17:53:27,888+0300 INFO  (jsonrpc/1) [vdsm.api] START
>> >>>>> >> > getSpmStatus(spUUID='940fe6f3-b0c6-4d0c-a921-198e7819c1cc',
>> >>>>> >> > options=None)
>> >>>>> >> > from=::ffff:10.35.161.127,53512,
>> >>>>> >> > task_id=c70ace39-dbfe-4f5c-ae49-a1e3a82c
>> >>>>> >> > 2758 (api:46)
>> >>>>> >> > 2018-05-05 17:53:27,909+0300 INFO  (vm/e6ce66ce) [root]
>> >>>>> >> > /usr/libexec/vdsm/hooks/before_device_create/10_allocate_net:
>> rc=2
>> >>>>> >> > err=vm
>> >>>>> >> > net allocation hook: [unexpected error]: Traceback (most recent
>> >>>>> call
>> >>>>> >> > last):
>> >>>>> >> >  File "/usr/libexec/vdsm/hooks/befor
>> e_device_create/10_allocate_ne
>> >>>>> t",
>> >>>>> >> > line
>> >>>>> >> > 105, in <module>
>> >>>>> >> >    main()
>> >>>>> >> >  File "/usr/libexec/vdsm/hooks/befor
>> e_device_create/10_allocate_ne
>> >>>>> t",
>> >>>>> >> > line
>> >>>>> >> > 93, in main
>> >>>>> >> >    allocate_random_network(device_xml)
>> >>>>> >> >  File "/usr/libexec/vdsm/hooks/befor
>> e_device_create/10_allocate_ne
>> >>>>> t",
>> >>>>> >> > line
>> >>>>> >> > 62, in allocate_random_network
>> >>>>> >> >    net = _get_random_network()
>> >>>>> >> >  File "/usr/libexec/vdsm/hooks/befor
>> e_device_create/10_allocate_ne
>> >>>>> t",
>> >>>>> >> > line
>> >>>>> >> > 50, in _get_random_network
>> >>>>> >> >    available_nets = _parse_nets()
>> >>>>> >> >  File "/usr/libexec/vdsm/hooks/befor
>> e_device_create/10_allocate_ne
>> >>>>> t",
>> >>>>> >> > line
>> >>>>> >> > 46, in _parse_nets
>> >>>>> >> >    return [net for net in os.environ[AVAIL_NETS_KEY].split()]
>> >>>>> >> >  File "/usr/lib64/python2.7/UserDict.py", line 23, in
>> __getitem__
>> >>>>> >> >    raise KeyError(key)
>> >>>>> >> > KeyError: 'equivnets'
>> >>>>> >> >
>> >>>>> >> >
>> >>>>> >> > (hooks:110)
>> >>>>> >> > 2018-05-05 17:53:27,915+0300 ERROR (vm/e6ce66ce) [virt.vm]
>> >>>>> >> > (vmId='e6ce66ce-852f-48c5-9997-5d2959432a27') The vm start
>> process
>> >>>>> >> > failed
>> >>>>> >> > (vm:943)
>> >>>>> >> > Traceback (most recent call last):
>> >>>>> >> >  File "/usr/lib/python2.7/site-packages/vdsm/virt/vm.py", line
>> >>>>> 872,
>> >>>>> in
>> >>>>> >> > _startUnderlyingVm
>> >>>>> >> >    self._run()
>> >>>>> >> >  File "/usr/lib/python2.7/site-packages/vdsm/virt/vm.py", line
>> >>>>> 2861,
>> >>>>> in
>> >>>>> >> > _run
>> >>>>> >> >    domxml = hooks.before_vm_start(self._buildDomainXML(),
>> >>>>> >> >  File "/usr/lib/python2.7/site-packages/vdsm/virt/vm.py", line
>> >>>>> 2254,
>> >>>>> in
>> >>>>> >> > _buildDomainXML
>> >>>>> >> >    dom, self.id, self._custom['custom'])
>> >>>>> >> >  File "/usr/lib/python2.7/site-packages/vdsm/virt/domxml_
>> >>>>> preprocess.py",
>> >>>>> >> > line 240, in replace_device_xml_with_hooks_xml
>> >>>>> >> >    dev_custom)
>> >>>>> >> >  File "/usr/lib/python2.7/site-packages/vdsm/common/hooks.py",
>> >>>>> line
>> >>>>> 134,
>> >>>>> >> > in
>> >>>>> >> > before_device_create
>> >>>>> >> >    params=customProperties)
>> >>>>> >> >  File "/usr/lib/python2.7/site-packages/vdsm/common/hooks.py",
>> >>>>> line
>> >>>>> 120,
>> >>>>> >> > in
>> >>>>> >> > _runHooksDir
>> >>>>> >> >    raise exception.HookError(err)
>> >>>>> >> > HookError: Hook Error: ('vm net allocation hook: [unexpected
>> >>>>> error]:
>> >>>>> >> > Traceback (most recent call last):\n  File
>> >>>>> >> > "/usr/libexec/vdsm/hooks/before_device_create/10_allocate_
>> net",
>> >>>>> line
>> >>>>> >> > 105, in
>> >>>>> >> > <module>\n    main()\n
>> >>>>> >> >  File "/usr/libexec/vdsm/hooks/befor
>> e_device_create/10_allocate_ne
>> >>>>> t",
>> >>>>> >> > line
>> >>>>> >> > 93, in main\n    allocate_random_network(device_xml)\n  File
>> >>>>> >> > "/usr/libexec/vdsm/hooks/before_device_create/10_allocate_
>> net",
>> >>>>> line
>> >>>>> 62,
>> >>>>> >> > i
>> >>>>> >> > n allocate_random_network\n    net = _get_random_network()\n
>> File
>> >>>>> >> > "/usr/libexec/vdsm/hooks/before_device_create/10_allocate_
>> net",
>> >>>>> line
>> >>>>> 50,
>> >>>>> >> > in
>> >>>>> >> > _get_random_network\n    available_nets = _parse_nets()\n  File
>> >>>>> "/us
>> >>>>> >> > r/libexec/vdsm/hooks/before_device_create/10_allocate_net",
>> line
>> >>>>> 46,
>> >>>>> in
>> >>>>> >> > _parse_nets\n    return [net for net in
>> >>>>> >> > os.environ[AVAIL_NETS_KEY].split()]\n  File
>> >>>>> >> > "/usr/lib64/python2.7/UserDict.py", line 23, in __getit
>> >>>>> >> > em__\n    raise KeyError(key)\nKeyError: \'equivnets\'\n\n\n',)
>> >>>>> >> >
>> >>>>> >> >
>> >>>>> >> >
>> >>>>> >> > Hence, the success rate was 28% against 100% running with d/s
>> >>>>> (d/s).
>> >>>>> If
>> >>>>> >> > needed, I'll compare against the latest master, but I think
>> you get
>> >>>>> the
>> >>>>> >> > picture with d/s.
>> >>>>> >> >
>> >>>>> >> > vdsm-4.20.27-3.gitfee7810.el7.centos.x86_64
>> >>>>> >> > libvirt-3.9.0-14.el7_5.3.x86_64
>> >>>>> >> > qemu-kvm-rhev-2.10.0-21.el7_5.2.x86_64
>> >>>>> >> > kernel 3.10.0-862.el7.x86_64
>> >>>>> >> > rhel7.5
>> >>>>> >> >
>> >>>>> >> >
>> >>>>> >> > Logs attached
>> >>>>> >> >
>> >>>>> >> > On Sat, May 5, 2018 at 1:26 PM, Elad Ben Aharon <
>> >>>>> ebena...@redhat.com>
>> >>>>> >> > wrote:
>> >>>>> >> >>
>> >>>>> >> >> nvm, found gluster 3.12 repo, managed to install vdsm
>> >>>>> >> >>
>> >>>>> >> >> On Sat, May 5, 2018 at 1:12 PM, Elad Ben Aharon <
>> >>>>> ebena...@redhat.com
>> >>>>> >
>> >>>>> >> >> wrote:
>> >>>>> >> >>>
>> >>>>> >> >>> No, vdsm requires it:
>> >>>>> >> >>>
>> >>>>> >> >>> Error: Package: vdsm-4.20.27-3.gitfee7810.el7.centos.x86_64
>> >>>>> >> >>> (/vdsm-4.20.27-3.gitfee7810.el7.centos.x86_64)
>> >>>>> >> >>>           Requires: glusterfs-fuse >= 3.12
>> >>>>> >> >>>           Installed: glusterfs-fuse-3.8.4-54.8.el7.x86_64
>> >>>>> (@rhv-4.2.3)
>> >>>>> >> >>>
>> >>>>> >> >>> Therefore, vdsm package installation is skipped upon force
>> >>>>> install.
>> >>>>> >> >>>
>> >>>>> >> >>> On Sat, May 5, 2018 at 11:42 AM, Michal Skrivanek
>> >>>>> >> >>> <michal.skriva...@redhat.com> wrote:
>> >>>>> >> >>>>
>> >>>>> >> >>>>
>> >>>>> >> >>>>
>> >>>>> >> >>>> On 5 May 2018, at 00:38, Elad Ben Aharon <
>> ebena...@redhat.com>
>> >>>>> wrote:
>> >>>>> >> >>>>
>> >>>>> >> >>>> Hi guys,
>> >>>>> >> >>>>
>> >>>>> >> >>>> The vdsm build from the patch requires glusterfs-fuse >
>> 3.12.
>> >>>>> This
>> >>>>> is
>> >>>>> >> >>>> while the latest 4.2.3-5 d/s build requires 3.8.4
>> >>>>> (3.4.0.59rhs-1.el7)
>> >>>>> >> >>>>
>> >>>>> >> >>>>
>> >>>>> >> >>>> because it is still oVirt, not a downstream build. We can’t
>> >>>>> really
>> >>>>> do
>> >>>>> >> >>>> downstream builds with unmerged changes:/
>> >>>>> >> >>>>
>> >>>>> >> >>>> Trying to get this gluster-fuse build, so far no luck.
>> >>>>> >> >>>> Is this requirement intentional?
>> >>>>> >> >>>>
>> >>>>> >> >>>>
>> >>>>> >> >>>> it should work regardless, I guess you can force install it
>> >>>>> without
>> >>>>> >> >>>> the
>> >>>>> >> >>>> dependency
>> >>>>> >> >>>>
>> >>>>> >> >>>>
>> >>>>> >> >>>> On Fri, May 4, 2018 at 2:38 PM, Michal Skrivanek
>> >>>>> >> >>>> <michal.skriva...@redhat.com> wrote:
>> >>>>> >> >>>>>
>> >>>>> >> >>>>> Hi Elad,
>> >>>>> >> >>>>> to make it easier to compare, Martin backported the change
>> to
>> >>>>> 4.2
>> >>>>> so
>> >>>>> >> >>>>> it
>> >>>>> >> >>>>> is actually comparable with a run without that patch.
>> Would you
>> >>>>> >> >>>>> please try
>> >>>>> >> >>>>> that out?
>> >>>>> >> >>>>> It would be best to have 4.2 upstream and this[1] run to
>> really
>> >>>>> >> >>>>> minimize the noise.
>> >>>>> >> >>>>>
>> >>>>> >> >>>>> Thanks,
>> >>>>> >> >>>>> michal
>> >>>>> >> >>>>>
>> >>>>> >> >>>>> [1]
>> >>>>> >> >>>>>
>> >>>>> >> >>>>> http://jenkins.ovirt.org/job/vdsm_4.2_build-artifacts-on-
>> >>>>> demand-el7-x86_64/28/
>> >>>>> >> >>>>>
>> >>>>> >> >>>>> On 27 Apr 2018, at 09:23, Martin Polednik <
>> >>>>> mpoled...@redhat.com>
>> >>>>> >> >>>>> wrote:
>> >>>>> >> >>>>>
>> >>>>> >> >>>>> On 24/04/18 00:37 +0300, Elad Ben Aharon wrote:
>> >>>>> >> >>>>>
>> >>>>> >> >>>>> I will update with the results of the next tier1 execution
>> on
>> >>>>> latest
>> >>>>> >> >>>>> 4.2.3
>> >>>>> >> >>>>>
>> >>>>> >> >>>>>
>> >>>>> >> >>>>> That isn't master but old branch though. Could you run it
>> >>>>> against
>> >>>>> >> >>>>> *current* VDSM master?
>> >>>>> >> >>>>>
>> >>>>> >> >>>>> On Mon, Apr 23, 2018 at 3:56 PM, Martin Polednik
>> >>>>> >> >>>>> <mpoled...@redhat.com>
>> >>>>> >> >>>>> wrote:
>> >>>>> >> >>>>>
>> >>>>> >> >>>>> On 23/04/18 01:23 +0300, Elad Ben Aharon wrote:
>> >>>>> >> >>>>>
>> >>>>> >> >>>>> Hi, I've triggered another execution [1] due to some
>> issues I
>> >>>>> saw
>> >>>>> in
>> >>>>> >> >>>>> the
>> >>>>> >> >>>>> first which are not related to the patch.
>> >>>>> >> >>>>>
>> >>>>> >> >>>>> The success rate is 78% which is low comparing to tier1
>> >>>>> executions
>> >>>>> >> >>>>> with
>> >>>>> >> >>>>> code from downstream builds (95-100% success rates) [2].
>> >>>>> >> >>>>>
>> >>>>> >> >>>>>
>> >>>>> >> >>>>> Could you run the current master (without the
>> dynamic_ownership
>> >>>>> >> >>>>> patch)
>> >>>>> >> >>>>> so that we have viable comparision?
>> >>>>> >> >>>>>
>> >>>>> >> >>>>> From what I could see so far, there is an issue with move
>> and
>> >>>>> copy
>> >>>>> >> >>>>>
>> >>>>> >> >>>>> operations to and from Gluster domains. For example [3].
>> >>>>> >> >>>>>
>> >>>>> >> >>>>> The logs are attached.
>> >>>>> >> >>>>>
>> >>>>> >> >>>>>
>> >>>>> >> >>>>> [1]
>> >>>>> >> >>>>> *https://rhv-jenkins.rhev-ci-v
>> ms.eng.rdu2.redhat.com/job/rhv
>> >>>>> >> >>>>> -4.2-ge-runner-tier1-after-upgrade/7/testReport/
>> >>>>> >> >>>>> <https://rhv-jenkins.rhev-ci-v
>> ms.eng.rdu2.redhat.com/job/rhv
>> >>>>> >> >>>>> -4.2-ge-runner-tier1-after-upgrade/7/testReport/>*
>> >>>>> >> >>>>>
>> >>>>> >> >>>>>
>> >>>>> >> >>>>>
>> >>>>> >> >>>>> [2]
>> >>>>> >> >>>>> https://rhv-jenkins.rhev-ci-vms.eng.rdu2.redhat.com/job/
>> >>>>> >> >>>>>
>> >>>>> >> >>>>> rhv-4.2-ge-runner-tier1-after-upgrade/7/
>> >>>>> >> >>>>>
>> >>>>> >> >>>>>
>> >>>>> >> >>>>>
>> >>>>> >> >>>>> [3]
>> >>>>> >> >>>>> 2018-04-22 13:06:28,316+0300 INFO  (jsonrpc/7) [vdsm.api]
>> >>>>> FINISH
>> >>>>> >> >>>>> deleteImage error=Image does not exist in domain:
>> >>>>> >> >>>>> 'image=cabb8846-7a4b-4244-9835-5f603e682f33,
>> >>>>> >> >>>>> domain=e5fd29c8-52ba-467e-be09-ca40ff054dd4'
>> >>>>> >> >>>>> from=:
>> >>>>> >> >>>>> :ffff:10.35.161.182,40936,
>> >>>>> >> >>>>> flow_id=disks_syncAction_ba6b2630-5976-4935,
>> >>>>> >> >>>>> task_id=3d5f2a8a-881c-409e-93e9-aaa643c10e42 (api:51)
>> >>>>> >> >>>>> 2018-04-22 13:06:28,317+0300 ERROR (jsonrpc/7)
>> >>>>> >> >>>>> [storage.TaskManager.Task]
>> >>>>> >> >>>>> (Task='3d5f2a8a-881c-409e-93e9-aaa643c10e42') Unexpected
>> error
>> >>>>> >> >>>>> (task:875)
>> >>>>> >> >>>>> Traceback (most recent call last):
>> >>>>> >> >>>>> File "/usr/lib/python2.7/site-packa
>> ges/vdsm/storage/task.py",
>> >>>>> line
>> >>>>> >> >>>>> 882,
>> >>>>> >> >>>>> in
>> >>>>> >> >>>>> _run
>> >>>>> >> >>>>>  return fn(*args, **kargs)
>> >>>>> >> >>>>> File "<string>", line 2, in deleteImage
>> >>>>> >> >>>>> File "/usr/lib/python2.7/site-packa
>> ges/vdsm/common/api.py",
>> >>>>> line
>> >>>>> 49,
>> >>>>> >> >>>>> in
>> >>>>> >> >>>>> method
>> >>>>> >> >>>>>  ret = func(*args, **kwargs)
>> >>>>> >> >>>>> File "/usr/lib/python2.7/site-packa
>> ges/vdsm/storage/hsm.py",
>> >>>>> line
>> >>>>> >> >>>>> 1503,
>> >>>>> >> >>>>> in
>> >>>>> >> >>>>> deleteImage
>> >>>>> >> >>>>>  raise se.ImageDoesNotExistInSD(imgUUID, sdUUID)
>> >>>>> >> >>>>> ImageDoesNotExistInSD: Image does not exist in domain:
>> >>>>> >> >>>>> 'image=cabb8846-7a4b-4244-9835-5f603e682f33,
>> >>>>> >> >>>>> domain=e5fd29c8-52ba-467e-be09-ca40ff054dd4'
>> >>>>> >> >>>>>
>> >>>>> >> >>>>> 2018-04-22 13:06:28,317+0300 INFO  (jsonrpc/7)
>> >>>>> >> >>>>> [storage.TaskManager.Task]
>> >>>>> >> >>>>> (Task='3d5f2a8a-881c-409e-93e9-aaa643c10e42') aborting:
>> Task
>> >>>>> is
>> >>>>> >> >>>>> aborted:
>> >>>>> >> >>>>> "Image does not exist in domain:
>> 'image=cabb8846-7a4b-4244-9835
>> >>>>> -
>> >>>>> >> >>>>> 5f603e682f33, domain=e5fd29c8-52ba-467e-be09-ca40ff054dd4'"
>> -
>> >>>>> code
>> >>>>> >> >>>>> 268
>> >>>>> >> >>>>> (task:1181)
>> >>>>> >> >>>>> 2018-04-22 13:06:28,318+0300 ERROR (jsonrpc/7)
>> >>>>> [storage.Dispatcher]
>> >>>>> >> >>>>> FINISH
>> >>>>> >> >>>>> deleteImage error=Image does not exist in domain:
>> >>>>> >> >>>>> 'image=cabb8846-7a4b-4244-9835-5f603e682f33,
>> >>>>> >> >>>>> domain=e5fd29c8-52ba-467e-be09
>> >>>>> >> >>>>> -ca40ff054d
>> >>>>> >> >>>>> d4' (dispatcher:82)
>> >>>>> >> >>>>>
>> >>>>> >> >>>>>
>> >>>>> >> >>>>>
>> >>>>> >> >>>>> On Thu, Apr 19, 2018 at 5:34 PM, Elad Ben Aharon
>> >>>>> >> >>>>> <ebena...@redhat.com>
>> >>>>> >> >>>>> wrote:
>> >>>>> >> >>>>>
>> >>>>> >> >>>>> Triggered a sanity tier1 execution [1] using [2], which
>> covers
>> >>>>> all
>> >>>>> >> >>>>> the
>> >>>>> >> >>>>>
>> >>>>> >> >>>>> requested areas, on iSCSI, NFS and Gluster.
>> >>>>> >> >>>>> I'll update with the results.
>> >>>>> >> >>>>>
>> >>>>> >> >>>>> [1]
>> >>>>> >> >>>>> https://rhv-jenkins.rhev-ci-vm
>> s.eng.rdu2.redhat.com/view/4.2
>> >>>>> >> >>>>> _dev/job/rhv-4.2-ge-flow-storage/1161/
>> >>>>> >> >>>>>
>> >>>>> >> >>>>> [2]
>> >>>>> >> >>>>> https://gerrit.ovirt.org/#/c/89830/
>> >>>>> >> >>>>> vdsm-4.30.0-291.git77aef9a.el7.x86_64
>> >>>>> >> >>>>>
>> >>>>> >> >>>>>
>> >>>>> >> >>>>>
>> >>>>> >> >>>>> On Thu, Apr 19, 2018 at 3:07 PM, Martin Polednik
>> >>>>> >> >>>>> <mpoled...@redhat.com>
>> >>>>> >> >>>>> wrote:
>> >>>>> >> >>>>>
>> >>>>> >> >>>>> On 19/04/18 14:54 +0300, Elad Ben Aharon wrote:
>> >>>>> >> >>>>>
>> >>>>> >> >>>>>
>> >>>>> >> >>>>> Hi Martin,
>> >>>>> >> >>>>>
>> >>>>> >> >>>>>
>> >>>>> >> >>>>> I see [1] requires a rebase, can you please take care?
>> >>>>> >> >>>>>
>> >>>>> >> >>>>>
>> >>>>> >> >>>>> Should be rebased.
>> >>>>> >> >>>>>
>> >>>>> >> >>>>> At the moment, our automation is stable only on iSCSI, NFS,
>> >>>>> Gluster
>> >>>>> >> >>>>> and
>> >>>>> >> >>>>>
>> >>>>> >> >>>>> FC.
>> >>>>> >> >>>>> Ceph is not supported and Cinder will be stabilized soon,
>> >>>>> AFAIR,
>> >>>>> >> >>>>> it's
>> >>>>> >> >>>>> not
>> >>>>> >> >>>>> stable enough at the moment.
>> >>>>> >> >>>>>
>> >>>>> >> >>>>>
>> >>>>> >> >>>>> That is still pretty good.
>> >>>>> >> >>>>>
>> >>>>> >> >>>>>
>> >>>>> >> >>>>> [1] https://gerrit.ovirt.org/#/c/89830/
>> >>>>> >> >>>>>
>> >>>>> >> >>>>>
>> >>>>> >> >>>>>
>> >>>>> >> >>>>> Thanks
>> >>>>> >> >>>>>
>> >>>>> >> >>>>> On Wed, Apr 18, 2018 at 2:17 PM, Martin Polednik
>> >>>>> >> >>>>> <mpoled...@redhat.com
>> >>>>> >> >>>>> >
>> >>>>> >> >>>>> wrote:
>> >>>>> >> >>>>>
>> >>>>> >> >>>>> On 18/04/18 11:37 +0300, Elad Ben Aharon wrote:
>> >>>>> >> >>>>>
>> >>>>> >> >>>>>
>> >>>>> >> >>>>> Hi, sorry if I misunderstood, I waited for more input
>> regarding
>> >>>>> what
>> >>>>> >> >>>>>
>> >>>>> >> >>>>> areas
>> >>>>> >> >>>>> have to be tested here.
>> >>>>> >> >>>>>
>> >>>>> >> >>>>>
>> >>>>> >> >>>>> I'd say that you have quite a bit of freedom in this
>> regard.
>> >>>>> >> >>>>>
>> >>>>> >> >>>>> GlusterFS
>> >>>>> >> >>>>> should be covered by Dennis, so iSCSI/NFS/ceph/cinder with
>> some
>> >>>>> >> >>>>> suite
>> >>>>> >> >>>>> that covers basic operations (start & stop VM, migrate it),
>> >>>>> >> >>>>> snapshots
>> >>>>> >> >>>>> and merging them, and whatever else would be important for
>> >>>>> storage
>> >>>>> >> >>>>> sanity.
>> >>>>> >> >>>>>
>> >>>>> >> >>>>> mpolednik
>> >>>>> >> >>>>>
>> >>>>> >> >>>>>
>> >>>>> >> >>>>> On Wed, Apr 18, 2018 at 11:16 AM, Martin Polednik <
>> >>>>> >> >>>>> mpoled...@redhat.com
>> >>>>> >> >>>>> >
>> >>>>> >> >>>>>
>> >>>>> >> >>>>> wrote:
>> >>>>> >> >>>>>
>> >>>>> >> >>>>>
>> >>>>> >> >>>>> On 11/04/18 16:52 +0300, Elad Ben Aharon wrote:
>> >>>>> >> >>>>>
>> >>>>> >> >>>>>
>> >>>>> >> >>>>> We can test this on iSCSI, NFS and GlusterFS. As for ceph
>> and
>> >>>>> >> >>>>> cinder,
>> >>>>> >> >>>>>
>> >>>>> >> >>>>> will
>> >>>>> >> >>>>>
>> >>>>> >> >>>>> have to check, since usually, we don't execute our
>> automation
>> >>>>> on
>> >>>>> >> >>>>> them.
>> >>>>> >> >>>>>
>> >>>>> >> >>>>>
>> >>>>> >> >>>>> Any update on this? I believe the gluster tests were
>> >>>>> successful,
>> >>>>> >> >>>>> OST
>> >>>>> >> >>>>>
>> >>>>> >> >>>>> passes fine and unit tests pass fine, that makes the
>> storage
>> >>>>> >> >>>>> backends
>> >>>>> >> >>>>> test the last required piece.
>> >>>>> >> >>>>>
>> >>>>> >> >>>>>
>> >>>>> >> >>>>> On Wed, Apr 11, 2018 at 4:38 PM, Raz Tamir <
>> rata...@redhat.com
>> >>>>> >
>> >>>>> >> >>>>> wrote:
>> >>>>> >> >>>>>
>> >>>>> >> >>>>>
>> >>>>> >> >>>>> +Elad
>> >>>>> >> >>>>>
>> >>>>> >> >>>>>
>> >>>>> >> >>>>>
>> >>>>> >> >>>>> On Wed, Apr 11, 2018 at 4:28 PM, Dan Kenigsberg <
>> >>>>> dan...@redhat.com
>> >>>>> >> >>>>>
>> >>>>> >> >>>>> >
>> >>>>> >> >>>>> wrote:
>> >>>>> >> >>>>>
>> >>>>> >> >>>>> On Wed, Apr 11, 2018 at 12:34 PM, Nir Soffer <
>> >>>>> nsof...@redhat.com>
>> >>>>> >> >>>>> wrote:
>> >>>>> >> >>>>>
>> >>>>> >> >>>>>
>> >>>>> >> >>>>> On Wed, Apr 11, 2018 at 12:31 PM Eyal Edri <
>> ee...@redhat.com>
>> >>>>> >> >>>>>
>> >>>>> >> >>>>> wrote:
>> >>>>> >> >>>>>
>> >>>>> >> >>>>>
>> >>>>> >> >>>>> Please make sure to run as much OST suites on this patch as
>> >>>>> >> >>>>>
>> >>>>> >> >>>>> possible
>> >>>>> >> >>>>>
>> >>>>> >> >>>>> before merging ( using 'ci please build' )
>> >>>>> >> >>>>>
>> >>>>> >> >>>>>
>> >>>>> >> >>>>>
>> >>>>> >> >>>>> But note that OST is not a way to verify the patch.
>> >>>>> >> >>>>>
>> >>>>> >> >>>>>
>> >>>>> >> >>>>> Such changes require testing with all storage types we
>> support.
>> >>>>> >> >>>>>
>> >>>>> >> >>>>> Nir
>> >>>>> >> >>>>>
>> >>>>> >> >>>>> On Tue, Apr 10, 2018 at 4:09 PM, Martin Polednik <
>> >>>>> >> >>>>> mpoled...@redhat.com
>> >>>>> >> >>>>> >
>> >>>>> >> >>>>>
>> >>>>> >> >>>>> wrote:
>> >>>>> >> >>>>>
>> >>>>> >> >>>>>
>> >>>>> >> >>>>> Hey,
>> >>>>> >> >>>>>
>> >>>>> >> >>>>>
>> >>>>> >> >>>>> I've created a patch[0] that is finally able to activate
>> >>>>> >> >>>>>
>> >>>>> >> >>>>> libvirt's
>> >>>>> >> >>>>> dynamic_ownership for VDSM while not negatively affecting
>> >>>>> >> >>>>> functionality of our storage code.
>> >>>>> >> >>>>>
>> >>>>> >> >>>>> That of course comes with quite a bit of code removal,
>> mostly
>> >>>>> >> >>>>> in
>> >>>>> >> >>>>> the
>> >>>>> >> >>>>> area of host devices, hwrng and anything that touches
>> devices;
>> >>>>> >> >>>>> bunch
>> >>>>> >> >>>>> of test changes and one XML generation caveat (storage is
>> >>>>> >> >>>>> handled
>> >>>>> >> >>>>> by
>> >>>>> >> >>>>> VDSM, therefore disk relabelling needs to be disabled on
>> the
>> >>>>> >> >>>>> VDSM
>> >>>>> >> >>>>> level).
>> >>>>> >> >>>>>
>> >>>>> >> >>>>> Because of the scope of the patch, I welcome
>> >>>>> >> >>>>> storage/virt/network
>> >>>>> >> >>>>> people to review the code and consider the implication this
>> >>>>> >> >>>>> change
>> >>>>> >> >>>>> has
>> >>>>> >> >>>>> on current/future features.
>> >>>>> >> >>>>>
>> >>>>> >> >>>>> [0] https://gerrit.ovirt.org/#/c/89830/
>> >>>>> >> >>>>>
>> >>>>> >> >>>>>
>> >>>>> >> >>>>> In particular:  dynamic_ownership was set to 0
>> prehistorically
>> >>>>> >> >>>>> (as
>> >>>>> >> >>>>>
>> >>>>> >> >>>>>
>> >>>>> >> >>>>> part
>> >>>>> >> >>>>>
>> >>>>> >> >>>>>
>> >>>>> >> >>>>> of https://bugzilla.redhat.com/show_bug.cgi?id=554961 )
>> >>>>> because
>> >>>>> >> >>>>>
>> >>>>> >> >>>>> libvirt,
>> >>>>> >> >>>>> running as root, was not able to play properly with
>> root-squash
>> >>>>> >> >>>>> nfs
>> >>>>> >> >>>>> mounts.
>> >>>>> >> >>>>>
>> >>>>> >> >>>>> Have you attempted this use case?
>> >>>>> >> >>>>>
>> >>>>> >> >>>>> I join to Nir's request to run this with storage QE.
>> >>>>> >> >>>>>
>> >>>>> >> >>>>>
>> >>>>> >> >>>>>
>> >>>>> >> >>>>>
>> >>>>> >> >>>>> --
>> >>>>> >> >>>>>
>> >>>>> >> >>>>>
>> >>>>> >> >>>>>
>> >>>>> >> >>>>> Raz Tamir
>> >>>>> >> >>>>> Manager, RHV QE
>> >>>>> >> >>>>>
>> >>>>> >> >>>>>
>> >>>>> >> >>>>>
>> >>>>> >> >>>>>
>> >>>>> >> >>>>>
>> >>>>> >> >>>>>
>> >>>>> >> >>>>>
>> >>>>> >> >>>>>
>> >>>>> >> >>>>> _______________________________________________
>> >>>>> >> >>>>> Devel mailing list
>> >>>>> >> >>>>> Devel@ovirt.org
>> >>>>> >> >>>>> http://lists.ovirt.org/mailman/listinfo/devel
>> >>>>> >> >>>>>
>> >>>>> >> >>>>>
>> >>>>> >> >>>>>
>> >>>>> >> >>>>
>> >>>>> >> >>>>
>> >>>>> >> >>>
>> >>>>> >> >>
>> >>>>> >> >
>> >>>>> >> > <logs.tar.gz>
>> >>>>> >> >
>> >>>>> >> >
>> >>>>> >
>> >>>>> >
>> >>>>>
>> >>>>>
>> >>
>> _______________________________________________
>> Devel mailing list -- devel@ovirt.org
>> To unsubscribe send an email to devel-le...@ovirt.org
>> Privacy Statement: https://www.ovirt.org/site/privacy-policy/
>> oVirt Code of Conduct: https://www.ovirt.org/communit
>> y/about/community-guidelines/
>> List Archives: https://lists.ovirt.org/archiv
>> es/list/devel@ovirt.org/message/WGMUK5T7PDYSIBIUZE3AJIYQAWJOILTC/
>>
>
_______________________________________________
Devel mailing list -- devel@ovirt.org
To unsubscribe send an email to devel-le...@ovirt.org
Privacy Statement: https://www.ovirt.org/site/privacy-policy/
oVirt Code of Conduct: 
https://www.ovirt.org/community/about/community-guidelines/
List Archives: 
https://lists.ovirt.org/archives/list/devel@ovirt.org/message/KF7W6RJE7WHYSBU4ZSUPRHCJEKLI4PKZ/

Reply via email to