[Yahoo-eng-team] [Bug 1936667] Re: Using or importing the ABCs from 'collections' instead of from 'collections.abc' is deprecated since Python 3.3

2021-08-18 Thread OpenStack Infra
Reviewed:  https://review.opendev.org/c/openstack/zaqar/+/801080
Committed: 
https://opendev.org/openstack/zaqar/commit/24538768a6ff007ca37a0be2fbddb3086eb01694
Submitter: "Zuul (22348)"
Branch:master

commit 24538768a6ff007ca37a0be2fbddb3086eb01694
Author: Takashi Kajinami 
Date:   Sat Jul 17 00:20:28 2021 +0900

Replace deprecated import of ABCs from collections

ABCs in collections should be imported from collections.abc and direct
import from collections is deprecated since Python 3.3.

Closes-Bug: #1936667
Change-Id: I60679f9e78e91bce1145d899332e871dd62d3a9c


** Changed in: zaqar
   Status: In Progress => Fix Released

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Identity (keystone).
https://bugs.launchpad.net/bugs/1936667

Title:
  Using or importing the ABCs from 'collections' instead of from
  'collections.abc' is deprecated since Python 3.3

Status in OpenStack Identity (keystone):
  In Progress
Status in OpenStack Shared File Systems Service (Manila):
  Fix Released
Status in Mistral:
  In Progress
Status in neutron:
  Fix Released
Status in OpenStack Object Storage (swift):
  Fix Released
Status in tacker:
  In Progress
Status in taskflow:
  Fix Released
Status in tempest:
  Fix Released
Status in zaqar:
  Fix Released

Bug description:
  Using or importing the ABCs from 'collections' instead of from
  'collections.abc' is deprecated since Python 3.3.

  For example:

  >>> import collections 
  >>> collections.Iterable
  :1: DeprecationWarning: Using or importing the ABCs from 'collections' 
instead of from 'collections.abc' is deprecated since Python 3.3, and in 3.10 
it will stop working
  

  >>> from collections import abc
  >>> abc.Iterable
  

To manage notifications about this bug go to:
https://bugs.launchpad.net/keystone/+bug/1936667/+subscriptions


-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1933502] Re: L3 DB FloatingIP callbacks improvements

2021-08-18 Thread OpenStack Infra
Reviewed:  https://review.opendev.org/c/openstack/neutron/+/801453
Committed: 
https://opendev.org/openstack/neutron/commit/b488fb8e223498703eed158e741c72d7ce03fe16
Submitter: "Zuul (22348)"
Branch:master

commit b488fb8e223498703eed158e741c72d7ce03fe16
Author: Szymon Wroblewski 
Date:   Mon Jul 19 12:25:32 2021 +0200

Use payloads for FloatingIP AFTER callbacks

Change FloatingIP AFTER_ notifications in l3_db to use publish calls.
Move association_event field into Payload metadata.

Closes-Bug: #1933502
Change-Id: Ie4c0f4a63a87c32026c49b03068e5f461deb38b6


** Changed in: neutron
   Status: In Progress => Fix Released

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to neutron.
https://bugs.launchpad.net/bugs/1933502

Title:
  L3 DB FloatingIP callbacks improvements

Status in neutron:
  Fix Released

Bug description:
  FloatingIP change notifications in l3_db use obsolete notify calls
  (instead of publish) and form that isn't compatible with
  OVOServerRpcInterface, which makes usage of RemoteResourceCache for
  FloatingIP by agents quite difficult.

To manage notifications about this bug go to:
https://bugs.launchpad.net/neutron/+bug/1933502/+subscriptions


-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1905944] Re: live-migration job not aborted when live_monitor thread fails

2021-08-18 Thread OpenStack Infra
Reviewed:  https://review.opendev.org/c/openstack/nova/+/764435
Committed: 
https://opendev.org/openstack/nova/commit/39f0af5d18d6bea34fa15b8f7778115b25432749
Submitter: "Zuul (22348)"
Branch:master

commit 39f0af5d18d6bea34fa15b8f7778115b25432749
Author: Alexandre Arents 
Date:   Thu Nov 26 15:24:19 2020 +

libvirt: Abort live-migration job when monitoring fails

During live migration process, a _live_migration_monitor thread
checks progress of migration on source host, if for any reason
we hit infrastructure issue involving a DB/RPC/libvirt-timeout
failure, an Exception is raised to the nova-compute service and
instance/migration is set to ERROR state.

The issue is that we may let live-migration job running out of nova
control. At the end of job, guest is resumed on target host while
nova still reports it on source host, this may lead to a split-brain
situation if instance is restarted.

This change proposes to abort live-migration job if issue occurs
during _live_migration_monitor.

Change-Id: Ia593b500425c81e54eb401e38264db5cc5fc1f93
Closes-Bug: #1905944


** Changed in: nova
   Status: In Progress => Fix Released

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1905944

Title:
  live-migration job not aborted when live_monitor thread fails

Status in OpenStack Compute (nova):
  Fix Released

Bug description:
  Description
  ===

  During live migration, a monitoring thread poll each 0.5s libvirt job
  progress and update db with with jobs stats. If there control pane issue
  like DB/RPC or libvirt unexpected Exception (timeout)
  exception handling do not properly interrupt libvirt job.

  Steps to reproduce
  ==
  On a multinode devstack master.

  #spawn instance on source_host
  1) openstack server create --flavor m1.small --image cirros-0.5.1-x86_64-disk 
\
  --nic net-id=private inst

  #ignite live block migration on dest_host, wait a bit( to be in monitoring 
thread),
  # and trigger an issue on DB for ex.
  2) nova live-migration inst ; sleep 6 ; sudo service mysql restart

  3) On source host you can survey libvirt job progess until it complete and 
disappear
  because libvirt resume guest on target host(starting writting data on target 
disk)
  source_host$ watch -n 1 virsh domjobinfo instance-000d

  4) on dest host you will find instance active
  dest_host$ virsh list
   Id   NameState
  ---
   20   instance-000d   running

  5) nova show inst show instance still on source host.
  $nova show inst | grep host
  | OS-EXT-SRV-ATTR:host | source_host

  
  if admin try to recover the instance on source on as it in on nova DB,
  we can fall in split-brain where 2 qemu running on two different disks on two 
host
  (true story..)

  Expected result
  ===
  If issue happen we must at least ensure that libvirt job is interrupted, 
avoiding
  the guest resume on target host.

  Actual result
  =
  If issue happen libvirt job continue and bring up guest on target host,
  nova still consider it on source.

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1905944/+subscriptions


-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1940460] [NEW] ORM fixes broke opportunistic testing on py36

2021-08-18 Thread Dan Smith
Public bug reported:

The patch 9e002a77f2131d3594a2a4029a147beaf37f5b17 which is aimed at
fixing things in advance of SQLAlchemy 2.0 seems to have broken our
opportunistic testing of DB migrations on py36 only. This manifests as a
total lockup of one worker during functional tests, which fails to
report anything and eventually times out the job.

** Affects: glance
 Importance: Undecided
 Status: New

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to Glance.
https://bugs.launchpad.net/bugs/1940460

Title:
  ORM fixes broke opportunistic testing on py36

Status in Glance:
  New

Bug description:
  The patch 9e002a77f2131d3594a2a4029a147beaf37f5b17 which is aimed at
  fixing things in advance of SQLAlchemy 2.0 seems to have broken our
  opportunistic testing of DB migrations on py36 only. This manifests as
  a total lockup of one worker during functional tests, which fails to
  report anything and eventually times out the job.

To manage notifications about this bug go to:
https://bugs.launchpad.net/glance/+bug/1940460/+subscriptions


-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1940425] [NEW] test_live_migration_with_trunk tempest test fails due to port remains in down state

2021-08-18 Thread Balazs Gibizer
Public bug reported:

Example failure is in [1]:

2021-08-18 10:40:52,334 124842 DEBUG[tempest.lib.common.utils.test_utils] 
Call _is_port_status_active returns false in 60.00 seconds
}}}

Traceback (most recent call last):
  File "/opt/stack/tempest/tempest/common/utils/__init__.py", line 89, in 
wrapper
return func(*func_args, **func_kwargs)
  File "/opt/stack/tempest/tempest/common/utils/__init__.py", line 70, in 
wrapper
return f(*func_args, **func_kwargs)
  File "/opt/stack/tempest/tempest/api/compute/admin/test_live_migration.py", 
line 281, in test_live_migration_with_trunk
self.assertTrue(
  File "/usr/lib/python3.8/unittest/case.py", line 765, in assertTrue
raise self.failureException(msg)
AssertionError: False is not true

Please note that a similar bug was reported and fixed previously
https://bugs.launchpad.net/tempest/+bug/1924258 It seems that fix did
not fully solved the issue.

It is not super frequent I saw 4 occasions in the last 30 days [2].

[1] 
https://zuul.opendev.org/t/openstack/build/fdbda223dc10456db58f922b6435f680/logs
[2] https://paste.opendev.org/show/808166/

** Affects: neutron
 Importance: Undecided
 Status: New


** Tags: tempest trunk

** Tags added: tempest trunk

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to neutron.
https://bugs.launchpad.net/bugs/1940425

Title:
  test_live_migration_with_trunk tempest test fails due to port remains
  in down state

Status in neutron:
  New

Bug description:
  Example failure is in [1]:

  2021-08-18 10:40:52,334 124842 DEBUG[tempest.lib.common.utils.test_utils] 
Call _is_port_status_active returns false in 60.00 seconds
  }}}

  Traceback (most recent call last):
File "/opt/stack/tempest/tempest/common/utils/__init__.py", line 89, in 
wrapper
  return func(*func_args, **func_kwargs)
File "/opt/stack/tempest/tempest/common/utils/__init__.py", line 70, in 
wrapper
  return f(*func_args, **func_kwargs)
File "/opt/stack/tempest/tempest/api/compute/admin/test_live_migration.py", 
line 281, in test_live_migration_with_trunk
  self.assertTrue(
File "/usr/lib/python3.8/unittest/case.py", line 765, in assertTrue
  raise self.failureException(msg)
  AssertionError: False is not true

  Please note that a similar bug was reported and fixed previously
  https://bugs.launchpad.net/tempest/+bug/1924258 It seems that fix did
  not fully solved the issue.

  It is not super frequent I saw 4 occasions in the last 30 days [2].

  [1] 
https://zuul.opendev.org/t/openstack/build/fdbda223dc10456db58f922b6435f680/logs
  [2] https://paste.opendev.org/show/808166/

To manage notifications about this bug go to:
https://bugs.launchpad.net/neutron/+bug/1940425/+subscriptions


-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1940009] Re: Some options are missing from neutron.conf generated by "tox -e genconfig"

2021-08-18 Thread OpenStack Infra
Reviewed:  https://review.opendev.org/c/openstack/neutron/+/804571
Committed: 
https://opendev.org/openstack/neutron/commit/c75df16e5021e28aaa9961ed0977f30fa278a1dc
Submitter: "Zuul (22348)"
Branch:master

commit c75df16e5021e28aaa9961ed0977f30fa278a1dc
Author: Takashi Kajinami 
Date:   Sat Aug 14 23:29:49 2021 +0900

Add missing options to generated neutron.conf

This change adds missing oslo.config.opts endpoints to the config file
to generated neutron.conf, so that the following options are included
in the neutron.conf generated by the oslo-config-generator command.

1) Some options of the oslo.service library, related to periodic tasks
   and base service implementation

2) Options of the oslo.reports library

3) Options of the healthcheck middleware from oslo.middleware

Closes-Bug: #1940009
Change-Id: I9f4763efc02e5cacca3a00e9f101f83d3f3fb1e7


** Changed in: neutron
   Status: In Progress => Fix Released

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to neutron.
https://bugs.launchpad.net/bugs/1940009

Title:
  Some options are missing from neutron.conf generated by "tox -e
  genconfig"

Status in neutron:
  Fix Released

Bug description:
  Looking at neutron.conf generated by "tox -e genconfig" I noticed the
  following items are missing from the generated neutron.conf.

  1)
  Some options of the oslo.service library like run_external_periodic_tasks

  2)
  Options of the oslo.reports library

  3)
  Options of the healthcheck middleware was added to api pipeline during 
Victoria cycle.
   https://review.opendev.org/c/openstack/neutron/+/724676

  This is because there are some oslo.config.opts endpoints missing from
  the config file passed to the oslo-config-generator command.

To manage notifications about this bug go to:
https://bugs.launchpad.net/neutron/+bug/1940009/+subscriptions


-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1940399] [NEW] vmware driver fails when setting vmware.maximum_objects to small value

2021-08-18 Thread Fabian Wiesel
Public bug reported:

Description
===

There are two functions in the code, which call (indirectly) RetrievePropertyEx 
with at most vmware.maximum_objects per result, but do not iterate over the 
results if there are more, nor does it 
By default the value is 100, which is usually sufficient, but if someone should 
set the value lower, it will cause the following problems.

Both are in the module nova.virt.vmwarapi.vm_util
1. get_all_cluster_mors
   The function is called to find the cluster reference for named cluster. If 
there are more clusters than maximum_objects, then there is a chance that the 
cluster won't be found by the driver despite being configured in the VCenter.

2. get_stats_from_cluster
   The function gets the statistics for all the hosts in a cluster. Since 
VSphere 7.0U1, the limit is 96 hosts per cluster, and quite likely the limit 
will increase past the default of 100, leaving the stats inaccurate. On top of 
it, the function does not call `cancel_retrieval` causing a leak in the VCenter.


Steps to reproduce
==

Take any VCenter version with more than one cluster and more than one host (I 
would suggest three or more), and any release of Nova configured to run against 
the VCenter.
I am not 100% according what the order is by which the clusters are returned, 
possibly in the order they have been created, or alphabetically.  I would 
suggest to create first a cluster-a, and then a cluster-b, and add the hosts to 
cluster-b. That reflects how our clusters are created (chronologically 
alphabetically sorted).

* Configure additionally
  [vmware]
  maximum_objects=1
* Try to start nova-compute

Expected result
===
The nova-compute service would start up and get the stats from all the hosts in 
the configured cluster.

Actual result
=

nova-compute fails to start with the error message:
> The specified cluster '' was not found in vCenter

If you have to clusters, and three hosts, then increasing
maximum_objects to two will get you around that failure, and will
trigger the second problem.

You can verify that by checking the resources of the nova-compute node,
which will report only the resources (CPUs,RAM...) of two of the ESXi-
hosts in the cluster.


Environment
===
1. Exact version of OpenStack you are running. 
   370830e944 Merge "libvirt: Enable 'vmcoreinfo' feature by default"

   As far as I can see, all Nova releases are affected by this behavior.

2. Which hypervisor did you use?
   VMware VSphere
   What's the version of that?
   7.0.1-17327586

3. Which networking type did you use?
   Neutron with NSX-T (https://github.com/sapcc/networking-nsx-t)

** Affects: nova
 Importance: Undecided
 Status: New

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1940399

Title:
  vmware driver fails when setting vmware.maximum_objects to small value

Status in OpenStack Compute (nova):
  New

Bug description:
  Description
  ===

  There are two functions in the code, which call (indirectly) 
RetrievePropertyEx with at most vmware.maximum_objects per result, but do not 
iterate over the results if there are more, nor does it 
  By default the value is 100, which is usually sufficient, but if someone 
should set the value lower, it will cause the following problems.

  Both are in the module nova.virt.vmwarapi.vm_util
  1. get_all_cluster_mors
 The function is called to find the cluster reference for named cluster. If 
there are more clusters than maximum_objects, then there is a chance that the 
cluster won't be found by the driver despite being configured in the VCenter.

  2. get_stats_from_cluster
 The function gets the statistics for all the hosts in a cluster. Since 
VSphere 7.0U1, the limit is 96 hosts per cluster, and quite likely the limit 
will increase past the default of 100, leaving the stats inaccurate. On top of 
it, the function does not call `cancel_retrieval` causing a leak in the VCenter.

  
  Steps to reproduce
  ==

  Take any VCenter version with more than one cluster and more than one host (I 
would suggest three or more), and any release of Nova configured to run against 
the VCenter.
  I am not 100% according what the order is by which the clusters are returned, 
possibly in the order they have been created, or alphabetically.  I would 
suggest to create first a cluster-a, and then a cluster-b, and add the hosts to 
cluster-b. That reflects how our clusters are created (chronologically 
alphabetically sorted).

  * Configure additionally
[vmware]
maximum_objects=1
  * Try to start nova-compute

  Expected result
  ===
  The nova-compute service would start up and get the stats from all the hosts 
in the configured cluster.

  Actual result
  =

  nova-compute fails to start with the error message:
  > The specif

[Yahoo-eng-team] [Bug 1940224] Re: UT neutron.tests.unit.services.ovn_l3.test_plugin.OVNL3ExtrarouteTests failing with neutron-lib from master

2021-08-18 Thread OpenStack Infra
Reviewed:  https://review.opendev.org/c/openstack/neutron-lib/+/804894
Committed: 
https://opendev.org/openstack/neutron-lib/commit/c6f576e66b577b9929f080f1ffd83cd3783794a4
Submitter: "Zuul (22348)"
Branch:master

commit c6f576e66b577b9929f080f1ffd83cd3783794a4
Author: Rodolfo Alonso Hernandez 
Date:   Tue Aug 17 16:39:38 2021 +

Use explicit "dict_or_nodata" validation for "external_gateway_info"

That API definition will be used during the policy check to create the
needed rule for each internal GW info element, specifically with the
external network fixed IPs. Only the owner of the external network or
the administrator can create fixed IPs on the external network.

Change-Id: Ia70485ed7713436dd6433ed39ae721500dbed7dd
Closes-Bug: #1940224


** Changed in: neutron
   Status: In Progress => Fix Released

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to neutron.
https://bugs.launchpad.net/bugs/1940224

Title:
  UT neutron.tests.unit.services.ovn_l3.test_plugin.OVNL3ExtrarouteTests
  failing with neutron-lib from master

Status in neutron:
  Fix Released

Bug description:
  Since 13.08.2021 one of our UT
  
neutron.tests.unit.services.ovn_l3.test_plugin.OVNL3ExtrarouteTests.test_router_create_with_gwinfo_ext_ip_non_admin
  is failing with neutron-lib from master.

  Failure example:
  
https://5095d1cf5e3173e1d222-5acdef5dc10478cee5291df1596ec66a.ssl.cf5.rackcdn.com/periodic/opendev.org/openstack/neutron/master/openstack-
  tox-py36-with-neutron-lib-master/292883e/testr_results.html

  Stacktrace:

  ft1.182: 
neutron.tests.unit.services.ovn_l3.test_plugin.OVNL3ExtrarouteTests.test_router_create_with_gwinfo_ext_ip_non_admintesttools.testresult.real._StringException:
 Traceback (most recent call last):
File "/home/zuul/src/opendev.org/openstack/neutron/neutron/tests/base.py", 
line 183, in func
  return f(self, *args, **kwargs)
File "/home/zuul/src/opendev.org/openstack/neutron/neutron/tests/base.py", 
line 183, in func
  return f(self, *args, **kwargs)
File 
"/home/zuul/src/opendev.org/openstack/neutron/neutron/tests/unit/extensions/test_l3.py",
 line 749, in test_router_create_with_gwinfo_ext_ip_non_admin
  self.assertEqual(exc.HTTPForbidden.code, res.status_int)
File 
"/home/zuul/src/opendev.org/openstack/neutron/.tox/py36/lib/python3.6/site-packages/testtools/testcase.py",
 line 393, in assertEqual
  self.assertThat(observed, matcher, message)
File 
"/home/zuul/src/opendev.org/openstack/neutron/.tox/py36/lib/python3.6/site-packages/testtools/testcase.py",
 line 480, in assertThat
  raise mismatch_error
  testtools.matchers._impl.MismatchError: 403 != 201

To manage notifications about this bug go to:
https://bugs.launchpad.net/neutron/+bug/1940224/+subscriptions


-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp