[Yahoo-eng-team] [Bug 2007986] Re: [neutron-lib][stable] Wallaby CI, tempest failing: unknown environment 'integrated-full'

2023-02-21 Thread yatin
Being fixed in Tempest
https://review.opendev.org/c/openstack/tempest/+/874704

** Changed in: neutron
   Status: New => Invalid

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to neutron.
https://bugs.launchpad.net/bugs/2007986

Title:
  [neutron-lib][stable] Wallaby CI, tempest failing: unknown environment
  'integrated-full'

Status in neutron:
  Invalid
Status in tempest:
  In Progress

Bug description:
  Neutron-lib Wallaby CI, job "tempest-full-py3" is failing with error:
controller | ERROR: unknown environment 'integrated-full'

  Logs:
  
https://storage.gra.cloud.ovh.net/v1/AUTH_dcaab5e32b234d56b626f72581e3644c/zuul_opendev_logs_cf8/874412/1/check/tempest-
  full-py3/cf87e9c/job-output.txt

To manage notifications about this bug go to:
https://bugs.launchpad.net/neutron/+bug/2007986/+subscriptions


-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 2007915] Re: Can nova-compute connect database directly?

2023-02-21 Thread shews
** Changed in: keystone
   Status: New => Invalid

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Identity (keystone).
https://bugs.launchpad.net/bugs/2007915

Title:
  Can nova-compute connect database directly?

Status in OpenStack Identity (keystone):
  Invalid

Bug description:
  Hello all.

  Now nova-compute get or update database via rabbitmq message to 
nova-conductor. If a region has a lot of nova-computes, this will increase the 
load of rabbitmq and nova-conductor.
  Can nova-compute connect to database directly? not via rabbitmq and 
nova-conductor. 

  Thank you.

To manage notifications about this bug go to:
https://bugs.launchpad.net/keystone/+bug/2007915/+subscriptions


-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 2008039] [NEW] Can nova-compute connect database directly?

2023-02-21 Thread shews
Public bug reported:

Hello all.

Now nova-compute get or update database via rabbitmq rpc message to 
nova-conductor. If a region has a lot of nova-computes, this will increase the 
load of rabbitmq and nova-conductor.
Can nova-compute connect to database directly? not via rabbitmq and 
nova-conductor.

Thank you.

** Affects: nova
 Importance: Undecided
 Status: New

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/2008039

Title:
  Can nova-compute connect database directly?

Status in OpenStack Compute (nova):
  New

Bug description:
  Hello all.

  Now nova-compute get or update database via rabbitmq rpc message to 
nova-conductor. If a region has a lot of nova-computes, this will increase the 
load of rabbitmq and nova-conductor.
  Can nova-compute connect to database directly? not via rabbitmq and 
nova-conductor.

  Thank you.

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/2008039/+subscriptions


-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 2008001] [NEW] [neutron-vpnaas] SQL error during the "vpnservice" creation

2023-02-21 Thread Rodolfo Alonso
Public bug reported:

During the VPN service creation, the Neutron API raises an SQL error
[1].

That happens when executing:
  $ openstack vpn service create vpn --router router

[1]https://paste.opendev.org/show/bnOLzHQNLRgkhoJMXw0g/

** Affects: neutron
 Importance: Undecided
 Status: New

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to neutron.
https://bugs.launchpad.net/bugs/2008001

Title:
  [neutron-vpnaas] SQL error during the "vpnservice" creation

Status in neutron:
  New

Bug description:
  During the VPN service creation, the Neutron API raises an SQL error
  [1].

  That happens when executing:
$ openstack vpn service create vpn --router router

  [1]https://paste.opendev.org/show/bnOLzHQNLRgkhoJMXw0g/

To manage notifications about this bug go to:
https://bugs.launchpad.net/neutron/+bug/2008001/+subscriptions


-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 2007992] [NEW] [FT] Random failures of ``test_good_address_allocation``

2023-02-21 Thread Rodolfo Alonso
Public bug reported:

Logs:
https://storage.gra.cloud.ovh.net/v1/AUTH_dcaab5e32b234d56b626f72581e3644c/zuul_opendev_logs_978/periodic/opendev.org/openstack/neutron/master/neutron-
functional/9789c89/testr_results.html

Error: https://paste.opendev.org/show/bSzMuP0DqW0yhBAhKBMF/

** Affects: neutron
 Importance: Undecided
 Status: New

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to neutron.
https://bugs.launchpad.net/bugs/2007992

Title:
  [FT] Random failures of ``test_good_address_allocation``

Status in neutron:
  New

Bug description:
  Logs:
  
https://storage.gra.cloud.ovh.net/v1/AUTH_dcaab5e32b234d56b626f72581e3644c/zuul_opendev_logs_978/periodic/opendev.org/openstack/neutron/master/neutron-
  functional/9789c89/testr_results.html

  Error: https://paste.opendev.org/show/bSzMuP0DqW0yhBAhKBMF/

To manage notifications about this bug go to:
https://bugs.launchpad.net/neutron/+bug/2007992/+subscriptions


-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 2007986] Re: [neutron-lib][stable] Wallaby CI, tempest failing: unknown environment 'integrated-full'

2023-02-21 Thread Slawek Kaplonski
** Tags added: tempest

** Also affects: tempest
   Importance: Undecided
   Status: New

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to neutron.
https://bugs.launchpad.net/bugs/2007986

Title:
  [neutron-lib][stable] Wallaby CI, tempest failing: unknown environment
  'integrated-full'

Status in neutron:
  New
Status in tempest:
  New

Bug description:
  Neutron-lib Wallaby CI, job "tempest-full-py3" is failing with error:
controller | ERROR: unknown environment 'integrated-full'

  Logs:
  
https://storage.gra.cloud.ovh.net/v1/AUTH_dcaab5e32b234d56b626f72581e3644c/zuul_opendev_logs_cf8/874412/1/check/tempest-
  full-py3/cf87e9c/job-output.txt

To manage notifications about this bug go to:
https://bugs.launchpad.net/neutron/+bug/2007986/+subscriptions


-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 2007986] [NEW] [neutron-lib][stable] Wallaby CI, tempest failing: unknown environment 'integrated-full'

2023-02-21 Thread Rodolfo Alonso
Public bug reported:

Neutron-lib Wallaby CI, job "tempest-full-py3" is failing with error:
  controller | ERROR: unknown environment 'integrated-full'

Logs:
https://storage.gra.cloud.ovh.net/v1/AUTH_dcaab5e32b234d56b626f72581e3644c/zuul_opendev_logs_cf8/874412/1/check/tempest-
full-py3/cf87e9c/job-output.txt

** Affects: neutron
 Importance: Critical
 Status: New

** Changed in: neutron
   Importance: Undecided => Critical

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to neutron.
https://bugs.launchpad.net/bugs/2007986

Title:
  [neutron-lib][stable] Wallaby CI, tempest failing: unknown environment
  'integrated-full'

Status in neutron:
  New

Bug description:
  Neutron-lib Wallaby CI, job "tempest-full-py3" is failing with error:
controller | ERROR: unknown environment 'integrated-full'

  Logs:
  
https://storage.gra.cloud.ovh.net/v1/AUTH_dcaab5e32b234d56b626f72581e3644c/zuul_opendev_logs_cf8/874412/1/check/tempest-
  full-py3/cf87e9c/job-output.txt

To manage notifications about this bug go to:
https://bugs.launchpad.net/neutron/+bug/2007986/+subscriptions


-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 2007985] [NEW] [ovn-octavia-provider] restore the member provisioning status to NO_MONITOR after delete the HM

2023-02-21 Thread Fernando Royo
Public bug reported:

When a HM is creating over a pool, the members associated to it change
theis provisioning_status to ONLINE, and if the health checks packets
detects an error on communication it will move to ERROR. When the HM is
deleted the provisioning status of those members should come back to
NO_MONITOR, but now is keeping on the last status (ONLINE or ERROR)

** Affects: neutron
 Importance: Undecided
 Assignee: Fernando Royo (froyoredhat)
 Status: New


** Tags: ovn-octavia-provider

** Changed in: neutron
 Assignee: (unassigned) => Fernando Royo (froyoredhat)

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to neutron.
https://bugs.launchpad.net/bugs/2007985

Title:
  [ovn-octavia-provider] restore the member provisioning status to
  NO_MONITOR after delete the HM

Status in neutron:
  New

Bug description:
  When a HM is creating over a pool, the members associated to it change
  theis provisioning_status to ONLINE, and if the health checks packets
  detects an error on communication it will move to ERROR. When the HM
  is deleted the provisioning status of those members should come back
  to NO_MONITOR, but now is keeping on the last status (ONLINE or ERROR)

To manage notifications about this bug go to:
https://bugs.launchpad.net/neutron/+bug/2007985/+subscriptions


-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 2007982] [NEW] keystone role cache misbehaving in HA setup

2023-02-21 Thread ebl...@nde.ag
Public bug reported:

Following up on two mailing list reports [1][2] which seem to have the same 
root cause. 
In a HA setup with 3 control nodes (Victoria, baremetal) terraform is used to 
deploy lots of different k8s clusters (and other stuff). We noticed keystone 
errors when a project is purged with terraform (cleanly) and a redeployment of 
the same project (with the same name) is started immediately after that. We did 
some tests to find out which exact keystone cache it is and it seems to be the 
role cache (default 600 seconds) which leads to an error in terraform, it 
reports that the project was not found and refers to the previous ID of the 
project which is already deleted from the database during the project cleanup.
The same deployment works in an identical cloud version except with only one 
control node, it just works although the cache is enabled as well.
I already tried to reduce the cache_time to 30 seconds but that doesn't help 
(although it takes more than 30 seconds until terraform is ready after the 
prechecks). I also disabled the role cache entirely which helps with the faster 
redeployment but the downside of disabling it leads to significantly longer 
response times when using the dashboard or querying the APIs.
Is there any way to tune the role cache in a way so we could have both a 
reasonable performance as well as being able to redeploy projects without a 
"sleep 600"?

Storage back end is Ceph (Pacific), keystone versions are:

control01:~ # rpm -qa | grep keystone
python3-keystonemiddleware-9.1.0-lp152.3.20.noarch
python3-keystone-18.0.1~dev11-lp152.1.21.noarch
python3-keystoneauth1-4.2.1-lp152.3.19.noarch
python3-keystoneclient-4.1.0-lp152.5.2.noarch
openstack-keystone-18.0.1~dev11-lp152.1.21.noarch

[1] 
https://lists.openstack.org/pipermail/openstack-discuss/2022-November/031122.html
[2] 
https://lists.openstack.org/pipermail/openstack-discuss/2023-February/032258.html

** Affects: keystone
 Importance: Undecided
 Status: New

** Project changed: nova => keystone

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/2007982

Title:
  keystone role cache misbehaving in HA setup

Status in OpenStack Identity (keystone):
  New

Bug description:
  Following up on two mailing list reports [1][2] which seem to have the same 
root cause. 
  In a HA setup with 3 control nodes (Victoria, baremetal) terraform is used to 
deploy lots of different k8s clusters (and other stuff). We noticed keystone 
errors when a project is purged with terraform (cleanly) and a redeployment of 
the same project (with the same name) is started immediately after that. We did 
some tests to find out which exact keystone cache it is and it seems to be the 
role cache (default 600 seconds) which leads to an error in terraform, it 
reports that the project was not found and refers to the previous ID of the 
project which is already deleted from the database during the project cleanup.
  The same deployment works in an identical cloud version except with only one 
control node, it just works although the cache is enabled as well.
  I already tried to reduce the cache_time to 30 seconds but that doesn't help 
(although it takes more than 30 seconds until terraform is ready after the 
prechecks). I also disabled the role cache entirely which helps with the faster 
redeployment but the downside of disabling it leads to significantly longer 
response times when using the dashboard or querying the APIs.
  Is there any way to tune the role cache in a way so we could have both a 
reasonable performance as well as being able to redeploy projects without a 
"sleep 600"?

  Storage back end is Ceph (Pacific), keystone versions are:

  control01:~ # rpm -qa | grep keystone
  python3-keystonemiddleware-9.1.0-lp152.3.20.noarch
  python3-keystone-18.0.1~dev11-lp152.1.21.noarch
  python3-keystoneauth1-4.2.1-lp152.3.19.noarch
  python3-keystoneclient-4.1.0-lp152.5.2.noarch
  openstack-keystone-18.0.1~dev11-lp152.1.21.noarch

  [1] 
https://lists.openstack.org/pipermail/openstack-discuss/2022-November/031122.html
  [2] 
https://lists.openstack.org/pipermail/openstack-discuss/2023-February/032258.html

To manage notifications about this bug go to:
https://bugs.launchpad.net/keystone/+bug/2007982/+subscriptions


-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 2006734] Re: [OVN] OVN trunk driver is not bumping the OVN LSP "external_ids:revision_number"

2023-02-21 Thread OpenStack Infra
Reviewed:  https://review.opendev.org/c/openstack/neutron/+/873296
Committed: 
https://opendev.org/openstack/neutron/commit/fce516e3461054716ff2d4623c240d3e006acbeb
Submitter: "Zuul (22348)"
Branch:master

commit fce516e3461054716ff2d4623c240d3e006acbeb
Author: Rodolfo Alonso Hernandez 
Date:   Thu Feb 9 16:36:14 2023 +0100

[OVN] Bump the port revision number in trunk driver

When the subport binding information is updated in the OVN trunk
driver, the OVN revision number should be updated too, same as
in other ``ovn_client`` method, for example.

Closes-Bug: #2006734
Change-Id: Icda34910ea7fe308814e0cc60999b465d3540b67


** Changed in: neutron
   Status: In Progress => Fix Released

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to neutron.
https://bugs.launchpad.net/bugs/2006734

Title:
  [OVN] OVN trunk driver is not bumping the OVN LSP
  "external_ids:revision_number"

Status in neutron:
  Fix Released

Bug description:
  The OVN trunk driver is not bumping the OVN LSP
  "external_ids:revision_number" after updating the Neutron DB register.
  The methods ``OVNTrunkHandler._set_binding_profile`` and
  ``OVNTrunkHandler._unset_binding_profile`` are called from
  ``_set_sub_ports`` and ``_unset_sub_ports`` respectively. These
  methods create a Neutron DB and a OVN DB transactions in the same with
  clause. Inside both transactions, the port register (Neutron DB and
  OVN DB) is modified. However, the OVN register revision_number is not
  bumped consequently. That forces the maintenance task to review the
  OVN register and force an update.

To manage notifications about this bug go to:
https://bugs.launchpad.net/neutron/+bug/2006734/+subscriptions


-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 2007968] [NEW] Flavors may not meet the image minimum requirement when resize

2023-02-21 Thread Kodanevhy Zhou
Public bug reported:

Description
===
When resize instance, the flavors returned may not meet the image minimum 
memory requirement, resizing instance ignores the minimum memory limit of the 
image, which may cause the resizing be successfully, but the instance fails to 
start because the memory is too small to run the system.

Steps to reproduce
==
1.create an instance with image min_ram 4096
2.resize the instance
3.watch the returned flavors

Expected result
===
do not include the flavors which memory less than 4096.

Actual result
=
returned all of the visible flavors.

Environment
===

Logs & Configs
==

** Affects: horizon
 Importance: Undecided
 Status: New

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Dashboard (Horizon).
https://bugs.launchpad.net/bugs/2007968

Title:
  Flavors may not meet the image minimum requirement when resize

Status in OpenStack Dashboard (Horizon):
  New

Bug description:
  Description
  ===
  When resize instance, the flavors returned may not meet the image minimum 
memory requirement, resizing instance ignores the minimum memory limit of the 
image, which may cause the resizing be successfully, but the instance fails to 
start because the memory is too small to run the system.

  Steps to reproduce
  ==
  1.create an instance with image min_ram 4096
  2.resize the instance
  3.watch the returned flavors

  Expected result
  ===
  do not include the flavors which memory less than 4096.

  Actual result
  =
  returned all of the visible flavors.

  Environment
  ===

  Logs & Configs
  ==

To manage notifications about this bug go to:
https://bugs.launchpad.net/horizon/+bug/2007968/+subscriptions


-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 2007959] [NEW] Package "python-devel [platform:rpm test]" defined by bindep fails on RHEL 8

2023-02-21 Thread Jorge San Emeterio
Public bug reported:

Description
===
On "stable/train" at "bindep.txt", line "python-devel [platform:rpm test]" 
should be "python3-devel [platform:rpm test]", otherwise during CI automation 
tools will fail while trying to install the package on RHEL 8. Versions after 
train have performed this change, so I expect this is not the first time the 
problem has happened.

Steps to reproduce
==
1.- On stock RHEL 8.4, run: dnf install python-devel
2.- RHEL will complain that the package is unknown.

Expected result
===
RHEL is capable of installing the "python36-devel" package and its dependencies 
through the "python3-devel" alias.

Actual result
=
RHEL does not find the "python-devel" package.

Environment
===
RHEL 8.4
Python 3.6

** Affects: nova
 Importance: Undecided
 Assignee: Jorge San Emeterio (jsanemet)
 Status: In Progress

** Changed in: nova
 Assignee: (unassigned) => Jorge San Emeterio (jsanemet)

** Changed in: nova
   Status: New => In Progress

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/2007959

Title:
  Package "python-devel [platform:rpm test]" defined by bindep fails on
  RHEL 8

Status in OpenStack Compute (nova):
  In Progress

Bug description:
  Description
  ===
  On "stable/train" at "bindep.txt", line "python-devel [platform:rpm test]" 
should be "python3-devel [platform:rpm test]", otherwise during CI automation 
tools will fail while trying to install the package on RHEL 8. Versions after 
train have performed this change, so I expect this is not the first time the 
problem has happened.

  Steps to reproduce
  ==
  1.- On stock RHEL 8.4, run: dnf install python-devel
  2.- RHEL will complain that the package is unknown.

  Expected result
  ===
  RHEL is capable of installing the "python36-devel" package and its 
dependencies through the "python3-devel" alias.

  Actual result
  =
  RHEL does not find the "python-devel" package.

  Environment
  ===
  RHEL 8.4
  Python 3.6

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/2007959/+subscriptions


-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 2007938] [NEW] [DHCP] Metadata IPv6 duplicated address when setting the interface

2023-02-21 Thread Rodolfo Alonso
Public bug reported:

When setting the IPv6 address for the metadata interface [2], it could
happen that the IPv6 address can be considered as duplicated [2].

I can't provide a way to replicate this issue as it is happening
randomly. The IPv6 address in the DHCP namespace is set as "dadfailed
tentative".

[2]https://github.com/openstack/neutron/blob/8455edda46b4a465e2f184b59ad31476ce79c6c4/neutron/agent/metadata/driver.py#L244-L246
[1]https://paste.opendev.org/show/bXDAHSHpvEAIHwdPjxrB/

** Affects: neutron
 Importance: Undecided
 Assignee: Rodolfo Alonso (rodolfo-alonso-hernandez)
 Status: New

** Changed in: neutron
 Assignee: (unassigned) => Rodolfo Alonso (rodolfo-alonso-hernandez)

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to neutron.
https://bugs.launchpad.net/bugs/2007938

Title:
  [DHCP] Metadata IPv6 duplicated address when setting the interface

Status in neutron:
  New

Bug description:
  When setting the IPv6 address for the metadata interface [2], it could
  happen that the IPv6 address can be considered as duplicated [2].

  I can't provide a way to replicate this issue as it is happening
  randomly. The IPv6 address in the DHCP namespace is set as "dadfailed
  tentative".

  
[2]https://github.com/openstack/neutron/blob/8455edda46b4a465e2f184b59ad31476ce79c6c4/neutron/agent/metadata/driver.py#L244-L246
  [1]https://paste.opendev.org/show/bXDAHSHpvEAIHwdPjxrB/

To manage notifications about this bug go to:
https://bugs.launchpad.net/neutron/+bug/2007938/+subscriptions


-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 2007922] [NEW] Cleanup pending instances in "building" state

2023-02-21 Thread ebl...@nde.ag
Public bug reported:

Following up on the ML thread [1], it was recommended to create a bug report.
After a network issue in a Victoria cluster (3 control nodes in HA mode, 26 
compute nodes) some instance builds were interrupted. Some of them could be 
cleaned up with 'openstack server delete' but two of them can not. They already 
have a mapping but can not be removed (or "reset-state") by nova. Those are 
both amphora instances from octavia:

control01:~ # openstack server list --project service -c ID -c Name -c Status 
-f value | grep BUILD
0453a7e5-e4f9-419b-ad71-d837a20ef6bb 
amphora-0ee32901-0c59-4752-8253-35b66da176ea BUILD
dc8cdc3a-f6b2-469b-af6f-ba2aa130ea9b 
amphora-4990a47b-fe8a-431a-90ec-5ac2368a5251 BUILD

control01:~ # openstack server delete 
amphora-0ee32901-0c59-4752-8253-35b66da176ea
No server with a name or ID of
'amphora-0ee32901-0c59-4752-8253-35b66da176ea' exists.

control01:~ # openstack server show 0453a7e5-e4f9-419b-ad71-d837a20ef6bb
ERROR (CommandError): No server with a name or ID of
'0453a7e5-e4f9-419b-ad71-d837a20ef6bb' exists.

The database tables referring to the UUID
0453a7e5-e4f9-419b-ad71-d837a20ef6bb are these:

nova_cell0/instance_id_mappings.ibd
nova_cell0/instance_info_caches.ibd
nova_cell0/instance_extra.ibd
nova_cell0/instances.ibd
nova_cell0/instance_system_metadata.ibd
octavia/amphora.ibd
nova_api/instance_mappings.ibd
nova_api/request_specs.ibd

I can provide both debug logs and database queries, just let me know
what exactly is required.

The storage back end is ceph (Pacific), we use neutron with OpenVSwitch,
the exact nova versions are:

control01:~ # rpm -qa | grep nova
openstack-nova-conductor-22.2.2~dev15-lp152.1.25.noarch
openstack-nova-api-22.2.2~dev15-lp152.1.25.noarch
openstack-nova-novncproxy-22.2.2~dev15-lp152.1.25.noarch
python3-novaclient-17.2.0-lp152.3.2.noarch
openstack-nova-scheduler-22.2.2~dev15-lp152.1.25.noarch
openstack-nova-22.2.2~dev15-lp152.1.25.noarch
python3-nova-22.2.2~dev15-lp152.1.25.noarch

[1] https://lists.openstack.org/pipermail/openstack-
discuss/2023-February/032308.html

** Affects: nova
 Importance: Undecided
 Status: New

** Description changed:

  Following up on the ML thread [1], it was recommended to create a bug report.
  After a network issue in a Victoria cluster (3 control nodes in HA mode, 26 
compute nodes) some instance builds were interrupted. Some of them could be 
cleaned up with 'openstack server delete' but two of them can not. They already 
have a mapping but can not be removed (or "reset-state") by nova. Those are 
both amphora instances from octavia:
  
  control01:~ # openstack server list --project service -c ID -c Name -c Status 
-f value | grep BUILD
  0453a7e5-e4f9-419b-ad71-d837a20ef6bb 
amphora-0ee32901-0c59-4752-8253-35b66da176ea BUILD
  dc8cdc3a-f6b2-469b-af6f-ba2aa130ea9b 
amphora-4990a47b-fe8a-431a-90ec-5ac2368a5251 BUILD
  
  control01:~ # openstack server delete 
amphora-0ee32901-0c59-4752-8253-35b66da176ea
- No server with a name or ID of  
+ No server with a name or ID of
  'amphora-0ee32901-0c59-4752-8253-35b66da176ea' exists.
  
  control01:~ # openstack server show 0453a7e5-e4f9-419b-ad71-d837a20ef6bb
- ERROR (CommandError): No server with a name or ID of  
+ ERROR (CommandError): No server with a name or ID of
  '0453a7e5-e4f9-419b-ad71-d837a20ef6bb' exists.
  
- The database tables referring to the UUID  
+ The database tables referring to the UUID
  0453a7e5-e4f9-419b-ad71-d837a20ef6bb are these:
  
  nova_cell0/instance_id_mappings.ibd
  nova_cell0/instance_info_caches.ibd
  nova_cell0/instance_extra.ibd
  nova_cell0/instances.ibd
  nova_cell0/instance_system_metadata.ibd
  octavia/amphora.ibd
  nova_api/instance_mappings.ibd
  nova_api/request_specs.ibd
  
  I can provide both debug logs and database queries, just let me know
  what exactly is required.
  
+ The storage back end is ceph (Pacific), we use neutron with OpenVSwitch,
+ the exact nova versions are:
+ 
+ control01:~ # rpm -qa | grep nova
+ openstack-nova-conductor-22.2.2~dev15-lp152.1.25.noarch
+ openstack-nova-api-22.2.2~dev15-lp152.1.25.noarch
+ openstack-nova-novncproxy-22.2.2~dev15-lp152.1.25.noarch
+ python3-novaclient-17.2.0-lp152.3.2.noarch
+ openstack-nova-scheduler-22.2.2~dev15-lp152.1.25.noarch
+ openstack-nova-22.2.2~dev15-lp152.1.25.noarch
+ python3-nova-22.2.2~dev15-lp152.1.25.noarch
+ 
  [1] https://lists.openstack.org/pipermail/openstack-
  discuss/2023-February/032308.html

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/2007922

Title:
  Cleanup pending instances in "building" state

Status in OpenStack Compute (nova):
  New

Bug description:
  Following up on the ML thread [1], it was recommended to create a bug report.
  After a network issue in a Victoria cluster (3 control nodes in HA mode, 26 
compute nodes) some instance builds were interrupted. Some of them could 

[Yahoo-eng-team] [Bug 2007924] [NEW] stevedore always shows error if boto3 is not installed

2023-02-21 Thread Takashi Kajinami
Public bug reported:

Currently stevedore always dump the following error in case boto3 is not
installed in the system.

ERROR stevedore.extension [-] Could not load 'glance.store.s3.Store': No module 
named 'boto3': ModuleNotFoundError: No module named 'boto3'
ERROR stevedore.extension [-] Could not load 's3': No module named 'boto3': 
ModuleNotFoundError: No module named 'boto3'

This error is red herring because missing boto3 does not harm unless s3
backend is actually used.

The other stores such as swift store ignores missing library during
loading drivers but fails in case swift store is actually requested.
It'd be helpful to follow that strategy for s3 backend to avoid
confusing error.

** Affects: glance
 Importance: Undecided
 Assignee: Takashi Kajinami (kajinamit)
 Status: New

** Changed in: glance
 Assignee: (unassigned) => Takashi Kajinami (kajinamit)

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to Glance.
https://bugs.launchpad.net/bugs/2007924

Title:
  stevedore always shows error if boto3 is not installed

Status in Glance:
  New

Bug description:
  Currently stevedore always dump the following error in case boto3 is
  not installed in the system.

  ERROR stevedore.extension [-] Could not load 'glance.store.s3.Store': No 
module named 'boto3': ModuleNotFoundError: No module named 'boto3'
  ERROR stevedore.extension [-] Could not load 's3': No module named 'boto3': 
ModuleNotFoundError: No module named 'boto3'

  This error is red herring because missing boto3 does not harm unless
  s3 backend is actually used.

  The other stores such as swift store ignores missing library during
  loading drivers but fails in case swift store is actually requested.
  It'd be helpful to follow that strategy for s3 backend to avoid
  confusing error.

To manage notifications about this bug go to:
https://bugs.launchpad.net/glance/+bug/2007924/+subscriptions


-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp