[Yahoo-eng-team] [Bug 1840862] Re: Missing description of hacking rule N363 in HACKING.rst

2019-08-21 Thread OpenStack Infra
Reviewed:  https://review.opendev.org/677589
Committed: 
https://git.openstack.org/cgit/openstack/nova/commit/?id=5ccdbc7189640d02cc17e2156c635e8f7eafb02e
Submitter: Zuul
Branch:master

commit 5ccdbc7189640d02cc17e2156c635e8f7eafb02e
Author: Takashi NATSUME 
Date:   Wed Aug 21 10:34:51 2019 +0900

Fix missing rule description in HACKING.rst

The description of N363 hacking rule is missing
in HACKING.rst.
Add the description.

Change-Id: I036a48612fcd256df4ccbd2ebba814bf3ed7a1c2
Closes-Bug: #1840862


** Changed in: nova
   Status: In Progress => Fix Released

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1840862

Title:
  Missing description of hacking rule N363 in HACKING.rst

Status in OpenStack Compute (nova):
  Fix Released

Bug description:
  The description of hacking rule N363 is missing in HACKING.rst.

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1840862/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1826695] Re: [L3][QoS] cache does not removed when router is down or deleted

2019-08-21 Thread OpenStack Infra
Reviewed:  https://review.opendev.org/656105
Committed: 
https://git.openstack.org/cgit/openstack/neutron/commit/?id=78629e0d3780af4a8a8cc1b5b0762e0bc8a48f1f
Submitter: Zuul
Branch:master

commit 78629e0d3780af4a8a8cc1b5b0762e0bc8a48f1f
Author: LIU Yulong 
Date:   Wed Mar 27 08:40:31 2019 +0800

Remove L3 IP QoS cache when router is down

When router admin-state is down or removed, fip-qos and gateway-ip-qos
extension should delete the router IPs' QoS rate limit cache. Then if
the router is up again the router floating IPs QoS can be reconfigured.
This patch achives these:
1. make sure floating IP or gateway IP QoS cache removed.
2. floating IP QoS can be re-configured to snat device when router
   doing admin_state down/up.

Closes-Bug: #1826695
Change-Id: I24fcecd9686ad17fa50093bb8bccab0d6c711298


** Changed in: neutron
   Status: In Progress => Fix Released

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to neutron.
https://bugs.launchpad.net/bugs/1826695

Title:
  [L3][QoS] cache does not removed when router is down or deleted

Status in neutron:
  Fix Released

Bug description:
  [L3][QoS] cache does not removed when router is down or deleted

  
  ENV: stable/queens (master has the same issue)

  When router is DOWN or deleted, the L3 IP QoS extensions (floating IP and 
gateway IP) should remove the QoS cache.
  Otherwise, if the router is UP again, the floating IP QoS rules may not set 
to the devices back because the bandwidth value does not change.

To manage notifications about this bug go to:
https://bugs.launchpad.net/neutron/+bug/1826695/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1840930] Re: Networking service in neutron - install guide says to configure nova with [neutron]/url which is deprecated

2019-08-21 Thread OpenStack Infra
Reviewed:  https://review.opendev.org/677810
Committed: 
https://git.openstack.org/cgit/openstack/neutron/commit/?id=e8653f28dee5044b943c4ca1ee0329e0bd7b34d2
Submitter: Zuul
Branch:master

commit e8653f28dee5044b943c4ca1ee0329e0bd7b34d2
Author: Matt Riedemann 
Date:   Wed Aug 21 16:09:09 2019 -0400

doc: remove deprecated [neutron]/url from compute install guide

The nova [neutron]/url config option was deprecated in Queens [1]
and is being removed in Train [2]. The neutron install guide
sections about configuring compute to work with neutron were still
using the url option so this change removes them. There are a few
things to note here:

1. The url option functionality is replaced with the endpoint_override
   option from keystoneauth1 but we don't really want users using that
   unless there is a real need. One of the main reasons for moving the
   nova configuration to use keystoneauth1 was so that the network
   service endpoint can be looked up via KSA dynamically based on the
   configurable interfaces (public, internal, private) and service types
   authority so the endpoint URL will just be pulled from the service
   catalog. That means not having to hard-code the endpoint URL in nova
   config which makes deployment and config management simpler. As such,
   the url option removed in the install guide here is not replaced with
   the endpoint_override option.

2. Following on #1, the install guide portion about the nova/neutron config
   is updated with a link back to the nova config guide for the full set
   of options in case an operator needs to tweak the config, e.g. to set
   valid_interfaces or endpoint_override because the KSA defaults don't work
   for their deployment.

3. With the old url option, if region_name was not specified, nova would
   default to 'RegionOne'. That is not the case if not using the url option
   so we leave the region_one config in the install guide example, otherwise
   region_name would default to None.

[1] I41724a612a5f3eabd504f3eaa9d2f9d141ca3f69
[2] I6c068a84c4c0bd88f088f9328d7897bfc1f843f1

Change-Id: I30445edeb8509330571db28c7d61dd63886e9e61
Closes-Bug: #1840930


** Changed in: neutron
   Status: In Progress => Fix Released

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to neutron.
https://bugs.launchpad.net/bugs/1840930

Title:
  Networking service in neutron - install guide says to configure nova
  with [neutron]/url which is deprecated

Status in neutron:
  Fix Released

Bug description:
  - [x] This doc is inaccurate in this way:

  The [neutron]/url option
  https://docs.openstack.org/nova/latest/configuration/config.html#neutron.url
  in nova has been deprecated since the Queens release and is being
  removed in Train. The neutron/compute config guide in the neutron
  install guides still says to use the url option though. Since Queens
  when nova started using KSA adapters for working with neutron config:

  https://review.opendev.org/#/c/509892/

  I think we want to avoid configuring the [neutron] section in
  nova.conf with url or endpoint_override but install rely on KSA to use
  the service types authority to find the endpoint to use based on
  service name/type and interface, in other words things should just
  work without needing to explicitly define an endpoint url to use for
  nova talking to neutron - nova can go through KSA and the service
  catalog to get the endpoint it needs.

  ---
  Release: 14.1.0.dev665 on 2017-06-30 05:58:47
  SHA: 490471ebd3ac56d0cee164b9c1c1211687e49437
  Source: https://opendev.org/openstack/neutron/src/doc/source/install/index.rst
  URL: https://docs.openstack.org/neutron/latest/install/

To manage notifications about this bug go to:
https://bugs.launchpad.net/neutron/+bug/1840930/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1840978] [NEW] nova-manage commands with unexpected errors returning 1 conflict with expected cases of 1 for flow control

2019-08-21 Thread Matt Riedemann
Public bug reported:

The archive_deleted_rows command returns 1 meaning some records were
archived and the code documents that if automating and not using
--until-complete, you should keep going while you get rc=1 until you get
rc=0:

https://github.com/openstack/nova/blob/0bf81cfe73340ba5cfd9cf44a38905014ba780f0/nova/cmd/manage.py#L505

The problem is if some unexpected error happens, let's say there is a
TypeError in the code or something, the command will also return 1:

https://github.com/openstack/nova/blob/0bf81cfe73340ba5cfd9cf44a38905014ba780f0/nova/cmd/manage.py#L2625

That unexpected error should probably be a 255 which generally means a
command failed in some unexpected way. There might be other nova-manage
commands that return 1 for flow control as well.

Note that changing the "unexpected error" code from 1 to 255 is an
upgrade impacting change worth a release note.

** Affects: nova
 Importance: Low
 Status: Triaged


** Tags: nova-manage

** Tags added: nova-manage

** Changed in: nova
   Importance: Undecided => Low

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1840978

Title:
  nova-manage commands with unexpected errors returning 1 conflict with
  expected cases of 1 for flow control

Status in OpenStack Compute (nova):
  Triaged

Bug description:
  The archive_deleted_rows command returns 1 meaning some records were
  archived and the code documents that if automating and not using
  --until-complete, you should keep going while you get rc=1 until you
  get rc=0:

  
https://github.com/openstack/nova/blob/0bf81cfe73340ba5cfd9cf44a38905014ba780f0/nova/cmd/manage.py#L505

  The problem is if some unexpected error happens, let's say there is a
  TypeError in the code or something, the command will also return 1:

  
https://github.com/openstack/nova/blob/0bf81cfe73340ba5cfd9cf44a38905014ba780f0/nova/cmd/manage.py#L2625

  That unexpected error should probably be a 255 which generally means a
  command failed in some unexpected way. There might be other nova-
  manage commands that return 1 for flow control as well.

  Note that changing the "unexpected error" code from 1 to 255 is an
  upgrade impacting change worth a release note.

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1840978/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1840979] [NEW] [L2] [opinion] update the port DB status directly in agent-side

2019-08-21 Thread LIU Yulong
Public bug reported:

When ovs-agent done processing the port, it will call neutron-server to make 
some DB update.
Especially when restart the ovs-agent, all ports in one agent will do such RPC 
and DB update again to make port status consistent. When a large number of 
concurrent agent restart happen, neutron-server may not work fine.
So how about making the following DB updating locally in neutron agent side 
directly? It may have some mechanism driver notification, IMO, this can also be 
done in agent-side.

def update_device_down(self, context, device, agent_id, host=None):
cctxt = self.client.prepare()
return cctxt.call(context, 'update_device_down', device=device,
  agent_id=agent_id, host=host)

def update_device_up(self, context, device, agent_id, host=None):
cctxt = self.client.prepare()
return cctxt.call(context, 'update_device_up', device=device,
  agent_id=agent_id, host=host)

def update_device_list(self, context, devices_up, devices_down,
ret = cctxt.call(context, 'update_device_list',

** Affects: neutron
 Importance: Undecided
 Status: New

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to neutron.
https://bugs.launchpad.net/bugs/1840979

Title:
  [L2] [opinion] update the port DB status directly in agent-side

Status in neutron:
  New

Bug description:
  When ovs-agent done processing the port, it will call neutron-server to make 
some DB update.
  Especially when restart the ovs-agent, all ports in one agent will do such 
RPC and DB update again to make port status consistent. When a large number of 
concurrent agent restart happen, neutron-server may not work fine.
  So how about making the following DB updating locally in neutron agent side 
directly? It may have some mechanism driver notification, IMO, this can also be 
done in agent-side.

  def update_device_down(self, context, device, agent_id, host=None):
  cctxt = self.client.prepare()
  return cctxt.call(context, 'update_device_down', device=device,
agent_id=agent_id, host=host)

  def update_device_up(self, context, device, agent_id, host=None):
  cctxt = self.client.prepare()
  return cctxt.call(context, 'update_device_up', device=device,
agent_id=agent_id, host=host)

  def update_device_list(self, context, devices_up, devices_down,
  ret = cctxt.call(context, 'update_device_list',

To manage notifications about this bug go to:
https://bugs.launchpad.net/neutron/+bug/1840979/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1840967] [NEW] nova-next job does not fail when 'nova-manage db purge' fails

2019-08-21 Thread melanie witt
Public bug reported:

Happened upon this while working on another patch to add more testing to
our post_test_hook.sh script, excerpt from the log [1]:

+ /usr/local/bin/nova-manage db purge --all --verbose --all-cells
+ RET=3
+ [[ 3 -eq 0 ]]
+ echo Purge failed with result 3
Purge failed with result 3
+ return 3
+ set -e
+ set +x
WARNING: setting legacy OS_TENANT_NAME to support cli tools.
+ /opt/stack/nova/gate/post_test_hook.sh:main:54 :   echo 'Verifying that 
instances were archived from all cells'
Verifying that instances were archived from all cells
++ /opt/stack/nova/gate/post_test_hook.sh:main:55 :   openstack server list 
--deleted --all-projects -c ID -f value
+ /opt/stack/nova/gate/post_test_hook.sh:main:55 :   
deleted_servers='e4727a33-796e-4173-b369-24d7ee45d7fd
b213a354-0830-4cc3-abf7-e9dd068cefa9
33569d93-d7b6-4a92-825e-f36e972722db
521e4a84-c313-433e-8cc7-6d66c821d78c

Because of a bug in my WIP patch, the purge command failed, but the job
continued to run and didn't fail at that point because the 'nova-manage
db purge' command comes before the 'set -e' command [that makes the
script exit with any non-zero return value].

So, we need to move the purge command after 'set -e'. Note that we
should *not* move the archive command though, because during its
intermediate runs, it is expected to return 1, and we don't want to fail
the job when that happens. The archive_deleted_rows function does its
own explicit exiting in the case of actual failures.

[1] https://object-storage-ca-
ymq-1.vexxhost.net/v1/86bbbcfa8ad043109d2d7af530225c72/logs_40/672840/8/check
/nova-next/9d13cfb/ara-report/result/d13f888f-d187-4c3b-b5ab-
9326f611e534/

** Affects: nova
 Importance: Undecided
 Assignee: melanie witt (melwitt)
 Status: New


** Tags: testing

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1840967

Title:
  nova-next job does not fail when 'nova-manage db purge' fails

Status in OpenStack Compute (nova):
  New

Bug description:
  Happened upon this while working on another patch to add more testing
  to our post_test_hook.sh script, excerpt from the log [1]:

  + /usr/local/bin/nova-manage db purge --all --verbose --all-cells
  + RET=3
  + [[ 3 -eq 0 ]]
  + echo Purge failed with result 3
  Purge failed with result 3
  + return 3
  + set -e
  + set +x
  WARNING: setting legacy OS_TENANT_NAME to support cli tools.
  + /opt/stack/nova/gate/post_test_hook.sh:main:54 :   echo 'Verifying that 
instances were archived from all cells'
  Verifying that instances were archived from all cells
  ++ /opt/stack/nova/gate/post_test_hook.sh:main:55 :   openstack server list 
--deleted --all-projects -c ID -f value
  + /opt/stack/nova/gate/post_test_hook.sh:main:55 :   
deleted_servers='e4727a33-796e-4173-b369-24d7ee45d7fd
  b213a354-0830-4cc3-abf7-e9dd068cefa9
  33569d93-d7b6-4a92-825e-f36e972722db
  521e4a84-c313-433e-8cc7-6d66c821d78c

  Because of a bug in my WIP patch, the purge command failed, but the
  job continued to run and didn't fail at that point because the 'nova-
  manage db purge' command comes before the 'set -e' command [that makes
  the script exit with any non-zero return value].

  So, we need to move the purge command after 'set -e'. Note that we
  should *not* move the archive command though, because during its
  intermediate runs, it is expected to return 1, and we don't want to
  fail the job when that happens. The archive_deleted_rows function does
  its own explicit exiting in the case of actual failures.

  [1] https://object-storage-ca-
  ymq-1.vexxhost.net/v1/86bbbcfa8ad043109d2d7af530225c72/logs_40/672840/8/check
  /nova-next/9d13cfb/ara-report/result/d13f888f-d187-4c3b-b5ab-
  9326f611e534/

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1840967/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1818734] Re: The endpoint group API doesn't use default roles

2019-08-21 Thread OpenStack Infra
Reviewed:  https://review.opendev.org/676115
Committed: 
https://git.openstack.org/cgit/openstack/keystone/commit/?id=cf572f9e843745fd3b0fb2040e8a7a9dec94432e
Submitter: Zuul
Branch:master

commit cf572f9e843745fd3b0fb2040e8a7a9dec94432e
Author: Vishakha Agarwal 
Date:   Tue Aug 13 12:07:19 2019 +0530

Add tests for project users interacting with endpoint_groups

This commit introduces some tests that show how project users
are expected to behave with the endpoint_groups API.

Change-Id: I0f32de4ea615c89a7500a8098c44ef543fe45a02
Closes-bug: #1818734


** Changed in: keystone
   Status: In Progress => Fix Released

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Identity (keystone).
https://bugs.launchpad.net/bugs/1818734

Title:
  The endpoint group API doesn't use default roles

Status in OpenStack Identity (keystone):
  Fix Released

Bug description:
  In Rocky, keystone implemented support to ensure at least three default roles 
were available [0].
  An endpoint group is a collection of endpoints that can be populated in a 
users service catalog through association to projects. Ultimately, endpoint 
groups are system-specific resources and shouldn't be accessible directly by 
domain or project users.

  The report is to track the work for implementing system `member` and
  system `reader` role support for endpoint groups.

  [0] 
http://specs.openstack.org/openstack/keystone-specs/specs/keystone/rocky/define-default-roles.html
  [1] 
http://git.openstack.org/cgit/openstack/keystone/tree/keystone/common/policies/endpoint_group.py?id=6e3f1f6e46787ed4542609c935c13cb85e91d7fc
  API Reference: 
https://developer.openstack.org/api-ref/identity/v3-ext/index.html#os-ep-filter-api

To manage notifications about this bug go to:
https://bugs.launchpad.net/keystone/+bug/1818734/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1840952] [NEW] no tenant network is available for allocation

2019-08-21 Thread brunotapdur
Public bug reported:

hi i don t know if it is a bug but i spent 3 days with this error. 
i read and redo the neutron option 2 network as the service
. i use stein version
. i installed a br0 bridge

my questions are :
. in linuxbridgeagent.ini must i set  provider:eno1 or provider:br0 ?
. how to create provider tenant with the new openstack client cli ?

this message rised with : openstack network create edge --enable
--internal

(all previous commands are all ok (horizon, controller, compute node,
domain, project, user, endpoint, etc). at startup and verification all
documentation commands are ok

i don t know where to investigate.
thx
Bruno

** Affects: neutron
 Importance: Undecided
 Status: New

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to neutron.
https://bugs.launchpad.net/bugs/1840952

Title:
  no tenant network is available for allocation

Status in neutron:
  New

Bug description:
  hi i don t know if it is a bug but i spent 3 days with this error. 
  i read and redo the neutron option 2 network as the service
  . i use stein version
  . i installed a br0 bridge

  my questions are :
  . in linuxbridgeagent.ini must i set  provider:eno1 or provider:br0 ?
  . how to create provider tenant with the new openstack client cli ?

  this message rised with : openstack network create edge --enable
  --internal

  (all previous commands are all ok (horizon, controller, compute node,
  domain, project, user, endpoint, etc). at startup and verification all
  documentation commands are ok

  i don t know where to investigate.
  thx
  Bruno

To manage notifications about this bug go to:
https://bugs.launchpad.net/neutron/+bug/1840952/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1704179] Re: Too many period db actions in large scale clusters increase the load of database

2019-08-21 Thread Matt Riedemann
*** This bug is a duplicate of bug 1729621 ***
https://bugs.launchpad.net/bugs/1729621

** This bug has been marked a duplicate of bug 1729621
   Inconsistent value for vcpu_used

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1704179

Title:
  Too many period db actions in large scale clusters increase the  load
  of database

Status in OpenStack Compute (nova):
  In Progress

Bug description:
  Too many period db actions in large scale clusters increase the  load
  of database,  especially un-necessary db update or query.

  For example, over 1000 nodes, it will be 2 * 1000=2000 db update for
  compute_node table in 60s in _update_available_resource, but this two
  db update is not necessary if resource usage not changed.

  Delete the first and second _update()  in _init_compute_node can
  reduce two db update for one node in 60s, if resource usage not
  changed for this compute_node.

  Then the funtion self._resource_change(compute_node) in _update() make
  sense

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1704179/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1729621] Re: Inconsistent value for vcpu_used

2019-08-21 Thread Matt Riedemann
** Also affects: nova/queens
   Importance: Undecided
   Status: New

** Also affects: nova/rocky
   Importance: Undecided
   Status: New

** Changed in: nova/queens
   Status: New => Fix Released

** Changed in: nova/rocky
   Status: New => Fix Released

** Changed in: nova/pike
 Assignee: Tony Breeds (o-tony) => Radoslav Gerganov (rgerganov)

** Changed in: nova/pike
   Status: In Progress => Won't Fix

** Changed in: nova/queens
 Assignee: (unassigned) => Radoslav Gerganov (rgerganov)

** Changed in: nova/rocky
 Assignee: (unassigned) => Radoslav Gerganov (rgerganov)

** No longer affects: nova/ocata

** Changed in: nova/queens
   Importance: Undecided => High

** Changed in: nova/rocky
   Importance: Undecided => High

** Changed in: nova/pike
   Importance: Undecided => High

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1729621

Title:
  Inconsistent value for vcpu_used

Status in OpenStack Compute (nova):
  Fix Released
Status in OpenStack Compute (nova) pike series:
  Won't Fix
Status in OpenStack Compute (nova) queens series:
  Fix Released
Status in OpenStack Compute (nova) rocky series:
  Fix Released

Bug description:
  Description
  ===

  Nova updates hypervisor resources using function called
  ./nova/compute/resource_tracker.py:update_available_resource().

  In case of *shutdowned* instances it could impact inconsistent values
  for resources like vcpu_used.

  Resources are taken from function self.driver.get_available_resource():
  
https://github.com/openstack/nova/blob/f974e3c3566f379211d7fdc790d07b5680925584/nova/compute/resource_tracker.py#L617
  
https://github.com/openstack/nova/blob/f974e3c3566f379211d7fdc790d07b5680925584/nova/virt/libvirt/driver.py#L5766

  This function calculates allocated vcpu's based on function _get_vcpu_total().
  
https://github.com/openstack/nova/blob/f974e3c3566f379211d7fdc790d07b5680925584/nova/virt/libvirt/driver.py#L5352

  As we see in _get_vcpu_total() function calls
  *self._host.list_guests()* without "only_running=False" parameter. So
  it doesn't respect shutdowned instances.

  At the end of resource update process function _update_available_resource() 
is beign called:
  > /opt/stack/nova/nova/compute/resource_tracker.py(733)

   677 @utils.synchronized(COMPUTE_RESOURCE_SEMAPHORE)
   678 def _update_available_resource(self, context, resources):
   679
   681 # initialize the compute node object, creating it
   682 # if it does not already exist.
   683 self._init_compute_node(context, resources)

  It initialize compute node object with resources that are calculated
  without shutdowned instances. If compute node object already exists it
  *UPDATES* its fields - *for a while nova-api has other resources
  values than it its in real.*

   731 # update the compute_node
   732 self._update(context, cn)

  The inconsistency is automatically fixed during other code execution:
  
https://github.com/openstack/nova/blob/f974e3c3566f379211d7fdc790d07b5680925584/nova/compute/resource_tracker.py#L709

  But for heavy-loaded hypervisors (like 100 active instances and 30
  shutdowned instances) it creates wrong informations in nova database
  for about 4-5 seconds (in my usecase) - it could impact other issues
  like spawning on already full hypervisor (because scheduler has wrong
  informations about hypervisor usage).

  Steps to reproduce
  ==

  1) Start devstack
  2) Create 120 instances
  3) Stop some instances
  4) Watch blinking values in nova hypervisor-show
  nova hypervisor-show e6dfc16b-7914-48fb-a235-6fe3a41bb6db

  Expected result
  ===
  Returned values should be the same during test.

  Actual result
  =
  while true; do echo -n "$(date) "; echo "select hypervisor_hostname, 
vcpus_used from compute_nodes where 
hypervisor_hostname='example.compute.node.com';" | mysql nova_cell1; sleep 0.3; 
done

  Thu Nov  2 14:50:09 UTC 2017 example.compute.node.com  120
  Thu Nov  2 14:50:10 UTC 2017 example.compute.node.com  120
  Thu Nov  2 14:50:10 UTC 2017 example.compute.node.com  120
  Thu Nov  2 14:50:10 UTC 2017 example.compute.node.com  120
  Thu Nov  2 14:50:11 UTC 2017 example.compute.node.com  120
  Thu Nov  2 14:50:11 UTC 2017 example.compute.node.com  120
  Thu Nov  2 14:50:11 UTC 2017 example.compute.node.com  120
  Thu Nov  2 14:50:11 UTC 2017 example.compute.node.com  120
  Thu Nov  2 14:50:12 UTC 2017 example.compute.node.com  117
  Thu Nov  2 14:50:12 UTC 2017 example.compute.node.com  117
  Thu Nov  2 14:50:12 UTC 2017 example.compute.node.com  117
  Thu Nov  2 14:50:13 UTC 2017 example.compute.node.com  117
  Thu Nov  2 14:50:13 UTC 2017 example.compute.node.com  117
  Thu Nov  2 14:50:13 UTC 2017 example.compute.node.com  117
  Thu Nov  2 14:50:14 UTC 20

[Yahoo-eng-team] [Bug 1789991] Re: nova-compute error after enrolling ironic baremetal nodes

2019-08-21 Thread Matt Riedemann
*** This bug is a duplicate of bug 1839674 ***
https://bugs.launchpad.net/bugs/1839674

** This bug has been marked a duplicate of bug 1839674
   ResourceTracker.compute_nodes won't try to create a ComputeNode a second 
time if the first create() fails

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1789991

Title:
  nova-compute error after enrolling ironic baremetal nodes

Status in OpenStack Compute (nova):
  New

Bug description:
  Description
  ===
  After enrolling some ironic baremetal nodes, I noticed the following in 
nova-compute.log (longer trace below): 

  2018-08-30 17:00:51.142 7 ERROR nova.compute.manager [req-73ba9d4b-
  b51d-4ab7-88c8-5fc3f27fd89e - - - - -] Error updating resources for
  node 0e5705cc-e872-49aa-aff4-1a91278b5cb3.: NotImplementedError:
  Cannot load 'id' in the base class

  Steps to reproduce
  ==

  * Enroll ironic baremetal nodes (openstack baremetal node provide)
  * Wait
  * Error repeatedly appears in nova-compute.log

  Expected result
  ===
  No errors in log

  Actual result
  =
  Errors in log

  Environment
  ===
  openstack-nova-compute-18.0.0-0.20180829095234.45fc232.el7.noarch
  puppet-nova-13.3.1-0.20180825165256.5d1889b.el7.noarch
  python-nova-18.0.0-0.20180829095234.45fc232.el7.noarch
  python-novajoin-1.0.19-0.20180828183900.3d58511.el7.noarch
  openstack-nova-common-18.0.0-0.20180829095234.45fc232.el7.noarch
  python2-novaclient-11.0.0-0.20180807085257.f1005ce.el7.noarch

  
  Logs & Configs
  =
  2018-08-30 17:00:51.142 7 DEBUG oslo_concurrency.lockutils 
[req-73ba9d4b-b51d-4ab7-88c8-5fc3f27fd89e - - - - -] Lock "compute_resources" 
release\
  d by "nova.compute.resource_tracker._update_available_resource" :: held 
0.001s inner /usr/lib/python2.7/site-packages/oslo_concurrency/lockutils\
  .py:285
  2018-08-30 17:00:51.142 7 ERROR nova.compute.manager 
[req-73ba9d4b-b51d-4ab7-88c8-5fc3f27fd89e - - - - -] Error updating resources 
for node 0e57\
  05cc-e872-49aa-aff4-1a91278b5cb3.: NotImplementedError: Cannot load 'id' in 
the base class
  2018-08-30 17:00:51.142 7 ERROR nova.compute.manager Traceback (most recent 
call last):
  2018-08-30 17:00:51.142 7 ERROR nova.compute.manager   File 
"/usr/lib/python2.7/site-packages/nova/compute/manager.py", line 7729, in 
_update_av\
  ailable_resource_for_node
  2018-08-30 17:00:51.142 7 ERROR nova.compute.manager 
rt.update_available_resource(context, nodename)
  2018-08-30 17:00:51.142 7 ERROR nova.compute.manager   File 
"/usr/lib/python2.7/site-packages/nova/compute/resource_tracker.py", line 700, 
in up\
  date_available_resource
  2018-08-30 17:00:51.142 7 ERROR nova.compute.manager 
self._update_available_resource(context, resources)
  2018-08-30 17:00:51.142 7 ERROR nova.compute.manager   File 
"/usr/lib/python2.7/site-packages/oslo_concurrency/lockutils.py", line 274, in 
inner
  2018-08-30 17:00:51.142 7 ERROR nova.compute.manager return f(*args, 
**kwargs)
  2018-08-30 17:00:51.142 7 ERROR nova.compute.manager   File 
"/usr/lib/python2.7/site-packages/nova/compute/resource_tracker.py", line 723, 
in _u\
  pdate_available_resource
  2018-08-30 17:00:51.142 7 ERROR nova.compute.manager 
self._init_compute_node(context, resources)
  2018-08-30 17:00:51.142 7 ERROR nova.compute.manager   File 
"/usr/lib/python2.7/site-packages/nova/compute/resource_tracker.py", line 563, 
in _i\
  nit_compute_node
  2018-08-30 17:00:51.142 7 ERROR nova.compute.manager 
self._setup_pci_tracker(context, cn, resources)
  2018-08-30 17:00:51.142 7 ERROR nova.compute.manager   File 
"/usr/lib/python2.7/site-packages/nova/compute/resource_tracker.py", line 594, 
in _s\
  etup_pci_tracker
  2018-08-30 17:00:51.142 7 ERROR nova.compute.manager n_id = 
compute_node.id
  2018-08-30 17:00:51.142 7 ERROR nova.compute.manager   File 
"/usr/lib/python2.7/site-packages/oslo_versionedobjects/base.py", line 67, in 
getter
  2018-08-30 17:00:51.142 7 ERROR nova.compute.manager 
self.obj_load_attr(name)
  2018-08-30 17:00:51.142 7 ERROR nova.compute.manager   File 
"/usr/lib/python2.7/site-packages/oslo_versionedobjects/base.py", line 603, in 
obj_l\
  oad_attr
  2018-08-30 17:00:51.142 7 ERROR nova.compute.manager _("Cannot load '%s' 
in the base class") % attrname)
  2018-08-30 17:00:51.142 7 ERROR nova.compute.manager NotImplementedError: 
Cannot load 'id' in the base class
  2018-08-30 17:00:51.142 7 ERROR nova.compute.manager

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1789991/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1838003] Re: Horizon hides password requirements

2019-08-21 Thread OpenStack Infra
Reviewed:  https://review.opendev.org/672927
Committed: 
https://git.openstack.org/cgit/openstack/horizon/commit/?id=8f5a3cd7dad96a0a764e0b7bd295fca8878a5166
Submitter: Zuul
Branch:master

commit 8f5a3cd7dad96a0a764e0b7bd295fca8878a5166
Author: Radomir Dopieralski 
Date:   Fri Jul 26 11:22:41 2019 +0200

Display password_regex_description when password change fails

When keystone has password_regex set and the new password doesn't match
it, the attempt to change the password will fail, and keystone will
helpfully include password_regex_description in its error message.
Unfortunately Horizon discards that and displays "Unable to update the
user password.", which is considerably less helpful.

This change extracts the description from the error message in this
case, and displays it to the user together with a translated error
message. Unfortunately we have no way of also translating the
description.

Closes-bug: #1838003
Change-Id: I535c020deed76188c0f7150072fbdced9c5454fc


** Changed in: horizon
   Status: In Progress => Fix Released

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Dashboard (Horizon).
https://bugs.launchpad.net/bugs/1838003

Title:
  Horizon hides password requirements

Status in OpenStack Dashboard (Horizon):
  Fix Released

Bug description:
  When keystone has password_regex configured, and a new password
  doesn't match it, it will include password_regex_description in the
  error message, so that the users have a chance at coming up with a
  password that matches the requirements. Unfortunately Horizon hides
  that message and displays a completely unhelpful "Unable to update the
  user password." instead.

  Ideally we would like to display password_regex_description right in
  the password change form, in its description, but there is currently
  no API in keystone that would let us retrieve it. So instead it would
  be nice to at least display the description from the error message.

To manage notifications about this bug go to:
https://bugs.launchpad.net/horizon/+bug/1838003/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1840930] [NEW] Networking service in neutron - install guide says to configure nova with [neutron]/url which is deprecated

2019-08-21 Thread Matt Riedemann
Public bug reported:

- [x] This doc is inaccurate in this way:

The [neutron]/url option
https://docs.openstack.org/nova/latest/configuration/config.html#neutron.url
in nova has been deprecated since the Queens release and is being
removed in Train. The neutron/compute config guide in the neutron
install guides still says to use the url option though. Since Queens
when nova started using KSA adapters for working with neutron config:

https://review.opendev.org/#/c/509892/

I think we want to avoid configuring the [neutron] section in nova.conf
with url or endpoint_override but install rely on KSA to use the service
types authority to find the endpoint to use based on service name/type
and interface, in other words things should just work without needing to
explicitly define an endpoint url to use for nova talking to neutron -
nova can go through KSA and the service catalog to get the endpoint it
needs.

---
Release: 14.1.0.dev665 on 2017-06-30 05:58:47
SHA: 490471ebd3ac56d0cee164b9c1c1211687e49437
Source: https://opendev.org/openstack/neutron/src/doc/source/install/index.rst
URL: https://docs.openstack.org/neutron/latest/install/

** Affects: neutron
 Importance: Undecided
 Status: New


** Tags: doc

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to neutron.
https://bugs.launchpad.net/bugs/1840930

Title:
  Networking service in neutron - install guide says to configure nova
  with [neutron]/url which is deprecated

Status in neutron:
  New

Bug description:
  - [x] This doc is inaccurate in this way:

  The [neutron]/url option
  https://docs.openstack.org/nova/latest/configuration/config.html#neutron.url
  in nova has been deprecated since the Queens release and is being
  removed in Train. The neutron/compute config guide in the neutron
  install guides still says to use the url option though. Since Queens
  when nova started using KSA adapters for working with neutron config:

  https://review.opendev.org/#/c/509892/

  I think we want to avoid configuring the [neutron] section in
  nova.conf with url or endpoint_override but install rely on KSA to use
  the service types authority to find the endpoint to use based on
  service name/type and interface, in other words things should just
  work without needing to explicitly define an endpoint url to use for
  nova talking to neutron - nova can go through KSA and the service
  catalog to get the endpoint it needs.

  ---
  Release: 14.1.0.dev665 on 2017-06-30 05:58:47
  SHA: 490471ebd3ac56d0cee164b9c1c1211687e49437
  Source: https://opendev.org/openstack/neutron/src/doc/source/install/index.rst
  URL: https://docs.openstack.org/neutron/latest/install/

To manage notifications about this bug go to:
https://bugs.launchpad.net/neutron/+bug/1840930/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1840927] [NEW] Rewrite doc on L2 population

2019-08-21 Thread Alexandra Settle
Public bug reported:

In the Networking Guide back in Kilo
https://docs.openstack.org/neutron/latest/admin/config-ml2.html there
was a section on L2 population. This section no longer exists, and
nothing similar is in existance.

As a replacement, this blog post was used in the documentation:
https://networkop.co.uk/blog/2016/05/06/neutron-l2pop/ as per this
review: https://review.opendev.org/#/c/676920/

It would be best if this knowledge could be transferred.

Thanks,

Alex

** Affects: neutron
 Importance: Undecided
 Status: New


** Tags: docs

** Tags added: docs

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to neutron.
https://bugs.launchpad.net/bugs/1840927

Title:
  Rewrite doc on L2 population

Status in neutron:
  New

Bug description:
  In the Networking Guide back in Kilo
  https://docs.openstack.org/neutron/latest/admin/config-ml2.html there
  was a section on L2 population. This section no longer exists, and
  nothing similar is in existance.

  As a replacement, this blog post was used in the documentation:
  https://networkop.co.uk/blog/2016/05/06/neutron-l2pop/ as per this
  review: https://review.opendev.org/#/c/676920/

  It would be best if this knowledge could be transferred.

  Thanks,

  Alex

To manage notifications about this bug go to:
https://bugs.launchpad.net/neutron/+bug/1840927/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1840912] [NEW] libvirt calls aren't reliably using tpool.Proxy

2019-08-21 Thread Matthew Booth
Public bug reported:

A customer is hitting an issue with symptoms identical to bug 1045152
(from 2012). Specifically, we are frequently seeing the compute host
being marked down. From log correlation, we can see that when this
occurs the relevant compute is always in the middle of executing
LibvirtDriver._get_disk_over_committed_size_total(). The reason for this
appears to be a long-running libvirt call which is not using
tpool.Proxy, and therefore blocks all other greenthreads during
execution. We do not yet know why the libvirt call is slow, but we have
identified the reason it is not using tpool.Proxy.

Because eventlet, we proxy libvirt calls at the point we create the
libvirt connection in libvirt.Host._connect:

return tpool.proxy_call(
(libvirt.virDomain, libvirt.virConnect),
libvirt.openAuth, uri, auth, flags)

This means: run libvirt.openAuth(uri, auth, flags) in a native thread.
If the returned object is a libvirt.virDomain or libvirt.virConnect,
wrap the returned object in a tpool.Proxy with the same autowrap rules.

There are 2 problems with this. Firstly, the autowrap list is
incomplete. At the very least we need to add libvirt.virNodeDevice,
libvirt.virSecret, and libvirt.NWFilter to this list as we use all of
these objects in Nova. Currently none of our interactions with these
objects are using the tpool proxy.

Secondly, and the specific root cause of this bug, it doesn't understand
lists:

https://github.com/eventlet/eventlet/blob/ca8dd0748a1985a409e9a9a517690f46e05cae99/eventlet/tpool.py#L149

In LibvirtDriver._get_disk_over_committed_size_total() we get a list of
running libvirt domains with libvirt.Host.list_instance_domains, which
calls virConnect.listAllDomains(). listAllDomains() returns a *list* of
virDomain, which the above code in tpool doesn't match. Consequently,
none of the subsequent virDomain calls use the tpool proxy, which
starves all other greenthreads.

** Affects: nova
 Importance: Undecided
 Status: New

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1840912

Title:
  libvirt calls aren't reliably using tpool.Proxy

Status in OpenStack Compute (nova):
  New

Bug description:
  A customer is hitting an issue with symptoms identical to bug 1045152
  (from 2012). Specifically, we are frequently seeing the compute host
  being marked down. From log correlation, we can see that when this
  occurs the relevant compute is always in the middle of executing
  LibvirtDriver._get_disk_over_committed_size_total(). The reason for
  this appears to be a long-running libvirt call which is not using
  tpool.Proxy, and therefore blocks all other greenthreads during
  execution. We do not yet know why the libvirt call is slow, but we
  have identified the reason it is not using tpool.Proxy.

  Because eventlet, we proxy libvirt calls at the point we create the
  libvirt connection in libvirt.Host._connect:

  return tpool.proxy_call(
  (libvirt.virDomain, libvirt.virConnect),
  libvirt.openAuth, uri, auth, flags)

  This means: run libvirt.openAuth(uri, auth, flags) in a native thread.
  If the returned object is a libvirt.virDomain or libvirt.virConnect,
  wrap the returned object in a tpool.Proxy with the same autowrap
  rules.

  There are 2 problems with this. Firstly, the autowrap list is
  incomplete. At the very least we need to add libvirt.virNodeDevice,
  libvirt.virSecret, and libvirt.NWFilter to this list as we use all of
  these objects in Nova. Currently none of our interactions with these
  objects are using the tpool proxy.

  Secondly, and the specific root cause of this bug, it doesn't
  understand lists:

  
https://github.com/eventlet/eventlet/blob/ca8dd0748a1985a409e9a9a517690f46e05cae99/eventlet/tpool.py#L149

  In LibvirtDriver._get_disk_over_committed_size_total() we get a list
  of running libvirt domains with libvirt.Host.list_instance_domains,
  which calls virConnect.listAllDomains(). listAllDomains() returns a
  *list* of virDomain, which the above code in tpool doesn't match.
  Consequently, none of the subsequent virDomain calls use the tpool
  proxy, which starves all other greenthreads.

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1840912/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1840898] [NEW] INSTR can not supported for postgresql

2019-08-21 Thread Abhishek Kekane
Public bug reported:

Recently added migration script doesn't support INSTR function. INSTR
function does support sqlite3 and mysql.

[1] https://review.opendev.org/#/c/665606/5

** Affects: glance
 Importance: High
 Assignee: Abhishek Kekane (abhishek-kekane)
 Status: New

** Changed in: glance
 Assignee: (unassigned) => Abhishek Kekane (abhishek-kekane)

** Changed in: glance
   Importance: Undecided => High

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to Glance.
https://bugs.launchpad.net/bugs/1840898

Title:
  INSTR can not supported for postgresql

Status in Glance:
  New

Bug description:
  Recently added migration script doesn't support INSTR function. INSTR
  function does support sqlite3 and mysql.

  [1] https://review.opendev.org/#/c/665606/5

To manage notifications about this bug go to:
https://bugs.launchpad.net/glance/+bug/1840898/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1840895] [NEW] segment parameter check failed when creating network

2019-08-21 Thread Jiafang Zhang
Public bug reported:

neutron net-create test --provider:network_type vlan
--provider:segmentation_id 0

Execute commands like this, all vlan in ml2_vlan_allocations table is
set to allocated, no vlan network can be created.

validate_provider_segment function should check whether
provider:segmentation_id is 0.

** Affects: neutron
 Importance: Undecided
 Status: New

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to neutron.
https://bugs.launchpad.net/bugs/1840895

Title:
  segment parameter check failed when creating network

Status in neutron:
  New

Bug description:
  neutron net-create test --provider:network_type vlan
  --provider:segmentation_id 0

  Execute commands like this, all vlan in ml2_vlan_allocations table is
  set to allocated, no vlan network can be created.

  validate_provider_segment function should check whether
  provider:segmentation_id is 0.

To manage notifications about this bug go to:
https://bugs.launchpad.net/neutron/+bug/1840895/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp