[Yahoo-eng-team] [Bug 1759473] Re: Avoid InstanceNotRunning reporting

2019-11-22 Thread OpenStack Infra
Reviewed:  https://review.opendev.org/541152
Committed: 
https://git.openstack.org/cgit/openstack/nova/commit/?id=ad78f24812df8e83c188b100d05b1c5a577c8f84
Submitter: Zuul
Branch:master

commit ad78f24812df8e83c188b100d05b1c5a577c8f84
Author: jichenjc 
Date:   Fri Feb 2 21:29:02 2018 +0800

Avoid raise InstanceNotRunning exception

when instance is being snapshot, it's possible that the instance
is deleted by another thread concurrently, libvirt driver translate
this InstanceNotFound exception into InstanceNotRunning and we didn't
catch and handle the exception and it leads to backtrace in compute log.

Also, fixes to catch the Exception and return a debug info
as the instance is anyway deleted already.

Change-Id: Ic6d0ac894fbbbf03fe967ee4f9ec93b2491cf073
Closes-Bug: 1759473


** Changed in: nova
   Status: In Progress => Fix Released

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1759473

Title:
  Avoid InstanceNotRunning reporting

Status in OpenStack Compute (nova):
  Fix Released

Bug description:
  This is found during bug analysis  1737024 and discussion on 
  https://review.openstack.org/#/c/541152/
  and it turn out InstanceNotRunning 
  is not handled in current implementation and can lead to potential back trace 

  so better to handle it

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1759473/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1853651] [NEW] Horizon Install Guide - missing 'WEBROOT' directive in horizon config file

2019-11-22 Thread Justin Goetz
Public bug reported:


This bug tracker is for errors with the documentation, use the following
as a template and remove or add fields as you see fit. Convert [ ] into
[x] to check boxes:

- [x] This doc is inaccurate in this way:

By following the instructions to setup the /etc/openstack-
dashboard/local_settings configuration file on CentOS 7, I received a
"Not Found" error when attempting to load horizon at
/dashboard in a browser. It appears horizon tries to re-
direct to a URL outside of the /dashboard URL, causing the 404. This is
due to not having a "WEBROOT" setting in the file.

By adding the following to the config file (and restarting httpd), the
horizon dashboard then loads without issue:

WEBROOT = '/dashboard'

My apologies if this is the incorrect place to request this
documentation addition.

Thanks!

If you have a troubleshooting or support issue, use the following
resources:

 - Ask OpenStack: http://ask.openstack.org
 - The mailing list: http://lists.openstack.org
 - IRC: 'openstack' channel on Freenode

---
Release:  on 2019-04-11 17:44:22
SHA: fd9839a0768842308b3f9ecabafc04b120453994
Source: 
https://opendev.org/openstack/horizon/src/doc/source/install/install-rdo.rst
URL: https://docs.openstack.org/horizon/train/install/install-rdo.html

** Affects: horizon
 Importance: Undecided
 Status: New


** Tags: documentation

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Dashboard (Horizon).
https://bugs.launchpad.net/bugs/1853651

Title:
  Horizon Install Guide - missing 'WEBROOT' directive in horizon config
  file

Status in OpenStack Dashboard (Horizon):
  New

Bug description:

  This bug tracker is for errors with the documentation, use the
  following as a template and remove or add fields as you see fit.
  Convert [ ] into [x] to check boxes:

  - [x] This doc is inaccurate in this way:

  By following the instructions to setup the /etc/openstack-
  dashboard/local_settings configuration file on CentOS 7, I received a
  "Not Found" error when attempting to load horizon at
  /dashboard in a browser. It appears horizon tries to
  re-direct to a URL outside of the /dashboard URL, causing the 404.
  This is due to not having a "WEBROOT" setting in the file.

  By adding the following to the config file (and restarting httpd), the
  horizon dashboard then loads without issue:

  WEBROOT = '/dashboard'

  My apologies if this is the incorrect place to request this
  documentation addition.

  Thanks!

  If you have a troubleshooting or support issue, use the following
  resources:

   - Ask OpenStack: http://ask.openstack.org
   - The mailing list: http://lists.openstack.org
   - IRC: 'openstack' channel on Freenode

  ---
  Release:  on 2019-04-11 17:44:22
  SHA: fd9839a0768842308b3f9ecabafc04b120453994
  Source: 
https://opendev.org/openstack/horizon/src/doc/source/install/install-rdo.rst
  URL: https://docs.openstack.org/horizon/train/install/install-rdo.html

To manage notifications about this bug go to:
https://bugs.launchpad.net/horizon/+bug/1853651/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1753676] Re: Live migration not working as Expected when Restarting nova-compute service while migration from source node

2019-11-22 Thread OpenStack Infra
Reviewed:  https://review.opendev.org/678016
Committed: 
https://git.openstack.org/cgit/openstack/nova/commit/?id=ebcf6e4ce576285949c5a202f2d7d21dc03156ef
Submitter: Zuul
Branch:master

commit ebcf6e4ce576285949c5a202f2d7d21dc03156ef
Author: Alexandre Arents 
Date:   Tue Aug 20 13:37:33 2019 +

Abort live-migration during instance_init

When compute service restart during a live-migration,
we lose live-migration monitoring thread. In that case
it is better to early abort live-migration job before resetting
state of instance, this will avoid API to accept further
action while unmanaged migration process still run in background.
It also avoid unexpected/dangerous behavior as describe in related bug.

Change-Id: Idec2d31cbba497dc4b20912f3388ad2341951d23
Closes-Bug: #1753676


** Changed in: nova
   Status: In Progress => Fix Released

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1753676

Title:
  Live migration not working as Expected when Restarting nova-compute
  service while migration from source node

Status in OpenStack Compute (nova):
  Fix Released

Bug description:
  Description
  ===

  Environment: Ubuntu 16.04
  Openstack Version: Pike

  I am trying to migrate VM ( live migration ( block migration ) ) form
  one compute node to another compute node...Everything looks good
  unless I restart nova-compute service, live migration still running
  underneath with help of libvirt, once the vm reaches destination,
  database is not updated properly.

  Steps to reproduce:
  ===

  nova.conf ( libvirt setting on both compute nodes )

  [libvirt]
  live_migration_bandwidth=1200
  live_migration_downtime=100
  live_migration_downtime_steps =3
  live_migration_downtime_delay=10
  live_migration_flag = 
VIR_MIGRATE_UNDEFINE_SOURCE,VIR_MIGRATE_PEER2PEER,VIR_MIGRATE_LIVE
  virt_type = kvm
  inject_password = False
  disk_cachemodes = network=writeback
  live_migration_uri = "qemu+tcp://nova@%s/system"
  live_migration_tunnelled = False
  block_migration_flag = 
VIR_MIGRATE_UNDEFINE_SOURCE,VIR_MIGRATE_PEER2PEER,VIR_MIGRATE_NON_SHARED_INC

  ( default openstack live migration configuration ( pre-copy with no tunneling 
)
  Source vm root disk ( boot from volume with one ephemernal disk (160GB) )

  Trying to migrate vm from compute1 to compute2, below is my source vm.

  | OS-EXT-SRV-ATTR:host | compute1 
   |
  | OS-EXT-SRV-ATTR:hostname | 
testcase1-all-ephemernal-boot-from-vol  
 |
  | OS-EXT-SRV-ATTR:hypervisor_hostname  | compute1 
   |
  | OS-EXT-SRV-ATTR:instance_name| instance-0153

  1) nova live-migration --block-migrate  compute2

  [req-48a3df61-3974-46ac-8019-c4c4a0f8a8c8
  4a8150eb246a4450829331e993f8c3fd f11a5d3631f14c4f879a2e7dddb96c06 -
  default default] pre_live_migration data is
  
LibvirtLiveMigrateData(bdms=,block_migration=True,disk_available_mb=6900736,disk_over_commit=,filename='tmpW5ApOS',graphics_listen_addr_spice=x.x.x.x,graphics_listen_addr_vnc=127.0.0.1,image_type='default',instance_relative_path='504028fc-1381
  -42ca-ad7c-
  
def7f749a722',is_shared_block_storage=False,is_shared_instance_path=False,is_volume_backed=True,migration=,serial_listen_addr=None,serial_listen_ports=,supported_perf_events=,target_connect_addr=)
  pre_live_migration /openstack/venvs/nova-16.0.6/lib/python2.7/site-
  packages/nova/compute/manager.py:5437

  Migration started, able to see the data and memory transfer ( using
  iftop )

  Data transfer between compute nodes using iftop
    
<=  
4.94Gb  4.99Gb  5.01Gb

  Restarted Nova-compute service on source compute node ( where the vm
  is migrating)

  Live migration still it is going, once migration completes, below is
  my total data transfer ( using iftop )

  TX: cum:   17.3MB   peak:   2.50Mb

  rates:   11.1Kb  7.11Kb   463Kb
  RX:97.7GB   4.97Gb

   3.82Kb  1.93Kb  1.87Gb
  TOTAL: 97.7GB   4.97Gb

  Once migration completes, from the destination compute node ( we can
  able to see the virsh domain running)

  root@compute2:~# virsh list --all
   IdName   State
  
   3 instance-0153  

[Yahoo-eng-team] [Bug 1853637] [NEW] Assign floating IP to port owned by another tenant is not override-able with RBAC policy

2019-11-22 Thread Michael Johnson
Public bug reported:

In neutron/db/l3_db.py:

def _internal_fip_assoc_data(self, context, fip, tenant_id):
"""Retrieve internal port data for floating IP.
Retrieve information concerning the internal port where
the floating IP should be associated to.
"""
internal_port = self._core_plugin.get_port(context, fip['port_id'])
if internal_port['tenant_id'] != tenant_id and not context.is_admin:
port_id = fip['port_id']
msg = (_('Cannot process floating IP association with '
 'Port %s, since that port is owned by a '
 'different tenant') % port_id)
raise n_exc.BadRequest(resource='floatingip', msg=msg)

This code does not allow operators to override the ability to assign
floating IPs to ports on another tenant using RBAC policy. It also does
not allow members of the advsvc role to take this action.

This code should be fixed to use the standard neutron RBAC and allow the
advsvc role to take this action.

** Affects: neutron
 Importance: Undecided
 Status: New

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to neutron.
https://bugs.launchpad.net/bugs/1853637

Title:
  Assign floating IP to port owned by another tenant is not override-
  able with RBAC policy

Status in neutron:
  New

Bug description:
  In neutron/db/l3_db.py:

  def _internal_fip_assoc_data(self, context, fip, tenant_id):
  """Retrieve internal port data for floating IP.
  Retrieve information concerning the internal port where
  the floating IP should be associated to.
  """
  internal_port = self._core_plugin.get_port(context, fip['port_id'])
  if internal_port['tenant_id'] != tenant_id and not context.is_admin:
  port_id = fip['port_id']
  msg = (_('Cannot process floating IP association with '
   'Port %s, since that port is owned by a '
   'different tenant') % port_id)
  raise n_exc.BadRequest(resource='floatingip', msg=msg)

  This code does not allow operators to override the ability to assign
  floating IPs to ports on another tenant using RBAC policy. It also
  does not allow members of the advsvc role to take this action.

  This code should be fixed to use the standard neutron RBAC and allow
  the advsvc role to take this action.

To manage notifications about this bug go to:
https://bugs.launchpad.net/neutron/+bug/1853637/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1840638] Re: network/subnet resources cannot be read and written separated

2019-11-22 Thread OpenStack Infra
Reviewed:  https://review.opendev.org/677166
Committed: 
https://git.openstack.org/cgit/openstack/neutron/commit/?id=4afba466e04fb74c8915f9f1b40686008b2503fb
Submitter: Zuul
Branch:master

commit 4afba466e04fb74c8915f9f1b40686008b2503fb
Author: zhanghao 
Date:   Sun Nov 17 20:17:35 2019 -0500

Make network support read and write separation

When 'slave_connection' is configured in neutron.conf, executing
the 'neutron port-show' command will read the port information from
the slave database nodes, but the network will not be able to read
the information from the slave database nodes.

Change-Id: I572cecace1016740584757e5f926d246ead2d9a0
Closes-Bug: #1840638


** Changed in: neutron
   Status: In Progress => Fix Released

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to neutron.
https://bugs.launchpad.net/bugs/1840638

Title:
  network/subnet resources cannot be read and written separated

Status in neutron:
  Fix Released

Bug description:
  openstack environment: 1 controller 1 compute(install mariadb)
  neutron.conf
  [database]
  connection = mysql+pymysql://neutron:***@controller/neutron
  slave_connection = mysql+pymysql://neutron: ***@compute/neutron

  create network(name=net10)、subnet(name=subnet10)、port(name=port01) in
  controller node.

  sync controller node database to compute node.

  update
  network.name=test_net10、subnet.name=test_subnet10、port.name=test_port01
  in compute node database.

  neutron port-show PORT_ID name:test_port01
  neutron net-show NET_ID   name:net10
  neutron subnet-show SUBNET_ID name:subnet10

  When slave_connection is configured in neutron.conf, executing the
  'neutron port-show' command will read the port information from the
  slave database nodes, but the network/subnet will not be able to read
  the information from the slave database nodes.

To manage notifications about this bug go to:
https://bugs.launchpad.net/neutron/+bug/1840638/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1853632] [NEW] designate dns driver does not use domain settings for auth

2019-11-22 Thread Manuel Torrinha
Public bug reported:

The designate external dns driver does not use domain settings for
authentication **if** there is more than one openstack domain.

If you have only the 'Default' domain the authentication system seems to have
no doubt on which domain to use so it will use that.

In our deployment we support federated authentication and we have this
issue.

The issue lies in the external dns driver as it also does not support
all of the documented options. You can test this by setting invalid one
of (or all):

  user_domain_name
  project_domain_name
  project_name

it should yield the same results.

@oammis initially found this issue, so credit where it's due.

We have a Queens deployment, although by what we can see from the code
it should be transverse to all releases.

I'll post a fix soon.

** Affects: neutron
 Importance: Undecided
 Status: New

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to neutron.
https://bugs.launchpad.net/bugs/1853632

Title:
  designate dns driver does not use domain settings for auth

Status in neutron:
  New

Bug description:
  The designate external dns driver does not use domain settings for
  authentication **if** there is more than one openstack domain.

  If you have only the 'Default' domain the authentication system seems to have
  no doubt on which domain to use so it will use that.

  In our deployment we support federated authentication and we have this
  issue.

  The issue lies in the external dns driver as it also does not support
  all of the documented options. You can test this by setting invalid
  one of (or all):

    user_domain_name
    project_domain_name
    project_name

  it should yield the same results.

  @oammis initially found this issue, so credit where it's due.

  We have a Queens deployment, although by what we can see from the code
  it should be transverse to all releases.

  I'll post a fix soon.

To manage notifications about this bug go to:
https://bugs.launchpad.net/neutron/+bug/1853632/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1853635] [NEW] Config_drive file throws Request is too large

2019-11-22 Thread bhreddy
Public bug reported:

Openstack Version: QUEENS
I am trying to instantiate VM with config-drive option with two files 
config1.xml(size 200KB) and config2.xml(100KB).
openstack server create config-drive=true --file config1.xml=config1.xml 
-flavor m1.tiny --image cirros

instance creation was failed with error "OverLimit: Request is too
large"

My Quota is already with 50

nova quota-show command output:

injected_file_content_bytes | 50 |

Even I have updated the following in nova.conf 
injected_file_content_bytes = 50

Any suggestion is highly appreciated.

** Affects: nova
 Importance: Undecided
 Status: New

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to OpenStack Compute (nova).
https://bugs.launchpad.net/bugs/1853635

Title:
  Config_drive file throws Request is too large

Status in OpenStack Compute (nova):
  New

Bug description:
  Openstack Version: QUEENS
  I am trying to instantiate VM with config-drive option with two files 
config1.xml(size 200KB) and config2.xml(100KB).
  openstack server create config-drive=true --file config1.xml=config1.xml 
-flavor m1.tiny --image cirros

  instance creation was failed with error "OverLimit: Request is too
  large"

  My Quota is already with 50

  nova quota-show command output:

  injected_file_content_bytes | 50 |

  Even I have updated the following in nova.conf 
  injected_file_content_bytes = 50

  Any suggestion is highly appreciated.

To manage notifications about this bug go to:
https://bugs.launchpad.net/nova/+bug/1853635/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1853613] [NEW] VMs don't get ip from dhcp after compute restart

2019-11-22 Thread Darragh O'Reilly
Public bug reported:

Env: pike + ovs + vxlan + l2pop + iptables_hybrid.
Dhcp agent on differnt node than compute.

Steps:
1. Boot 4 or more vms to same compute and same vxlan net.
2. Wait until they are fully running and reboot compute node.
3. After boot the vms are in status SHUTOFF. Start the vms.

Vms don't get an ip address from neutron dhcp. The flood to tunnels flow
(br-tun table 22) for the network is missing, so broadcasts like dhcp
requests don't get on a tunnel to the node with dhcp agent. Neutron
server did not send the flooding entry to the agent. It only does that
for the first or second active port, or if the agent is restarted.

After the compute boots, neutron-ovs-cleanup runs first and deletes the
qvo ports from br-int [4]. Then the ovs-agent starts and nova-compute
after it. Nova-compute destroys the domains and moves the vms to SHUTOFF
status. It also (for some reason) recreates the qbr linux bridges and
qvb/qvo veths connected to br-int. So neutron continues to see the ports
as ACTIVE even though the vms are SHUTOFF, and agent_active_ports [1]
never drops below 3. Also nova-compute might start a short time after
the ovs-agent and the new ports are not detected in first iteration of
the ovs agent loop, so agent_restarted will be false here [2].

Before [3] agent_restarted was true if the agent was running for less
than agent_boot_time (default 180 sec) and the problem did not show.

It does not happen if neutron-ovs-cleanup is disabled. Then the ovs
agent first treats them as skipped_devices and they get status DOWN.

[1] 
https://github.com/openstack/neutron/blob/21a52f7ae597f7992f32ff41cedff0c31e35c762/neutron/plugins/ml2/drivers/l2pop/mech_driver.py#L306
[2] 
https://github.com/openstack/neutron/blob/21a52f7ae597f7992f32ff41cedff0c31e35c762/neutron/plugins/ml2/drivers/l2pop/mech_driver.py#L310
 
[3] 
https://opendev.org/openstack/neutron/commit/62fe7852bbd70a24174853997096c52ee015e269
[4] https://bugs.launchpad.net/neutron/+bug/1853582

** Affects: neutron
 Importance: Undecided
 Status: New


** Tags: l2-pop ovs

** Tags added: l2-pop ovs

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to neutron.
https://bugs.launchpad.net/bugs/1853613

Title:
  VMs don't get ip from dhcp after compute restart

Status in neutron:
  New

Bug description:
  Env: pike + ovs + vxlan + l2pop + iptables_hybrid.
  Dhcp agent on differnt node than compute.

  Steps:
  1. Boot 4 or more vms to same compute and same vxlan net.
  2. Wait until they are fully running and reboot compute node.
  3. After boot the vms are in status SHUTOFF. Start the vms.

  Vms don't get an ip address from neutron dhcp. The flood to tunnels
  flow (br-tun table 22) for the network is missing, so broadcasts like
  dhcp requests don't get on a tunnel to the node with dhcp agent.
  Neutron server did not send the flooding entry to the agent. It only
  does that for the first or second active port, or if the agent is
  restarted.

  After the compute boots, neutron-ovs-cleanup runs first and deletes
  the qvo ports from br-int [4]. Then the ovs-agent starts and nova-
  compute after it. Nova-compute destroys the domains and moves the vms
  to SHUTOFF status. It also (for some reason) recreates the qbr linux
  bridges and qvb/qvo veths connected to br-int. So neutron continues to
  see the ports as ACTIVE even though the vms are SHUTOFF, and
  agent_active_ports [1] never drops below 3. Also nova-compute might
  start a short time after the ovs-agent and the new ports are not
  detected in first iteration of the ovs agent loop, so agent_restarted
  will be false here [2].

  Before [3] agent_restarted was true if the agent was running for less
  than agent_boot_time (default 180 sec) and the problem did not show.

  It does not happen if neutron-ovs-cleanup is disabled. Then the ovs
  agent first treats them as skipped_devices and they get status DOWN.

  [1] 
https://github.com/openstack/neutron/blob/21a52f7ae597f7992f32ff41cedff0c31e35c762/neutron/plugins/ml2/drivers/l2pop/mech_driver.py#L306
  [2] 
https://github.com/openstack/neutron/blob/21a52f7ae597f7992f32ff41cedff0c31e35c762/neutron/plugins/ml2/drivers/l2pop/mech_driver.py#L310
 
  [3] 
https://opendev.org/openstack/neutron/commit/62fe7852bbd70a24174853997096c52ee015e269
  [4] https://bugs.launchpad.net/neutron/+bug/1853582

To manage notifications about this bug go to:
https://bugs.launchpad.net/neutron/+bug/1853613/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1853603] [NEW] SG rules not enforced

2019-11-22 Thread Michal Dulko
Public bug reported:

We see this in kuryr-kubernetes and test_namespace_sg_isolation test.
More info can be found in bugzilla:
https://bugzilla.redhat.com/show_bug.cgi?id=1688323

** Affects: neutron
 Importance: Undecided
 Status: New

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to neutron.
https://bugs.launchpad.net/bugs/1853603

Title:
  SG rules not enforced

Status in neutron:
  New

Bug description:
  We see this in kuryr-kubernetes and test_namespace_sg_isolation test.
  More info can be found in bugzilla:
  https://bugzilla.redhat.com/show_bug.cgi?id=1688323

To manage notifications about this bug go to:
https://bugs.launchpad.net/neutron/+bug/1853603/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp


[Yahoo-eng-team] [Bug 1853587] [NEW] nova compute fails to update inventory in placement after changing [DEFAULT]/host

2019-11-22 Thread Balazs Gibizer
Public bug reported:


* build a single node devstack (hostname=aio)
* when everything up and running change the /etc/nova/nova-cpu.conf 
[DEFAULT]/host to something else than the compute host hostname
* restart the nova-compute service

The below exception is periodically visible in the nova-compute logs

Nov 22 11:50:52 aio nova-compute[15324]: ERROR nova.compute.manager [None 
req-46df956c-7a08-4217-a658-e44a1a8edb28 None None] Error updating resources 
for node aio.: nova.exception.R
Nov 22 11:50:52 aio nova-compute[15324]: ERROR nova.compute.manager Traceback 
(most recent call last):
Nov 22 11:50:52 aio nova-compute[15324]: ERROR nova.compute.manager   File 
"/opt/stack/nova/nova/compute/manager.py", line 9152, in 
_update_available_resource_for_node
Nov 22 11:50:52 aio nova-compute[15324]: ERROR nova.compute.manager 
startup=startup)
Nov 22 11:50:52 aio nova-compute[15324]: ERROR nova.compute.manager   File 
"/opt/stack/nova/nova/compute/resource_tracker.py", line 844, in 
update_available_resource
Nov 22 11:50:52 aio nova-compute[15324]: ERROR nova.compute.manager 
self._update_available_resource(context, resources, startup=startup)
Nov 22 11:50:52 aio nova-compute[15324]: ERROR nova.compute.manager   File 
"/usr/local/lib/python3.6/dist-packages/oslo_concurrency/lockutils.py", line 
328, in inner
Nov 22 11:50:52 aio nova-compute[15324]: ERROR nova.compute.manager return 
f(*args, **kwargs)
Nov 22 11:50:52 aio nova-compute[15324]: ERROR nova.compute.manager   File 
"/opt/stack/nova/nova/compute/resource_tracker.py", line 929, in 
_update_available_resource
Nov 22 11:50:52 aio nova-compute[15324]: ERROR nova.compute.manager 
self._update(context, cn, startup=startup)
Nov 22 11:50:52 aio nova-compute[15324]: ERROR nova.compute.manager   File 
"/opt/stack/nova/nova/compute/resource_tracker.py", line 1178, in _update
Nov 22 11:50:52 aio nova-compute[15324]: ERROR nova.compute.manager 
self._update_to_placement(context, compute_node, startup)
Nov 22 11:50:52 aio nova-compute[15324]: ERROR nova.compute.manager   File 
"/usr/local/lib/python3.6/dist-packages/retrying.py", line 49, in wrapped_f
Nov 22 11:50:52 aio nova-compute[15324]: ERROR nova.compute.manager return 
Retrying(*dargs, **dkw).call(f, *args, **kw)
Nov 22 11:50:52 aio nova-compute[15324]: ERROR nova.compute.manager   File 
"/usr/local/lib/python3.6/dist-packages/retrying.py", line 206, in call
Nov 22 11:50:52 aio nova-compute[15324]: ERROR nova.compute.manager return 
attempt.get(self._wrap_exception)
Nov 22 11:50:52 aio nova-compute[15324]: ERROR nova.compute.manager   File 
"/usr/local/lib/python3.6/dist-packages/retrying.py", line 247, in get
Nov 22 11:50:52 aio nova-compute[15324]: ERROR nova.compute.manager 
six.reraise(self.value[0], self.value[1], self.value[2])
Nov 22 11:50:52 aio nova-compute[15324]: ERROR nova.compute.manager   File 
"/usr/local/lib/python3.6/dist-packages/six.py", line 696, in reraise
Nov 22 11:50:52 aio nova-compute[15324]: ERROR nova.compute.manager raise 
value
Nov 22 11:50:52 aio nova-compute[15324]: ERROR nova.compute.manager   File 
"/usr/local/lib/python3.6/dist-packages/retrying.py", line 200, in call
Nov 22 11:50:52 aio nova-compute[15324]: ERROR nova.compute.manager attempt 
= Attempt(fn(*args, **kwargs), attempt_number, False)
Nov 22 11:50:52 aio nova-compute[15324]: ERROR nova.compute.manager   File 
"/opt/stack/nova/nova/compute/resource_tracker.py", line 1112, in 
_update_to_placement
Nov 22 11:50:52 aio nova-compute[15324]: ERROR nova.compute.manager 
context, compute_node.uuid, name=compute_node.hypervisor_hostname)
Nov 22 11:50:52 aio nova-compute[15324]: ERROR nova.compute.manager   File 
"/opt/stack/nova/nova/scheduler/client/report.py", line 857, in 
get_provider_tree_and_ensure_root
Nov 22 11:50:52 aio nova-compute[15324]: ERROR nova.compute.manager 
parent_provider_uuid=parent_provider_uuid)
Nov 22 11:50:52 aio nova-compute[15324]: ERROR nova.compute.manager   File 
"/opt/stack/nova/nova/scheduler/client/report.py", line 644, in 
_ensure_resource_provider
Nov 22 11:50:52 aio nova-compute[15324]: ERROR nova.compute.manager 
parent_provider_uuid=parent_provider_uuid)
Nov 22 11:50:52 aio nova-compute[15324]: ERROR nova.compute.manager   File 
"/opt/stack/nova/nova/scheduler/client/report.py", line 72, in wrapper
Nov 22 11:50:52 aio nova-compute[15324]: ERROR nova.compute.manager return 
f(self, *a, **k)
Nov 22 11:50:52 aio nova-compute[15324]: ERROR nova.compute.manager   File 
"/opt/stack/nova/nova/scheduler/client/report.py", line 574, in 
_create_resource_provider
Nov 22 11:50:52 aio nova-compute[15324]: ERROR nova.compute.manager raise 
exception.ResourceProviderCreationFailed(name=name)
Nov 22 11:50:52 aio nova-compute[15324]: ERROR nova.compute.manager 
nova.exception.ResourceProviderCreationFailed: Failed to create resource 
provider aio
Nov 22 11:50:52 aio nova-compute[15324]: ERROR nova.compute.manager 

Problems:
Nova fails to

[Yahoo-eng-team] [Bug 1853582] [NEW] ovs cleanup deletes nova ports

2019-11-22 Thread Darragh O'Reilly
Public bug reported:

The help for ovs_all_ports says that when False, only ports that were
created by Neutron are deleted.

https://github.com/openstack/neutron/blob/67b613b795416406fb4fab143b3ec9ba8657711f/neutron/conf/agent/cmd.py#L41-L46
cfg.BoolOpt('ovs_all_ports',
default=False,
help=_('True to delete all ports on all the OpenvSwitch '
   'bridges. False to delete ports created by '
   'Neutron on integration and external network '
   'bridges.'))


But actually ports created by Nova are also deleted. Those also have the 
iface-id and attached-mac fields.
https://github.com/openstack/neutron/blob/67b613b795416406fb4fab143b3ec9ba8657711f/neutron/agent/ovsdb/impl_idl.py#L80-L81


# ovs-vsctl --columns=name,external_ids --format=table list Interface
name external_ids   

 
 
-
...
"qr-868ef235-bb" {attached-mac="fa:16:3e:bb:94:ca", 
iface-id="868ef235-bba4-4776-aac9-4b46a9a17def", iface-status=active}
...
"qvof43ccb24-99" {attached-mac="fa:16:3e:70:5c:b0", 
iface-id="f43ccb24-999e-4f9d-a9e4-07d11b9e5128", iface-status=active, 
vm-uuid="33c6fffa-f58b-4aab-b5ba-9960c1e02004"}


Not sure which is wrong, the help or the script. Is it supposed to delete ports 
that were created by nova?

** Affects: neutron
 Importance: Undecided
 Status: New


** Tags: ovs

** Tags added: ovs

-- 
You received this bug notification because you are a member of Yahoo!
Engineering Team, which is subscribed to neutron.
https://bugs.launchpad.net/bugs/1853582

Title:
  ovs cleanup deletes nova ports

Status in neutron:
  New

Bug description:
  The help for ovs_all_ports says that when False, only ports that were
  created by Neutron are deleted.

  
https://github.com/openstack/neutron/blob/67b613b795416406fb4fab143b3ec9ba8657711f/neutron/conf/agent/cmd.py#L41-L46
  cfg.BoolOpt('ovs_all_ports',
  default=False,
  help=_('True to delete all ports on all the OpenvSwitch '
 'bridges. False to delete ports created by '
 'Neutron on integration and external network '
 'bridges.'))

  
  But actually ports created by Nova are also deleted. Those also have the 
iface-id and attached-mac fields.
  
https://github.com/openstack/neutron/blob/67b613b795416406fb4fab143b3ec9ba8657711f/neutron/agent/ovsdb/impl_idl.py#L80-L81

  
  # ovs-vsctl --columns=name,external_ids --format=table list Interface
  name external_ids 

   
   
-
  ...
  "qr-868ef235-bb" {attached-mac="fa:16:3e:bb:94:ca", 
iface-id="868ef235-bba4-4776-aac9-4b46a9a17def", iface-status=active}
  ...
  "qvof43ccb24-99" {attached-mac="fa:16:3e:70:5c:b0", 
iface-id="f43ccb24-999e-4f9d-a9e4-07d11b9e5128", iface-status=active, 
vm-uuid="33c6fffa-f58b-4aab-b5ba-9960c1e02004"}

  
  Not sure which is wrong, the help or the script. Is it supposed to delete 
ports that were created by nova?

To manage notifications about this bug go to:
https://bugs.launchpad.net/neutron/+bug/1853582/+subscriptions

-- 
Mailing list: https://launchpad.net/~yahoo-eng-team
Post to : yahoo-eng-team@lists.launchpad.net
Unsubscribe : https://launchpad.net/~yahoo-eng-team
More help   : https://help.launchpad.net/ListHelp